For the latest, visit Trending Questions on AI and Emerging Technologies. Here, find past questions and answers on emerging technologies.
For the latest, visit Trending Questions on AI and Emerging Technologies. Here, find past questions and answers on emerging technologies.
June 2025
Gartner defines artificial general intelligence (AGI) as the (currently hypothetical) capability of a machine that can match or surpass the capabilities of humans across all cognitive tasks. In addition, AGI will be able to autonomously learn and adapt in pursuit of predetermined or novel goals in a wide range of physical and virtual environments.
Narrow AI is designed to perform specific tasks (such as image recognition, language translation or playing chess) or solve particular problems. Narrow AI systems rely on supervised learning and are limited to the data they are trained on. They cannot learn or adapt beyond their specific programming and training, making them less flexible in unfamiliar situations.
AGI is currently theoretical — no existing systems embody AGI capabilities. While the pursuit of AGI is often associated with ambitious goals in AI research and development, the technology raises significant ethical, societal and regulatory questions. Gartner’s position is that AGI will not be a reality in the next 10 years. Yet there will be progress toward AGI — and every step along the way can be turned into new types of business value.
AGI shouldn’t mimic human intelligence. It should, however, lead to new types of value. AGI also should not become a single-entity super intelligence, as technology should always support people in their endeavours. Should AGI appear, we believe it will be best achieved by connecting countless narrow AI agents to create a synergetic type of AI.
Consider a comprehensive investment strategy with five key components:
1. Infrastructure planning: Invest in AI-optimized infrastructure as a service (IaaS), which is projected to grow significantly over the next three years. This infrastructure should be designed to support AI workloads efficiently, leveraging cloud scalability and pay-as-you-go pricing models. Implement AI supercomputing architectures that integrate AI accelerators, optimized storage and high-speed networking. This architecture is crucial for handling the extreme-scale parallelism required for training generative AI models.
2. GPU acquisition and management: Use features like dynamic workload scheduling and GPU sharing to maximize the use of GPU resources. This is particularly important given ongoing GPU supply constraints and the increasing demand for inference workloads. Also reduce switching costs from traditional GPUs to cloud-optimized AI processors by investing in open standards and ecosystems.
3. Networking and storage enhancements: Build out dedicated networks for AI workloads, ensuring low-latency and lossless networking capabilities. This includes using InfiniBand or high-speed ethernet to connect GPUs effectively. Implement storage solutions that support high throughput and low latency, such as parallel file systems and object storage. These systems should be capable of handling the large datasets required for AI training and inference.
4. Cost management and efficiency: Integrate monitoring tools to track GPU usage and power consumption, which allows for better management of operational IT costs and efficiency gains. As AI workloads can significantly increase energy consumption, consider investing in advanced cooling technologies (like immersion cooling) to manage the high power density of AI computing infrastructure.
5. Training and skill development: Develop training programs for infrastructure and operations (I&O) teams to enhance their skills in managing AI infrastructure, including prompt engineering and AI data center design.
By Mordecai .
Establishing an AI center of excellence (CoE) is a strategic initiative that can significantly enhance an organization's ability to leverage AI technologies effectively. Important steps and considerations include;
Establish executive support: Secure an executive sponsor who will be actively involved in the CoE's projects and help navigate organizational challenges, and ensure the sponsor communicates the CoE's value to other business leaders.
Define CoE objectives and scope: Co-create a detailed charter that includes goals, principles, roles, metrics and communication procedures and aligns the CoE with organizational priorities Establish success metrics that are specific, measurable and relevant to leadership.
Assemble a diverse team: Ensure a generalist mindset is active and the focus not on the technology alone.Include those with a mix of technical and soft skills, such as AI architects, data scientists, and change management specialists, to manage designated activities effectively. Consider using consultants for initial projects to fill skills gaps and facilitate knowledge transfer.
Select impactful metrics: They should cover inputs, processes and outputs to provide a comprehensive evaluation of effectiveness, reflect organizational needs and link technology-driven metrics to business outcomes.
Determine the right organizational structure: Choose a model that can evolve as the enterprise matures. Common models include centralized, decentralized and federated structures.
Foster a culture of continuous improvement: Implement training and development programs to enhance the skills of CoE staff. Create recognition programs to celebrate successes and encourage engagement.
Leverage existing resources and tools: Conduct a thorough assessment of the existing IT landscape to identify tools that can be leveraged for AI initiatives. Establish partnerships with technology vendors to enhance the CoE's capabilities.
Create knowledge platforms (Innovation as a Service): Hub playbooks, technical guides, basic knowledge, and key use cases for access across the enterprise.
Monitor and evaluate effectiveness: Conduct surveys to gather insights on the CoE's impact and areas for improvement. Focus on measures that go beyond or are the result of increased productivity and efficiencies. Report metrics and success stories to senior leaders to secure ongoing support for the CoE.
There is a lot of distrust around AI due to the incidence of a "black box" effect — we see results but cannot understand how they were arrived at. Lack of AI trust, risk and security (AI TRiSM) is further exacerbated by the lack of visibility in the underlying data that AI models used to create their results. AI-powered robots ultimately need humans to take accountability such that they perform their functions in a safe, secure, fair and socially beneficial manner.
Whether evaluating AI-powered robots or other AI use cases, Gartner recommends:
Collaborating with relevant stakeholders — enterprise architecture, IT, HR, legal and data analytics leaders — to embrace the potential of enhanced ethical AI governance platforms. Operationalize the key foundational principles of ethics and embed them in the broader AI strategy.
Addressing ethical dilemmas around the use and development of AI systems by adding an AI ethicist to the AI central team and/or forming a digital ethics advisory board to validate otherwise ambiguous AI outcomes and improve their contextual accuracy.
Continuously monitoring the behavior of AI systems by implementing a robust AI testing program, aligned with basic ethical principles to avoid potential negative consequences, such as legal backlash.
As of now, only 6% of infrastructure and operations (I&O) organizations have achieved the highest maturity level necessary for the successful adoption of agentic AI. At most organizations, agentic AI is limited to isolated deployments that provide some benefits but are not transformative.
The adoption roadmap for agentic AI requires a structured, multi-prong approach:
Assessment and readiness: Begin by assessing your current state and identifying gaps in your infrastructure and operational models that may hinder the adoption of agentic AI. This includes evaluating data quality, governance and existing technology capabilities.
Strategic vision: Create a vision for agentic AI adoption that identifies constraints in decision-making and explores how these could be alleviated through agentic AI. This vision should align with the organization's overall IT operating model.
Investment in infrastructure: Focus future investments on ensuring that new infrastructure is compatible with agentic AI. This includes adopting intelligent infrastructures and platforms that can support the autonomous capabilities of AI agents.
Governance and management: Establish a governance framework that includes a Center of Excellence (CoE), staffed with IT professionals, business technologists and business leaders, to oversee the implementation and scaling of agentic AI initiatives.
Continuous improvement: Regularly reassess your AI maturity and adjust strategies based on evolving technologies and market demands. This includes refreshing the AI maturity assessment annually.
Implementation and iteration: After securing funding and developing a clear implementation plan, execute strategies iteratively.
April 2025
By Daniel Sun
Manus AI, developed by the Chinese startup Monica.im and launched on March 6, 2025, is a general-purpose (large language model) LLM assistant with the potential to evolve into a general-purpose AI agent. It is a step up in AI technology, focused on enhancing human-machine collaboration and task automations of AI applications.
Capabilities: Manus AI offers a wide range of functionalities, including task automation, data processing, analysis and decision making and content generation. Manus AI can also learn from interactions and adapt its actions accordingly — for example by self-prompting tasks or modifying plans for a new context.
Technology: Manus AI leverages existing lLLMs, including Anthropic's Claude and Alibaba's Qwen, and employs a multiagent architecture. This allows it to utilize various tools and open-source software, enhancing its capabilities in task execution and interaction with APIs.
User Interface: A chatbot with a simplified user experience makes it accessible for users to interact with the AI while it manages tasks in the background.
Applications: The potential use cases of Manus AI for business include streamlining complex processes, enhancing customer engagement through automated responses and assisting in tasks like financial analysis.
Challenges: Operational stability can be an issue since Manus AI relies on third-party systems and APIs, which may pose risks in enterprise environments. Ethical and regulatory concerns also exist. These include questions of accountability and liability when AI makes critical decisions and uncertainty remains about its data sources and operational mechanisms. Its pricing is not clearly defined.
Small language models (SLMs) are a category of language models that are sized to be more efficient and cost-effective alternatives to large language models (LLMs).
SLMs are of the order of size of 10 billion parameters. Their strength is realized when tailored for specific tasks and domains. The smaller size makes them easier to fine-tune and deploy than their LLM counterparts.
Cost-effectiveness: SLMs are relatively low cost to fine-tune and serve, making them suitable for organizations with budget constraints.
Efficiency: They require significantly fewer computational resources, which allows for deployment on-device, on-premises or in private cloud environments, enhancing data privacy and security.
Task specialization: When appropriately aligned to a specific use case and fine-tuned for that implementation, they can outperform larger models. This makes them particularly useful in scenarios where larger language capabilities are not necessary.
SLMs are ideal for applications that involve high user interaction volumes, low-latency requirements, and sensitive data scenarios. They can be effectively used in document chatbots, customer service applications, and organizational language models.
They can also be integrated into multimodel systems, where multiple task-specialized models work together to address complex AI challenges.
While SLMs do not match the generic language reasoning and generation capabilities of LLMs, they excel in specific applications where their smaller size and lower resource requirements provide a significant advantage.
SLMs represent a practical and efficient approach to implementing generative AI solutions, particularly for organizations that require specialized capabilities without the overhead of larger models. For this reason, SLM adoption is expected to grow. Gartner projects that by 2026, more GenAI queries and requests will be processed on-device than in the cloud, enabling intelligent assistants and other proactive applications.
Modern edge devices are becoming increasingly capable of running complex algorithms. At the same time, AI is being reshaped and adapted to the capabilities of those devices, allowing smarter, more efficient and autonomous applications. As AI models progress, we will see significant efficiency improvements in on-device inference with limited computational capacity and can expect even greater integration, driving innovation across various sectors.
Real-time processing: AI enables edge devices to perform real-time data processing and analytics locally, reducing latency and bandwidth usage and removing the dependency on network infrastructure. This is crucial for applications that require immediate decision making and robust processes, such as autonomous vehicles and industrial automation.
Privacy and security: By computing data locally, AI helps address privacy concerns, as sensitive information does not need to be transmitted to the cloud, but gets processed during runtime without storing raw data. This is particularly important for processing personal information, like biometrics, or in other heavily regulated industries where data sovereignty is a concern.
Integration with IoT: Edge devices are increasingly integrated into Internet of Things (IoT) meshes, allowing for smarter, more autonomous operations. Integrating AI onto the edge enables devices to learn from their environment and improve their performance over time, leading to more efficient operations and better user experiences.
Use of advanced techniques: Techniques such as TinyML, transfer learning and federated learning are being employed to optimize model training for edge devices. These advancements enable edge devices to perform machine learning tasks without needing extensive computational resources.
Cost efficiency: By reducing the need for data transmission to centralized systems, AI at the edge can lower operational costs long term, while initially increasing capital expenditures. This is particularly beneficial for applications that process large volumes of data, but only need to transfer a federated or aggregated summary, as bandwidth requirements and associated costs are minimized, compared to processing everything in the cloud.
Scalability and flexibility: AI enhances the scalability of edge solutions by allowing for distributed processing and decision making. New nodes on the edge can be applied without increasing the demand for scaling central IT structures significantly. This flexibility is essential for adapting to various use cases across different industries, from smart cities to health care.
Energy efficiency: Processing data directly on the edge device significantly reduces the overall energy consumption of the whole network infrastructure by minimizing unnecessary traffic or eliminating additional handling steps like safeguarding algorithms for sensitive raw data.
Flexible infrastructure: AI-powered edge devices can be part of larger composite structures to provide advanced functionalities via cloud-to-edge and edge-to-cloud integrations. This allows for synchronization with larger orchestrations while maintaining local execution.
By Alizeh Khare
DeepSeek made a significant impact on the AI market in January 2025 with the launch of its R1 model, developed using distillation techniques.
DeepSeek R1 is a family of AI models focused on logical and reasoning tasks, available in two versions: DeepSeek R1, launched on January 20, 2025, and accessible on multiple platforms, and DeepSeek R1 Zero, which excels in accuracy and performance but is not yet released.
The architecture employs a mixture-of-experts (MoE) framework, activating only 37 billion of its 671 billion parameters per query, enhancing efficiency and resource conservation. According to DeepSeek, its R1 model is 20 to 50 times less expensive to use than OpenAI o1 model, depending on the task, and has dramatically smaller graphics processing unit (GPU) requirements.
DeepSeek’s models have demonstrated competitive performance, but the bold claims of cost and development efficiency haven’t been independently reviewed or validated. Still, the innovative approach and transparency allows researchers, competitors, and developers to understand both the strengths and limitations of the R1 model, which can potentially change the AI landscape by making it less expensive for providers to build and run AI models.
Use of open-source principles in DeepSeek model development
The techniques that DeepSeek used to achieve model-training breakthroughs are not proprietary. They are part of the open-source toolset and therefore available for other providers to study and learn from. This approach reinforces the notion that open-source AI is becoming more competitive with, and perhaps even surpassing, the closed, proprietary models of major technology firms.
The open source approach also aligns with the growing calls for ethical AI development, as it allows for greater scrutiny and accountability in how AI models are built and deployed.
Even with DeepSeek's assurances of robust data security, users are concerned about the management and potential sharing of their data. The absence of clear data-handling policies could undermine trust, particularly in regions with strict privacy regulations like the EU's GDPR, where DeepSeek faces scrutiny and restrictions from numerous companies and government agencies.
Still, the advancements made by DeepSeek could lead to a broader adoption of domain-specific AI models and a focus on sustainable AI practices.
Domain-specific AI models (DSAMs) and generic large language models (LLMs) serve different purposes and have distinct characteristics that make them suitable for various enterprise applications. While LLMs are designed to be used across any and all classes of tasks, issues and problems in a domain-agnostic manner, DSAMs are designed to address issues in a specific industry, business function or a class of problem (resource optimization). Thus, the training methodology used and the benefits each brings will be distinct.
Purpose and specialization
DSAMs: Tailored for specific industries, business functions, or tasks and optimized to understand and generate content relevant to particular domains, such as healthcare, finance, or legal sectors, providing greater accuracy, reliability, and contextual understanding.
Generic LLMs: Designed to handle a wide range of tasks across various domains, they excel in general language understanding and generation.
Training data
DSAMs: Often fine-tuned on datasets that are specific to their target domain, which helps them learn the nuances and terminologies relevant to that field. This reduces the need for extensive prompt engineering and enhances the model's performance in specific tasks.
Generic LLMs: Trained on vast amounts of unlabeled text data from diverse sources, which gives them broad capabilities.
Cost and resource efficiency
DSAMs: Tend to be more cost-effective for enterprises as they require less computational power and data for fine-tuning than generic LLMs, making them easier to deploy and maintain, especially in environments with limited resources.
Generic LLMs: Can be resource-intensive to deploy and fine-tune, requiring more parameters, energy, and higher cost to achieve the same performance level as DSAMs.
Use cases
DSAMs: Ideal for applications that require deep knowledge of a specific field, such as medical diagnosis, legal document analysis, or financial forecasting. They can provide more relevant and accurate outputs for these specialized tasks.
Generic LLMs: Suitable for a wide range of general tasks, such as content creation, summarization and information retrieval for conversational agents.
By Gary Olliffe
As enterprises seek to implement AI agents that take on more complex tasks across workflows and processes, those agents will need to access a variety of data, information and functional resources. Model context protocol (MCP) is an open standard interface that defines a consistent and flexible way to connect AI applications and agents to these resources.
MCP defines how AI applications access and interact with other resources — such as information stores, APIs and other applications — in a model and application independent way. As AI agents evolve from isolated assistants into collaborative systems, they are increasingly relied upon to automate operations, augment productivity and support end-to-end business processes that require flexible integration.
Without MCP, each AI application framework or platform will define its own proprietary approach to connecting external resources to AI. With MCP, organizations will be able to share resources across a variety of AI use cases in a consistent and manageable way, reducing duplication and allowing developers and users to connect their AI applications to the most appropriate resources for a task or use case.
In software development, for example, agents will support tasks such as code modernization, test automation and issue resolution, ensuring each agent has the correct tools and context across stages is essential for efficiency and accuracy. MCP provides a standard mechanism for integrating these capabilities into off-the-shelf or custom-developed AI applications..
As AI agents become more autonomous and interdependent, IT executives should evaluate whether their AI architectures are designed to support flexible integration with shareable context. MCP is in the early stages of its lifecycle, and will need to mature rapidly to meet the enterprise needs, but its rapid growth in popularity highlights the demand for a standard interface for LLM-based AI applications and agents to connect to external context.