agentic ai Archives - AI News https://www.artificialintelligence-news.com/news/tag/agentic-ai/ Artificial Intelligence News Fri, 25 Apr 2025 14:07:36 +0000 en-GB hourly 1 https://wordpress.org/?v=6.8.1 https://www.artificialintelligence-news.com/wp-content/uploads/2020/09/cropped-ai-icon-32x32.png agentic ai Archives - AI News https://www.artificialintelligence-news.com/news/tag/agentic-ai/ 32 32 Beyond acceleration: the rise of Agentic AI https://www.artificialintelligence-news.com/news/beyond-acceleration-the-rise-of-agentic-ai/ https://www.artificialintelligence-news.com/news/beyond-acceleration-the-rise-of-agentic-ai/#respond Mon, 07 Apr 2025 06:31:00 +0000 https://www.artificialintelligence-news.com/?p=105001 We already find ourselves at an inflection point with AI. According to a recent study by McKinsey, we’ve reached the turning point where ‘businesses must look beyond automation and towards AI-driven reinvention’ to stay ahead of the competition. While the era of AI-driven acceleration isn’t over, a new phase has already begun – one that […]

The post Beyond acceleration: the rise of Agentic AI appeared first on AI News.

]]>
We already find ourselves at an inflection point with AI. According to a recent study by McKinsey, we’ve reached the turning point where ‘businesses must look beyond automation and towards AI-driven reinvention’ to stay ahead of the competition. While the era of AI-driven acceleration isn’t over, a new phase has already begun – one that goes beyond making existing workflows more efficient and moves toward replacing existing workflows and/or creating new ones.

This is the age of Agentic AI.

Truly autonomous AI agents are capable of reshaping operations entirely. Systems can act autonomously, make decisions, and adapt dynamically. These agents will go beyond conversational interfaces, responding to user input and proactively managing tasks, navigating complex IT environments, and orchestrating business processes.

However, this shift isn’t just about technology — it also comes with a few considerations. Companies will need to address regulatory challenges, build AI literacy, and focus on applied use cases with clear ROI if the evolution is to succeed.

Moving from acceleration to transformation

So far, companies have primarily used AI to accelerate existing processes, whether through chatbots improving customer interactions or AI-driven analytics optimising workflows. In the end, these implementations make businesses more efficient.

But acceleration alone is no longer enough to stay ahead in the game. The real opportunity lies in replacing outdated workflows entirely and creating new, previously impossible capabilities.

For example, AI plays a vital role in automating troubleshooting and enhancing security within the network industry. But what if AI could autonomously anticipate and predict failures, reconfigure networks proactively to avoid service level degradations in real time, and optimise performance without human intervention? As AI becomes more autonomous, its ability to not just assist but act independently will be key to unlocking new levels of productivity and innovation.

That’s what Agentic AI is about.

Navigating the AI regulatory landscape

However, as AI becomes more autonomous, the regulatory landscape governing its deployment will evolve in parallel. The introduction of the EU AI Act, alongside global regulatory frameworks, means companies must already navigate new compliance requirements related to AI transparency, bias mitigation, and ethical deployment.

That means AI governance can no longer be an afterthought.

AI-powered systems must be designed with built-in compliance mechanisms, data privacy protections, and explainability features to build trust among users and regulators alike. Zero-trust security models will also be crucial in mitigating risks, enforcing strict access controls, and ensuring that AI decisions remain auditable and secure.

The importance of AI literacy

As stated, the success of Agentic AI’s era will depend on more than just technical capabilities – it will require alignment between leadership, developers, and end-users. As AI becomes more advanced, AI literacy becomes a key differentiator, and companies must invest in upskilling their workforce to understand AI’s capabilities, limitations, and ethical considerations. A recent report by the ICT Workforce Consortium found that 92% of information and communication technology jobs are expected to undergo significant transformation due to advancements in AI. So, without proper AI education, businesses risk misalignment between AI implementers and those who use the technology.

This can lead to a lack of trust, slow adoption, and ineffective deployment, which can impact the bottom line. So, to unlock the full potential of Agentic AI, it’s essential to build AI literacy across all levels of the organisation.

As this new era of AI blooms, companies must learn from the current era of AI adoption: focus on applied use cases with tangible ROI. The days of experimenting with AI for innovation’s sake are ending – the next generation of AI deployments must prove their worth.

In networking, it could be projects such as AI-powered autonomous network optimisation. These systems do more than automate tasks; they continuously monitor network traffic, predict congestion points, and autonomously adjust configurations to ensure optimal performance. By providing proactive insights and real-time adjustments, these AI-driven solutions help companies prevent issues and outages before they occur.

This level of AI autonomy reduces human intervention and enhances overall security and operational efficiency.

Identifying and implementing high-value, high-impact Agentic AI use cases such as these will be vital.

Trust as the adoption hurdle

While we’re entering a new era, trust plays a key role in widespread AI adoption. Users must feel confident that AI decisions are accurate, fair, and explainable. Even the most advanced AI models will face challenges gaining acceptance without transparency.

This is particularly relevant as AI transitions from assisting users to making autonomous decisions. Whether AI agents manage IT infrastructure or drive customer interactions, organisations must ensure that AI decisions are auditable, unbiased, and aligned with business objectives.

Without transparency and accountability, companies may face resistance from both employees and customers.

The future of AI

Looking ahead, 2025 holds exciting potential for AI. As it reaches a new level of maturity, its success will depend on how well organisations, governments, and individuals adapt to its growing presence in everyday life. Moving beyond efficiency and automation, AI has the opportunity to become a powerful driver of intelligent decision-making, problem-solving, and innovation.

Organisations that harness Agentic AI effectively – balancing autonomy with oversight – will see the greatest benefits. However, success will require a commitment to transparency, education, and ethical deployment to build trust and ensure AI is a true enabler of progress.

Because AI is no longer just an accelerant, it is a transformative force reshaping how we work, communicate, and interact with technology.

Photo by Ryan De Hamer on Unsplash

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation ConferenceBlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Beyond acceleration: the rise of Agentic AI appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/beyond-acceleration-the-rise-of-agentic-ai/feed/ 0
Amazon Nova Act: A step towards smarter, web-native AI agents https://www.artificialintelligence-news.com/news/amazon-nova-act-step-towards-smarter-web-native-ai-agents/ https://www.artificialintelligence-news.com/news/amazon-nova-act-step-towards-smarter-web-native-ai-agents/#respond Tue, 01 Apr 2025 16:57:43 +0000 https://www.artificialintelligence-news.com/?p=105105 Amazon has introduced Nova Act, an advanced AI model engineered for smarter agents that can execute tasks within web browsers. While large language models popularised the concept of “agents” as tools that answer queries or retrieve information via methods such as Retrieval-Augmented Generation (RAG), Amazon envisions something more robust. The company defines agents not just […]

The post Amazon Nova Act: A step towards smarter, web-native AI agents appeared first on AI News.

]]>
Amazon has introduced Nova Act, an advanced AI model engineered for smarter agents that can execute tasks within web browsers.

While large language models popularised the concept of “agents” as tools that answer queries or retrieve information via methods such as Retrieval-Augmented Generation (RAG), Amazon envisions something more robust. The company defines agents not just as responders but as entities capable of performing tangible, multi-step tasks in diverse digital and physical environments.

“Our dream is for agents to perform wide-ranging, complex, multi-step tasks like organising a wedding or handling complex IT tasks to increase business productivity,” said Amazon.

Current market offerings often fall short, with many agents requiring continuous human supervision and their functionality dependent on comprehensive API integration—something not feasible for all tasks. Nova Act is Amazon’s answer to these limitations.

Alongside the model, Amazon is releasing a research preview of the Amazon Nova Act SDK. Using the SDK, developers can create agents capable of automating web tasks like submitting out-of-office notifications, scheduling calendar holds, or enabling automatic email replies.

The SDK aims to break down complex workflows into dependable “atomic commands” such as searching, checking out, or interacting with specific interface elements like dropdowns or popups. Detailed instructions can be added to refine these commands, allowing developers to, for instance, instruct an agent to bypass an insurance upsell during checkout.

To further enhance accuracy, the SDK supports browser manipulation via Playwright, API calls, Python integrations, and parallel threading to overcome web page load delays.

Nova Act: Exceptional performance on benchmarks

Unlike other generative models that showcase middling accuracy on complex tasks, Nova Act prioritises reliability. Amazon highlights its model’s impressive scores of over 90% on internal evaluations for specific capabilities that typically challenge competitors. 

Nova Act achieved a near-perfect 0.939 on the ScreenSpot Web Text benchmark, which measures natural language instructions for text-based interactions, such as adjusting font sizes. Competing models such as Claude 3.7 Sonnet (0.900) and OpenAI’s CUA (0.883) trail behind by significant margins.

Similarly, Nova Act scored 0.879 in the ScreenSpot Web Icon benchmark, which tests interactions with visual elements like rating stars or icons. While the GroundUI Web test, designed to assess an AI’s proficiency in navigating various user interface elements, showed Nova Act slightly trailing competitors, Amazon sees this as an area ripe for improvement as the model evolves.

Amazon stresses its focus on delivering practical reliability. Once an agent built using Nova Act functions as expected, developers can deploy it headlessly, integrate it as an API, or even schedule it to run tasks asynchronously. In one demonstrated use case, an agent automatically orders a salad for delivery every Tuesday evening without requiring ongoing user intervention.

Amazon sets out its vision for scalable and smart AI agents

One of Nova Act’s standout features is its ability to transfer its user interface understanding to new environments with minimal additional training. Amazon shared an instance where Nova Act performed admirably in browser-based games, even though its training had not included video game experiences. This adaptability positions Nova Act as a versatile agent for diverse applications.

This capability is already being leveraged in Amazon’s own ecosystem. Within Alexa+, Nova Act enables self-directed web navigation to complete tasks for users, even when API access is not comprehensive enough. This represents a step towards smarter AI assistants that can function independently, harnessing their skills in more dynamic ways.

Amazon is clear that Nova Act represents the first stage in a broader mission to craft intelligent, reliable AI agents capable of handling increasingly complex, multi-step tasks. 

Expanding beyond simple instructions, Amazon’s focus is on training agents through reinforcement learning across varied, real-world scenarios rather than overly simplistic demonstrations. This foundational model serves as a checkpoint in a long-term training curriculum for Nova models, indicating the company’s ambition to reshape the AI agent landscape.

“The most valuable use cases for agents have yet to be built,” Amazon noted. “The best developers and designers will discover them. This research preview of our Nova Act SDK enables us to iterate alongside these builders through rapid prototyping and iterative feedback.”

Nova Act is a step towards making AI agents truly useful for complex, digital tasks. From rethinking benchmarks to emphasising reliability, its design philosophy is centred around empowering developers to move beyond what’s possible with current-generation tools. 

See also: Anthropic provides insights into the ‘AI biology’ of Claude

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Amazon Nova Act: A step towards smarter, web-native AI agents appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/amazon-nova-act-step-towards-smarter-web-native-ai-agents/feed/ 0
NVIDIA Dynamo: Scaling AI inference with open-source efficiency https://www.artificialintelligence-news.com/news/nvidia-dynamo-scaling-ai-inference-open-source-efficiency/ https://www.artificialintelligence-news.com/news/nvidia-dynamo-scaling-ai-inference-open-source-efficiency/#respond Wed, 19 Mar 2025 16:49:21 +0000 https://www.artificialintelligence-news.com/?p=104933 NVIDIA has launched Dynamo, an open-source inference software designed to accelerate and scale reasoning models within AI factories. Efficiently managing and coordinating AI inference requests across a fleet of GPUs is a critical endeavour to ensure that AI factories can operate with optimal cost-effectiveness and maximise the generation of token revenue. As AI reasoning becomes […]

The post NVIDIA Dynamo: Scaling AI inference with open-source efficiency appeared first on AI News.

]]>
NVIDIA has launched Dynamo, an open-source inference software designed to accelerate and scale reasoning models within AI factories.

Efficiently managing and coordinating AI inference requests across a fleet of GPUs is a critical endeavour to ensure that AI factories can operate with optimal cost-effectiveness and maximise the generation of token revenue.

As AI reasoning becomes increasingly prevalent, each AI model is expected to generate tens of thousands of tokens with every prompt, essentially representing its “thinking” process. Enhancing inference performance while simultaneously reducing its cost is therefore crucial for accelerating growth and boosting revenue opportunities for service providers.

A new generation of AI inference software

NVIDIA Dynamo, which succeeds the NVIDIA Triton Inference Server, represents a new generation of AI inference software specifically engineered to maximise token revenue generation for AI factories deploying reasoning AI models.

Dynamo orchestrates and accelerates inference communication across potentially thousands of GPUs. It employs disaggregated serving, a technique that separates the processing and generation phases of large language models (LLMs) onto distinct GPUs. This approach allows each phase to be optimised independently, catering to its specific computational needs and ensuring maximum utilisation of GPU resources.

“Industries around the world are training AI models to think and learn in different ways, making them more sophisticated over time,” stated Jensen Huang, founder and CEO of NVIDIA. “To enable a future of custom reasoning AI, NVIDIA Dynamo helps serve these models at scale, driving cost savings and efficiencies across AI factories.”

Using the same number of GPUs, Dynamo has demonstrated the ability to double the performance and revenue of AI factories serving Llama models on NVIDIA’s current Hopper platform. Furthermore, when running the DeepSeek-R1 model on a large cluster of GB200 NVL72 racks, NVIDIA Dynamo’s intelligent inference optimisations have shown to boost the number of tokens generated by over 30 times per GPU.

To achieve these improvements in inference performance, NVIDIA Dynamo incorporates several key features designed to increase throughput and reduce operational costs.

Dynamo can dynamically add, remove, and reallocate GPUs in real-time to adapt to fluctuating request volumes and types. The software can also pinpoint specific GPUs within large clusters that are best suited to minimise response computations and efficiently route queries. Dynamo can also offload inference data to more cost-effective memory and storage devices while retrieving it rapidly when required, thereby minimising overall inference costs.

NVIDIA Dynamo is being released as a fully open-source project, offering broad compatibility with popular frameworks such as PyTorch, SGLang, NVIDIA TensorRT-LLM, and vLLM. This open approach supports enterprises, startups, and researchers in developing and optimising novel methods for serving AI models across disaggregated inference infrastructures.

NVIDIA expects Dynamo to accelerate the adoption of AI inference across a wide range of organisations, including major cloud providers and AI innovators like AWS, Cohere, CoreWeave, Dell, Fireworks, Google Cloud, Lambda, Meta, Microsoft Azure, Nebius, NetApp, OCI, Perplexity, Together AI, and VAST.

NVIDIA Dynamo: Supercharging inference and agentic AI

A key innovation of NVIDIA Dynamo lies in its ability to map the knowledge that inference systems hold in memory from serving previous requests, known as the KV cache, across potentially thousands of GPUs.

The software then intelligently routes new inference requests to the GPUs that possess the best knowledge match, effectively avoiding costly recomputations and freeing up other GPUs to handle new incoming requests. This smart routing mechanism significantly enhances efficiency and reduces latency.

“To handle hundreds of millions of requests monthly, we rely on NVIDIA GPUs and inference software to deliver the performance, reliability and scale our business and users demand,” said Denis Yarats, CTO of Perplexity AI.

“We look forward to leveraging Dynamo, with its enhanced distributed serving capabilities, to drive even more inference-serving efficiencies and meet the compute demands of new AI reasoning models.”

AI platform Cohere is already planning to leverage NVIDIA Dynamo to enhance the agentic AI capabilities within its Command series of models.

“Scaling advanced AI models requires sophisticated multi-GPU scheduling, seamless coordination and low-latency communication libraries that transfer reasoning contexts seamlessly across memory and storage,” explained Saurabh Baji, SVP of engineering at Cohere.

“We expect NVIDIA Dynamo will help us deliver a premier user experience to our enterprise customers.”

Support for disaggregated serving

The NVIDIA Dynamo inference platform also features robust support for disaggregated serving. This advanced technique assigns the different computational phases of LLMs – including the crucial steps of understanding the user query and then generating the most appropriate response – to different GPUs within the infrastructure.

Disaggregated serving is particularly well-suited for reasoning models, such as the new NVIDIA Llama Nemotron model family, which employs advanced inference techniques for improved contextual understanding and response generation. By allowing each phase to be fine-tuned and resourced independently, disaggregated serving improves overall throughput and delivers faster response times to users.

Together AI, a prominent player in the AI Acceleration Cloud space, is also looking to integrate its proprietary Together Inference Engine with NVIDIA Dynamo. This integration aims to enable seamless scaling of inference workloads across multiple GPU nodes. Furthermore, it will allow Together AI to dynamically address traffic bottlenecks that may arise at various stages of the model pipeline.

“Scaling reasoning models cost effectively requires new advanced inference techniques, including disaggregated serving and context-aware routing,” stated Ce Zhang, CTO of Together AI.

“The openness and modularity of NVIDIA Dynamo will allow us to seamlessly plug its components into our engine to serve more requests while optimising resource utilisation—maximising our accelerated computing investment. We’re excited to leverage the platform’s breakthrough capabilities to cost-effectively bring open-source reasoning models to our users.”

Four key innovations of NVIDIA Dynamo

NVIDIA has highlighted four key innovations within Dynamo that contribute to reducing inference serving costs and enhancing the overall user experience:

  • GPU Planner: A sophisticated planning engine that dynamically adds and removes GPUs based on fluctuating user demand. This ensures optimal resource allocation, preventing both over-provisioning and under-provisioning of GPU capacity.
  • Smart Router: An intelligent, LLM-aware router that directs inference requests across large fleets of GPUs. Its primary function is to minimise costly GPU recomputations of repeat or overlapping requests, thereby freeing up valuable GPU resources to handle new incoming requests more efficiently.
  • Low-Latency Communication Library: An inference-optimised library designed to support state-of-the-art GPU-to-GPU communication. It abstracts the complexities of data exchange across heterogeneous devices, significantly accelerating data transfer speeds.
  • Memory Manager: An intelligent engine that manages the offloading and reloading of inference data to and from lower-cost memory and storage devices. This process is designed to be seamless, ensuring no negative impact on the user experience.

NVIDIA Dynamo will be made available within NIM microservices and will be supported in a future release of the company’s AI Enterprise software platform. 

See also: LG EXAONE Deep is a maths, science, and coding buff

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post NVIDIA Dynamo: Scaling AI inference with open-source efficiency appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/nvidia-dynamo-scaling-ai-inference-open-source-efficiency/feed/ 0
ServiceNow deploys AI agents to boost enterprise workflows https://www.artificialintelligence-news.com/news/servicenow-deploys-ai-agents-boost-enterprise-workflows/ https://www.artificialintelligence-news.com/news/servicenow-deploys-ai-agents-boost-enterprise-workflows/#respond Thu, 13 Mar 2025 16:40:58 +0000 https://www.artificialintelligence-news.com/?p=104777 ServiceNow has launched its Yokohama platform which introduces AI agents across various sectors to boost workflows and maximise end-to-end business impact. The Yokohama platform release features teams of preconfigured AI agents designed to deliver immediate productivity gains. These agents operate on a single, unified platform, ensuring seamless integration and coordination across different business functions. The […]

The post ServiceNow deploys AI agents to boost enterprise workflows appeared first on AI News.

]]>
ServiceNow has launched its Yokohama platform which introduces AI agents across various sectors to boost workflows and maximise end-to-end business impact.

The Yokohama platform release features teams of preconfigured AI agents designed to deliver immediate productivity gains. These agents operate on a single, unified platform, ensuring seamless integration and coordination across different business functions. The platform also includes capabilities to build, onboard, and manage the entire AI agent lifecycle, making it easier for enterprises to adopt and scale AI solutions.

Data is the lifeblood of AI, and ServiceNow recognises this by expanding its Knowledge Graph with advancements to its Common Service Data Model (CSDM). This expansion aims to break down barriers among data sources, enabling more connected and intelligent AI agents. By unifying data from various sources, ServiceNow’s platform ensures that AI agents can operate with a comprehensive view of the enterprise, driving more informed decisions and actions.

The growing need for ‘Guardian Agents’

According to Gartner, by 2028, 40% of CIOs will demand ‘Guardian Agents’ to autonomously track, oversee, or contain the results of AI agent actions. This underscores the growing need for a coordinated, enterprise-wide approach to AI deployment and management.

ServiceNow’s Yokohama release addresses this need by serving as the AI agent control tower for enterprises. The platform removes common roadblocks such as data fragmentation, governance gaps, and real-time performance challenges, ensuring seamless data connectivity with Workflow Data Fabric.

Unlike other AI providers that operate in silos or require complex integrations, ServiceNow AI Agents are built on a single, enterprise-wide platform. This ensures seamless data connectivity and provides a single view of all workflows, AI, and automation needs.

Amit Zavery, President, Chief Product Officer, and Chief Operating Officer at ServiceNow, commented: “Agentic AI is the new frontier. Enterprise leaders are no longer just experimenting with AI agents; they’re demanding AI solutions that can help them achieve productivity at scale.

“ServiceNow’s industry‑leading agentic AI framework meets this need by delivering predictability and efficiency from the start. With the combination of agentic AI, data fabric, and workflow automation all on one platform, we’re making it easier for organisations to embed connected AI where work happens and both measure and drive business outcomes faster, smarter, and at scale.”

New AI agents from ServiceNow aim to accelerate productivity

ServiceNow’s new AI Agents are now available to accelerate productivity at scale. These agents are designed to drive real outcomes for enterprise-wide use cases. For example:

  • Security Operations (SecOps) expert AI agents: These agents transform security operations by streamlining the entire incident lifecycle, eliminating repetitive tasks, and empowering SecOps teams to focus on stopping real threats quickly.
  • Autonomous change management AI agents: Acting like seasoned change managers, these agents generate custom implementation, test, and backout plans by analysing impact, historical data, and similar changes, ensuring seamless execution with minimal risk.
  • Proactive network test & repair AI agents: These AI-powered troubleshooters automatically detect, diagnose, and resolve network issues before they impact performance.

ServiceNow AI Agent Orchestrator and AI Agent Studio are now generally available with expanded capabilities to govern the complete AI agent lifecycle.

These tools help to streamline the setup process with guided instructions, making it easier to design and configure new AI agents using natural language descriptions. Their expanded performance management capabilities include an analytics dashboard for visualising AI agent usage, quality, and value—ensuring that AI agent performance and ROI can be easily tracked.

At the core of the ServiceNow Platform is Workflow Data Fabric, enabling AI-powered workflows that integrate with an organisation’s data, regardless of the system or source. This fabric allows businesses to gain deeper insights through AI-driven contextualisation and decision intelligence while automating manual work and creating process efficiencies.

The Yokohama release continues to expand ServiceNow’s Knowledge Graph data capabilities with enhancements to its Common Service Data Model (CSDM). CSDM provides a standardised framework for managing IT and business services to accelerate quick, safe, and compliant technology deployments.

Several customers and partners have already seen the benefits of ServiceNow’s AI solutions. CANCOM, Cognizant, Davies, and Sentara have all praised the platform’s ability to drive efficiency, cost savings, and productivity. These organisations have successfully integrated ServiceNow’s AI agents into their operations.

Jason Wojahn, Global Head of the ServiceNow Business Group at Cognizant, said: “At Cognizant, we are helping companies harness the next phase of AI with agentic AI workflows that could bring unparalleled efficiency. We were the first to bring ServiceNow’s Workflow Data Fabric to market and are working to help our clients to seamlessly connect their data with AI.

“With the Yokohama release and the integration of AI agents onto the Now Platform, clients can now operate their agents virtually effortlessly with connected data, driving productivity and ROI across their entire business.”

Darrell Burnell, Group Head of Technology at Davies, added: “Agility is essential for Davies, given our work with clients in heavily regulated markets. We’ve transformed our agent experience with ServiceNow’s generative AI, deploying Now Assist for ITSM in just six weeks to streamline information retrieval and accelerate resolution times.”

ServiceNow’s Yokohama platform release is a step forward in the evolution of AI for business transformation. By unleashing new AI agents and expanding data capabilities, ServiceNow aims to empower businesses to achieve faster and smarter workflows to maximise impact.

(Image by Thomas Fengler)

See also: Opera introduces browser-integrated AI agent

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post ServiceNow deploys AI agents to boost enterprise workflows appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/servicenow-deploys-ai-agents-boost-enterprise-workflows/feed/ 0
Opera introduces browser-integrated AI agent https://www.artificialintelligence-news.com/news/opera-introduces-browser-integrated-ai-agent/ https://www.artificialintelligence-news.com/news/opera-introduces-browser-integrated-ai-agent/#respond Mon, 03 Mar 2025 16:34:09 +0000 https://www.artificialintelligence-news.com/?p=104668 Opera has introduced “Browser Operator,” a native AI agent designed to perform tasks for users directly within the browser. Rather than acting as a separate tool, Browser Operator is an extension of the browser itself—designed to empower users by automating repetitive tasks like purchasing products, completing online forms, and gathering web content. Unlike server-based AI […]

The post Opera introduces browser-integrated AI agent appeared first on AI News.

]]>
Opera has introduced “Browser Operator,” a native AI agent designed to perform tasks for users directly within the browser.

Rather than acting as a separate tool, Browser Operator is an extension of the browser itself—designed to empower users by automating repetitive tasks like purchasing products, completing online forms, and gathering web content.

Unlike server-based AI integrations which require sensitive data to be sent to third-party servers, Browser Operator processes tasks locally within the Opera browser.

Opera’s demonstration video showcases how Browser Operator can streamline an everyday task like buying socks. Instead of manually scrolling through product pages or filling out payment forms, users could delegate the entire process to Browser Operator—allowing them to shift focus to activities that matter more to them, such as spending time with loved ones.

Harnessing natural language processing powered by Opera’s AI Composer Engine, Browser Operator interprets written instructions from users and executes corresponding tasks within the browser. All operations occur locally on a user’s device, leveraging the browser’s own infrastructure to safely and swiftly complete commands.  

If Browser Operator encounters a sensitive step in the process, such as entering payment details or approving an order, it pauses and requests the user’s input. You also have the freedom to intervene and take control of the process at any time.  

Every step Browser Operator takes is transparent and fully reviewable, providing users a clear understanding of how tasks are being executed. If mistakes occur – like placing an incorrect order – you can further instruct the AI agent to make amends, such as cancelling the order or adjusting a form.

The key differentiators: Privacy, performance, and precision  

What sets Browser Operator apart from other AI-integrated tools is its localised, privacy-first architecture. Unlike competitors that depend on screenshots or video recordings to understand webpage content, Opera’s approach uses the Document Object Model (DOM) Tree and browser layout data—a textual representation of the webpage.  

This difference offers several key advantages:

  • Faster task completion: Browser Operator doesn’t need to “see” and interpret pixels on the screen or emulate mouse movements. Instead, it accesses web page elements directly, avoiding unnecessary overhead and allowing it to process pages holistically without scrolling.
  • Enhanced privacy: With all operations conducted on the browser itself, user data – including logins, cookies, and browsing history – remains secure on the local device. No screenshots, keystrokes, or personal information are sent to Opera’s servers.
  • Easier interaction with page elements: The AI can engage with elements hidden from the user’s view, such as behind cookie popups or verification dialogs, enabling seamless access to web page content.

By enabling the browser to autonomously perform tasks, Opera is taking a significant step forward in making browsers “agentic”—not just tools for accessing the internet, but assistants that actively enhance productivity.  

See also: You.com ARI: Professional-grade AI research agent for businesses

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Opera introduces browser-integrated AI agent appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/opera-introduces-browser-integrated-ai-agent/feed/ 0
Fetch.ai launches first Web3 agentic AI model https://www.artificialintelligence-news.com/news/fetch-ai-launches-first-web3-agentic-ai-model/ https://www.artificialintelligence-news.com/news/fetch-ai-launches-first-web3-agentic-ai-model/#respond Tue, 25 Feb 2025 16:50:45 +0000 https://www.artificialintelligence-news.com/?p=104610 Fetch.ai has launched ASI-1 Mini, a native Web3 large language model designed to support complex agentic AI workflows. Described as a gamechanger for AI accessibility and performance, ASI-1 Mini is heralded for delivering results on par with leading LLMs but at significantly reduced hardware costs—a leap forward in making AI enterprise-ready. ASI-1 Mini integrates into […]

The post Fetch.ai launches first Web3 agentic AI model appeared first on AI News.

]]>
Fetch.ai has launched ASI-1 Mini, a native Web3 large language model designed to support complex agentic AI workflows.

Described as a gamechanger for AI accessibility and performance, ASI-1 Mini is heralded for delivering results on par with leading LLMs but at significantly reduced hardware costs—a leap forward in making AI enterprise-ready.

ASI-1 Mini integrates into Web3 ecosystems, enabling secure and autonomous AI interactions. Its release sets the foundation for broader innovation within the AI sector—including the imminent launch of the Cortex suite, which will further enhance the use of large language models and generalised intelligence.

“This launch marks the beginning of ASI-1 Mini’s rollout and a new era of community-owned AI. By decentralising AI’s value chain, we’re empowering the Web3 community to invest in, train, and own foundational AI models,” said Humayun Sheikh, CEO of Fetch.ai and Chairman of the Artificial Superintelligence Alliance.

“We’ll soon introduce advanced agentic tool integration, multi-modal capabilities, and deeper Web3 synergy to enhance ASI-1 Mini’s automation capabilities while keeping AI’s value creation in the hands of its contributors.”

Democratising AI with Web3: Decentralised ownership and shared value  

Key to Fetch.ai’s vision is the democratisation of foundational AI models, allowing the Web3 community to not just use, but also train and own proprietary LLMs like ASI-1 Mini. 

This decentralisation unlocks opportunities for individuals to directly benefit from the economic growth of cutting-edge AI models, which could achieve multi-billion-dollar valuations.  

Through Fetch.ai’s platform, users can invest in curated AI model collections, contribute to their development, and share in generated revenues. For the first time, decentralisation is driving AI model ownership—ensuring financial benefits are more equitably distributed.

Advanced reasoning and tailored performance  

ASI-1 Mini introduces adaptability in decision-making with four dynamic reasoning modes: Multi-Step, Complete, Optimised, and Short Reasoning. This flexibility allows it to balance depth and precision based on the specific task at hand.  

Whether performing intricate, multi-layered problem-solving or delivering concise, actionable insights, ASI-1 Mini adapts dynamically for maximum efficiency. Its Mixture of Models (MoM) and Mixture of Agents (MoA) frameworks further enhance this versatility.  

Mixture of Models (MoM):  

ASI-1 Mini selects relevant models dynamically from a suite of specialised AI models, which are optimised for specific tasks or datasets. This ensures high efficiency and scalability, especially for multi-modal AI and federated learning.  

Mixture of Agents (MoA):  

Independent agents with unique knowledge and reasoning capabilities work collaboratively to solve complex tasks. The system’s coordination mechanism ensures efficient task distribution, paving the way for decentralised AI models that thrive in dynamic, multi-agent systems.  

This sophisticated architecture is built on three interacting layers:  

  1. Foundational layer: ASI-1 Mini serves as the core intelligence and orchestration hub.  
  2. Specialisation layer (MoM Marketplace): Houses diverse expert models, accessible through the ASI platform.  
  3. Action layer (AgentVerse): Features agents capable of managing live databases, integrating APIs, facilitating decentralised workflows, and more.  

By selectively activating only necessary models and agents, the system ensures performance, precision, and scalability in real-time tasks.  

Transforming AI efficiency and accessibility

Unlike traditional LLMs, which come with high computational overheads, ASI-1 Mini is optimised for enterprise-grade performance on just two GPUs, reducing hardware costs by a remarkable eightfold. For businesses, this means reduced infrastructure costs and increased scalability, breaking down financial barriers to high-performance AI integration.  

On benchmark tests like Massive Multitask Language Understanding (MMLU), ASI-1 Mini matches or surpasses leading LLMs in specialised domains such as medicine, history, business, and logical reasoning.  

Rolling out in two phases, ASI-1 Mini will soon process vastly larger datasets with upcoming context window expansions:  

  • Up to 1 million tokens: Allows the model to analyse complex documents or technical manuals.
  • Up to 10 million tokens: Enables high-stakes applications like legal record review, financial analysis, and enterprise-scale datasets.  

These enhancements will make ASI-1 Mini invaluable for complex and multi-layered tasks.  

Tackling the “black-box” problem  

The AI industry has long faced the challenge of addressing the black-box problem, where deep learning models reach conclusions without clear explanations.

ASI-1 Mini mitigates this issue with continuous multi-step reasoning, facilitating real-time corrections and optimised decision-making. While it doesn’t entirely eliminate opacity, ASI-1 provides more explainable outputs—critical for industries like healthcare and finance.  

Its multi-expert model architecture not only ensures transparency but also optimises complex workflows across diverse sectors. From managing databases to executing real-time business logic, ASI-1 outperforms traditional models in both speed and reliability.  

AgentVerse integration: Building the agentic AI economy

ASI-1 Mini is set to connect with AgentVerse, Fetch.ai’s agent marketplace, providing users with the tools to build and deploy autonomous agents capable of real-world task execution via simple language commands. For example, users could automate trip planning, restaurant reservations, or financial transactions through “micro-agents” hosted on the platform.

This ecosystem enables open-source AI customisation and monetisation, creating an “agentic economy” where developers and businesses thrive symbiotically. Developers can monetise micro-agents, while users gain seamless access to tailored AI solutions.  

As its agentic ecosystem matures, ASI-1 Mini aims to evolve into a multi-modal powerhouse capable of processing structured text, images, and complex datasets with context-aware decision-making.  

See also: Endor Labs: AI transparency vs ‘open-washing’

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Fetch.ai launches first Web3 agentic AI model appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/fetch-ai-launches-first-web3-agentic-ai-model/feed/ 0
ChatGPT gains agentic capability for complex research https://www.artificialintelligence-news.com/news/chatgpt-gains-agentic-capability-for-complex-research/ https://www.artificialintelligence-news.com/news/chatgpt-gains-agentic-capability-for-complex-research/#respond Mon, 03 Feb 2025 17:22:06 +0000 https://www.artificialintelligence-news.com/?p=104108 OpenAI is releasing a powerful agentic capability that enables ChatGPT to conduct complex, multi-step research tasks online. The feature, called Deep Research, reportedly achieves in tens of minutes what could take a human researcher hours or even days. OpenAI describes Deep Research as a significant milestone in its journey toward artificial general intelligence (AGI). “The […]

The post ChatGPT gains agentic capability for complex research appeared first on AI News.

]]>
OpenAI is releasing a powerful agentic capability that enables ChatGPT to conduct complex, multi-step research tasks online. The feature, called Deep Research, reportedly achieves in tens of minutes what could take a human researcher hours or even days.

OpenAI describes Deep Research as a significant milestone in its journey toward artificial general intelligence (AGI).

“The ability to synthesise knowledge is a prerequisite for creating new knowledge,” says OpenAI. “For this reason, Deep Research marks a significant step toward our broader goal of developing AGI.”

Agentic AI enables ChatGPT to assist with complex research

Deep Research empowers ChatGPT to find, analyse, and synthesise information from hundreds of online sources autonomously. With just a prompt from the user, the tool can deliver a comprehensive report, comparable to the output of a research analyst, according to OpenAI.

Drawing capabilities from a variant of OpenAI’s upcoming “o3” model, the aim is to free users from time-consuming, labour-intensive information gathering. Whether it’s a competitive analysis of streaming platforms, an informed policy review, or even personalised recommendations for a new commuter bike, Deep Research promises precise and reliable results.

Importantly, every output includes full citations and transparent documentation—enabling users to verify the findings with ease.

The tool appears particularly adept at uncovering niche or non-intuitive insights, making it an invaluable asset across industries like finance, science, policymaking, and engineering. But OpenAI also envisions Deep Research being useful for the average user, such as shoppers looking for hyper-personalised recommendations or a specific product.

This latest agentic capability operates through the user interface of ChatGPT; users simply select the “Deep Research” option in the message composer and type their query. Supporting files or spreadsheets can also be uploaded for additional context.

Once initiated, the AI embarks on a rigorous multi-step process, which may take 5-30 minutes to complete. A sidebar provides updates on the actions taken and the sources consulted. Users can carry on with other tasks and will be notified when the final report is ready. 

The results are presented in the chat as detailed, well-documented reports. In the coming weeks, OpenAI plans to enhance these outputs further by embedding images, data visualisations, and graphs to deliver even greater clarity and context.

Unlike GPT-4o – which excels in real-time, multimodal conversations – Deep Research prioritises depth and detail. Its ability to rigorously cite sources and provide comprehensive analysis sets it apart—shifting the focus from fast, summarised answers to well-documented, research-grade insights.

Built for real-world challenges

Deep Rsearch leverages sophisticated training methodologies, grounded in real-world browsing and reasoning tasks across diverse domains. Its model was trained via reinforcement learning to autonomously plan and execute multi-step research processes, including backtracking and adaptively refining its approach as new information becomes available. 

The tool can browse user-uploaded files, generate and iterate on graphs using Python, embed media such as generated images and web pages into responses, and cite exact sentences or passages from its sources. The result of this extensive training is a highly capable agent for tackling complex real-world problems.

OpenAI evaluated Deep Research across a broad set of expert-level exams known as “Humanity’s Last Exam”. The exams – comprising over 3,000 questions covering topics from rocket science and linguistics to ecology and classics – test an AI’s competence in solving multifaceted problems.

The results were impressive, with the model achieving a record-breaking 26.6% accuracy across these domains:

  • GPT-4o: 3.3%
  • Grok-2: 3.8%
  • Claude 3.5 Sonnet: 4.3%
  • OpenAI o1: 9.1%
  • DeepSeek-R1: 9.4%
  • Deep research: 26.6% (with browsing + Python tools)

Deep Research also reached a new state-of-the-art performance on the GAIA benchmark, which evaluates AI models on real-world questions requiring reasoning, multi-modal fluency, and tool-use proficiency. Deep Research topped the leaderboard with a score of 72.57%.

Limitations and challenges

While the Deep Research agentic AI capability in ChatGPT signifies a bold step forward, OpenAI acknowledges that the technology is still in its early stages and comes with limitations.

The system occasionally “hallucinates” facts or offers incorrect inferences, albeit at a notably reduced rate compared to existing GPT models, according to OpenAI. It also faces challenges in differentiating between authoritative sources and speculative content, and it struggles to calibrate its confidence levels—often displaying undue certainty for potentially uncertain findings.

Minor formatting errors in reports and citations, as well as delays in initiating tasks, could also frustrate initial users. OpenAI says these issues are expected to improve over time with more usage and iterative refinements.

OpenAI is rolling out the capability gradually, starting with Pro users, who will have access to up to 100 queries per month. Plus and Team tiers will follow suit, with Enterprise access arriving next. 

UK, Swiss, and European Economic Area residents are not yet able to access the feature, but OpenAI says it’s working on expanding its rollout to these regions.

In the weeks ahead, OpenAI will expand the feature to ChatGPT’s mobile and desktop platforms. The long-term vision includes enabling connections to subscription-based or proprietary data sources, further enhancing the robustness and personalisation of its outputs.

Looking further ahead, OpenAI envisions integrating Deep Research with “Operator,” an existing chatbot capability that takes real-world actions. This integration would allow ChatGPT to seamlessly handle tasks that require both asynchronous online research and real-world execution.

(Photo by John Schnobrich)

See also: Microsoft and OpenAI probe alleged data theft by DeepSeek

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post ChatGPT gains agentic capability for complex research appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/chatgpt-gains-agentic-capability-for-complex-research/feed/ 0
Yiannis Antoniou, Lab49: OpenAI Operator kickstarts era of browser AI agents https://www.artificialintelligence-news.com/news/yiannis-antoniou-lab49-openai-operator-era-browser-ai-agents/ https://www.artificialintelligence-news.com/news/yiannis-antoniou-lab49-openai-operator-era-browser-ai-agents/#respond Fri, 24 Jan 2025 14:03:14 +0000 https://www.artificialintelligence-news.com/?p=16963 OpenAI has unveiled Operator, a tool that integrates seamlessly with web browsers to perform tasks autonomously. From filling out forms to ordering groceries, Operator promises to simplify repetitive online activities by interacting directly with websites through clicks, typing, and scrolling. Designed around a new model called the Computer-Using Agent (CUA), Operator combines GPT-4o’s vision recognition […]

The post Yiannis Antoniou, Lab49: OpenAI Operator kickstarts era of browser AI agents appeared first on AI News.

]]>
OpenAI has unveiled Operator, a tool that integrates seamlessly with web browsers to perform tasks autonomously. From filling out forms to ordering groceries, Operator promises to simplify repetitive online activities by interacting directly with websites through clicks, typing, and scrolling.

Designed around a new model called the Computer-Using Agent (CUA), Operator combines GPT-4o’s vision recognition with advanced reasoning capabilities—allowing it to function as a virtual “human-in-the-browser.” Yet, for all its innovation, industry experts see room for refinement.

Yiannis Antoniou, Head of AI, Data, and Analytics at specialist consultancy Lab49, shared his insights on Operator’s significance and positioning in the competitive landscape of agent AI systems.

Agentic AI through a familiar interface

“OpenAI’s announcement of Operator, its latest foray into the agentic AI wars, is both fascinating and incomplete,” said Antoniou, who has over two decades of experience designing AI systems for financial services firms.

Headshot of Yiannis Antoniou, Head of AI, Data, and Analytics at specialist consultancy Lab49, for an article on how OpenAI operator is kickstarting the era of browser AI agents.

“Clearly influenced by Anthropic Claude’s Computer Use system, introduced back in October, Operator streamlines the experience by removing the need for complex infrastructure and focusing on a familiar interface: the browser.”

By designing Operator to operate within an environment users already understand, the web browser, OpenAI sidesteps the need for bespoke APIs or integrations.

“By leveraging the world’s most popular interface, OpenAI enhances the user experience and captures immediate interest from the general public. This browser-centric approach creates significant potential for widespread adoption, something Anthropic – despite its early-mover advantage – has struggled to achieve.”

Unlike some competing systems that may feel technical or niche in their application, Operator’s browser-focused framework lowers the barrier to entry and is a step forward in OpenAI’s efforts to democratise AI.

Unique take on usability and security

One of the hallmarks of Operator is its emphasis on adaptability and security, implemented through human-in-the-loop protocols. Antoniou acknowledged these thoughtful usability features but noted that more work is needed.

“Architecturally, Operator’s browser integration closely mirrors Claude’s system. Both involve taking screenshots of the user’s browser and sending them for analysis, as well as controlling the screen via virtual keystrokes and mouse movements. However, Operator introduces thoughtful usability touches. 

“Features like custom instructions for specific websites add a layer of personalisation, and the emphasis on human-in-the-loop safeguards against unauthorised actions – such as purchases, sending emails, or applying for jobs – demonstrate OpenAI’s awareness of potential security risks posed by malicious websites, but more work is clearly needed to make this system widely safe across a variety of scenarios.”

OpenAI has implemented a multi-layered safety framework for Operator, including takeover mode for secure inputs, user confirmations prior to significant actions, and monitoring systems to detect adversarial behavior. Furthermore, users can delete browsing data and manage privacy settings directly within the tool.

However, Antoniou emphasised that these measures are still evolving—particularly as Operator encounters complex or sensitive tasks. 

OpenAI Operator further democratises AI

Antoniou also sees the release of Operator as a pivotal moment for the consumer AI landscape, albeit one that is still in its early stages. 

“Overall, this is an excellent first attempt at building an agentic system for everyday users, designed around how they naturally interact with technology. As the system develops – with added capabilities and more robust security controls – this limited rollout, priced at $200/month, will serve as a testing ground. 

“Once matured and extended to lower subscription tiers and the free version, Operator has the potential to usher in the era of consumer-facing agents, further democratising AI and embedding it into daily life.”

Designed initially for Pro users at a premium price point, Operator provides OpenAI with an opportunity to learn from early adopters and refine its capabilities.

Antoniou noted that while $200/month might not yet justify the system’s value for most users, investment in making Operator more powerful and accessible could lead to significant competitive advantages for OpenAI in the long run.

“Is it worth $200/month? Perhaps not yet. But as the system evolves, OpenAI’s moat will grow, making it harder for competitors to catch up. Now, the challenge shifts back to Anthropic and Google – both of whom have demonstrated similar capabilities in niche or engineering-focused products – to respond and stay in the game,” he concludes.

As OpenAI continues to fine-tune Operator, the potential to revolutionise how people interact with technology becomes apparent. From collaborations with companies like Instacart, DoorDash, and Uber to use cases in the public sector, Operator aims to balance innovation with trust and safety.

While early limitations and pricing may deter widespread adoption for now, these hurdles might only be temporary as OpenAI commits to enhancing usability and accessibility over time.

See also: OpenAI argues against ChatGPT data deletion in Indian court

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Yiannis Antoniou, Lab49: OpenAI Operator kickstarts era of browser AI agents appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/yiannis-antoniou-lab49-openai-operator-era-browser-ai-agents/feed/ 0
NVIDIA advances AI frontiers with CES 2025 announcements https://www.artificialintelligence-news.com/news/nvidia-advances-ai-frontiers-with-ces-2025-announcements/ https://www.artificialintelligence-news.com/news/nvidia-advances-ai-frontiers-with-ces-2025-announcements/#respond Tue, 07 Jan 2025 11:25:09 +0000 https://www.artificialintelligence-news.com/?p=16818 NVIDIA CEO and founder Jensen Huang took the stage for a keynote at CES 2025 to outline the company’s vision for the future of AI in gaming, autonomous vehicles (AVs), robotics, and more. “AI has been advancing at an incredible pace,” Huang said. “It started with perception AI — understanding images, words, and sounds. Then […]

The post NVIDIA advances AI frontiers with CES 2025 announcements appeared first on AI News.

]]>
NVIDIA CEO and founder Jensen Huang took the stage for a keynote at CES 2025 to outline the company’s vision for the future of AI in gaming, autonomous vehicles (AVs), robotics, and more.

“AI has been advancing at an incredible pace,” Huang said. “It started with perception AI — understanding images, words, and sounds. Then generative AI — creating text, images, and sound. Now, we’re entering the era of ‘physical AI,’ AI that can perceive, reason, plan, and act.”

With NVIDIA’s platforms and GPUs at the core, Huang explained how the company continues to fuel breakthroughs across multiple industries while unveiling innovations such as the Cosmos platform, next-gen GeForce RTX 50 Series GPUs, and compact AI supercomputer Project DIGITS. 

RTX 50 series: “The GPU is a beast”

One of the most significant announcements during CES 2025 was the introduction of the GeForce RTX 50 Series, powered by NVIDIA Blackwell architecture. Huang debuted the flagship RTX 5090 GPU, boasting 92 billion transistors and achieving an impressive 3,352 trillion AI operations per second (TOPS).

“GeForce enabled AI to reach the masses, and now AI is coming home to GeForce,” said Huang.

Holding the blacked-out GPU, Huang called it “a beast,” highlighting its advanced features, including dual cooling fans and its ability to leverage AI for revolutionary real-time graphics.

Set for a staggered release in early 2025, the RTX 50 Series includes the flagship RTX 5090 and RTX 5080 (available 30 January), followed by the RTX 5070 Ti and RTX 5070 (February). Laptop GPUs join the lineup in March.

In addition, NVIDIA introduced DLSS 4 – featuring ‘Multi-Frame Generation’ technology – which boosts gaming performance up to eightfold by generating three additional frames for every frame rendered.

Other advancements, such as RTX Neural Shaders and RTX Mega Geometry, promise heightened realism in video games, including precise face and hair rendering using generative AI.

Cosmos: Ushering in physical AI

NVIDIA took another step forward with the Cosmos platform at CES 2025, which Huang described as a “game-changer” for robotics, industrial AI, and AVs. Much like the impact of large language models on generative AI, Cosmos represents a new frontier for AI applications in robotics and autonomous systems.

“The ChatGPT moment for general robotics is just around the corner,” Huang declared.

Cosmos integrates generative models, tokenisers, and video processing frameworks to enable robots and vehicles to simulate potential outcomes and predict optimal actions. By ingesting text, image, and video prompts, Cosmos can generate “virtual world states,” tailored for complex robotics and AV use cases involving real-world environments and lighting.

Top robotics and automotive leaders – including XPENG, Hyundai Motor Group, and Uber – are among the first to adopt Cosmos, which is available on GitHub via an open licence.

Pras Velagapudi, CTO at Agility, comments: “Data scarcity and variability are key challenges to successful learning in robot environments. Cosmos’ text-, image- and video-to-world capabilities allow us to generate and augment photorealistic scenarios for a variety of tasks that we can use to train models without needing as much expensive, real-world data capture.”

Empowering developers with AI models

NVIDIA also unveiled new AI foundation models for RTX PCs, which aim to supercharge content creation, productivity, and enterprise applications. These models, presented as NVIDIA NIM (Neural Interaction Model) microservices, are designed to integrate with the RTX 50 Series hardware.

Huang emphasised the accessibility of these tools: “These AI models run in every single cloud because NVIDIA GPUs are now available in every cloud.”

NVIDIA is doubling down on its push to equip developers with advanced tools for building AI-driven solutions. The company introduced AI Blueprints: pre-configured tools for crafting agents tailored to specific enterprise needs, such as content generation, fraud detection, and video management.

“They are completely open source, so you could take it and modify the blueprints,” explains Huang.

Huang also announced the release of Llama Nemotron, designed for developers to build and deploy powerful AI agents.

Ahmad Al-Dahle, VP and Head of GenAI at Meta, said: “Agentic AI is the next frontier of AI development, and delivering on this opportunity requires full-stack optimisation across a system of LLMs to deliver efficient, accurate AI agents.

“Through our collaboration with NVIDIA and our shared commitment to open models, the NVIDIA Llama Nemotron family built on Llama can help enterprises quickly create their own custom AI agents.”

Philipp Herzig, Chief AI Officer at SAP, added: “AI agents that collaborate to solve complex tasks across multiple lines of the business will unlock a whole new level of enterprise productivity beyond today’s generative AI scenarios.

“Through SAP’s Joule, hundreds of millions of enterprise users will interact with these agents to accomplish their goals faster than ever before. NVIDIA’s new open Llama Nemotron model family will foster the development of multiple specialised AI agents to transform business processes.”

Safer and smarter autonomous vehicles

NVIDIA’s announcements extended to the automotive industry, where its DRIVE Hyperion AV platform is fostering a safer and smarter future for AVs. Built on the new NVIDIA AGX Thor system-on-a-chip (SoC), the platform allows vehicles to achieve next-level functional safety and autonomous capabilities using generative AI models.

“The autonomous vehicle revolution is here,” Huang said. “Building autonomous vehicles, like all robots, requires three computers: NVIDIA DGX to train AI models, Omniverse to test-drive and generate synthetic data, and DRIVE AGX, a supercomputer in the car.”

Huang explained that synthetic data is critical for AV development, as it dramatically enhances real-world datasets. NVIDIA’s AI data factories – powered by Omniverse and Cosmos platforms – generate synthetic driving scenarios, increasing the effectiveness of training data exponentially.

Toyota, the world’s largest automaker, is committed to using NVIDIA DRIVE AGX Orin and the safety-certified NVIDIA DriveOS to develop its next-generation vehicles. Heavyweights such as JLR, Mercedes-Benz, and Volvo Cars have also adopted DRIVE Hyperion.

Project DIGITS: Compact AI supercomputer

Huang concluded his NVIDIA keynote at CES 2025 with a final “one more thing” announcement: Project DIGITS, NVIDIA’s smallest yet most powerful AI supercomputer, powered by the cutting-edge GB10 Grace Blackwell Superchip.

“This is NVIDIA’s latest AI supercomputer,” Huang declared, revealing its compact size, claiming it’s portable enough to “practically fit in a pocket.”

Project DIGITS enables developers and engineers to train and deploy AI models directly from their desks, providing the full power of NVIDIA’s AI stack in a compact form.

Image of Project DIGITS on a desk, a compact AI supercomputer by NVIDIA debuted at CES 2025.

Set to launch in May, Project DIGITS represents NVIDIA’s push to make AI supercomputing accessible to individuals as well as organisations.

Vision for tomorrow

Reflecting on NVIDIA’s journey since inventing the programmable GPU in 1999, Huang described the past 12 years of AI-driven change as transformative.

“Every single layer of the technology stack has been fundamentally transformed,” he said.

With advancements spanning gaming, AI-driven agents, robotics, and autonomous vehicles, Huang foresees an exciting future.

“All of the enabling technologies I’ve talked about today will lead to surprising breakthroughs in general robotics and AI over the coming years,” Huang concludes.

(Image Credit: NVIDIA)

See also: Sam Altman, OpenAI: ‘Lucky and humbling’ to work towards superintelligence

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post NVIDIA advances AI frontiers with CES 2025 announcements appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/nvidia-advances-ai-frontiers-with-ces-2025-announcements/feed/ 0
Gemini 2.0: Google ushers in the agentic AI era  https://www.artificialintelligence-news.com/news/gemini-2-0-google-ushers-in-agentic-ai-era/ https://www.artificialintelligence-news.com/news/gemini-2-0-google-ushers-in-agentic-ai-era/#respond Wed, 11 Dec 2024 16:52:09 +0000 https://www.artificialintelligence-news.com/?p=16694 Google CEO Sundar Pichai has announced the launch of Gemini 2.0, a model that represents the next step in Google’s ambition to revolutionise AI. A year after introducing the Gemini 1.0 model, this major upgrade incorporates enhanced multimodal capabilities, agentic functionality, and innovative user tools designed to push boundaries in AI-driven technology. Leap towards transformational […]

The post Gemini 2.0: Google ushers in the agentic AI era  appeared first on AI News.

]]>
Google CEO Sundar Pichai has announced the launch of Gemini 2.0, a model that represents the next step in Google’s ambition to revolutionise AI.

A year after introducing the Gemini 1.0 model, this major upgrade incorporates enhanced multimodal capabilities, agentic functionality, and innovative user tools designed to push boundaries in AI-driven technology.

Leap towards transformational AI  

Reflecting on Google’s 26-year mission to organise and make the world’s information accessible, Pichai remarked, “If Gemini 1.0 was about organising and understanding information, Gemini 2.0 is about making it much more useful.”

Gemini 1.0, released in December 2022, was notable for being Google’s first natively multimodal AI model. The first iteration excelled at understanding and processing text, video, images, audio, and code. Its enhanced 1.5 version became widely embraced by developers for its long-context understanding, enabling applications such as the productivity-focused NotebookLM.

Now, with Gemini 2.0, Google aims to accelerate the role of AI as a universal assistant capable of native image and audio generation, better reasoning and planning, and real-world decision-making capabilities. In Pichai’s words, the development represents the dawn of an “agentic era.”

“We have been investing in developing more agentic models, meaning they can understand more about the world around you, think multiple steps ahead, and take action on your behalf, with your supervision,” Pichai explained.

Gemini 2.0: Core features and availability

At the heart of today’s announcement is the experimental release of Gemini 2.0 Flash, the flagship model of Gemini’s second generation. It builds upon the foundations laid by its predecessors while delivering faster response times and advanced performance.

Gemini 2.0 Flash supports multimodal inputs and outputs, including the ability to generate native images in conjunction with text and produce steerable text-to-speech multilingual audio. Additionally, users can benefit from native tool integration such as Google Search and even third-party user-defined functions.

Developers and businesses will gain access to Gemini 2.0 Flash via the Gemini API in Google AI Studio and Vertex AI, while larger model sizes are scheduled for broader release in January 2024.

For global accessibility, the Gemini app now features a chat-optimised version of the 2.0 Flash experimental model. Early adopters can experience this updated assistant on desktop and mobile, with a mobile app rollout imminent.

Products such as Google Search are also being enhanced with Gemini 2.0, unlocking the ability to handle complex queries like advanced math problems, coding enquiries, and multimodal questions.

Comprehensive suite of AI innovations  

The launch of Gemini 2.0 comes with compelling new tools that showcase its capabilities.

One such feature, Deep Research, functions as an AI research assistant, simplifying the process of investigating complex topics by compiling information into comprehensive reports. Another upgrade enhances Search with Gemini-enabled AI Overviews that tackle intricate, multi-step user queries.

The model was trained using Google’s sixth-generation Tensor Processing Units (TPUs), known as Trillium, which Pichai notes “powered 100% of Gemini 2.0 training and inference.”

Trillium is now available for external developers, allowing them to benefit from the same infrastructure that supports Google’s own advancements.

Pioneering agentic experiences  

Accompanying Gemini 2.0 are experimental “agentic” prototypes built to explore the future of human-AI collaboration, including:

  • Project Astra: A universal AI assistant

First introduced at I/O earlier this year, Project Astra taps into Gemini 2.0’s multimodal understanding to improve real-world AI interactions. Trusted testers have trialled the assistant on Android, offering feedback that has helped refine its multilingual dialogue, memory retention, and integration with Google tools like Search, Lens, and Maps. Astra has also demonstrated near-human conversational latency, with further research underway for its application in wearable technology, such as prototype AI glasses.

  • Project Mariner: Redefining web automation 

Project Mariner is an experimental web-browsing assistant that uses Gemini 2.0’s ability to reason across text, images, and interactive elements like forms within a browser. In initial tests, it achieved an 83.5% success rate on the WebVoyager benchmark for completing end-to-end web tasks. Early testers using a Chrome extension are helping to refine Mariner’s capabilities while Google evaluates safety measures that ensure the technology remains user-friendly and secure.

  • Jules: A coding agent for developers  

Jules, an AI-powered assistant built for developers, integrates directly into GitHub workflows to address coding challenges. It can autonomously propose solutions, generate plans, and execute code-based tasks—all under human supervision. This experimental endeavour is part of Google’s long-term goal to create versatile AI agents across various domains.

  • Gaming applications and beyond  

Extending Gemini 2.0’s reach into virtual environments, Google DeepMind is working with gaming partners like Supercell on intelligent game agents. These experimental AI companions can interpret game actions in real-time, suggest strategies, and even access broader knowledge via Search. Research is also being conducted into how Gemini 2.0’s spatial reasoning could support robotics, opening doors for physical-world applications in the future.

Addressing responsibility in AI development

As AI capabilities expand, Google emphasises the importance of prioritising safety and ethical considerations.

Google claims Gemini 2.0 underwent extensive risk assessments, bolstered by the Responsibility and Safety Committee’s oversight to mitigate potential risks. Additionally, its embedded reasoning abilities allow for advanced “red-teaming,” enabling developers to evaluate security scenarios and optimise safety measures at scale.

Google is also exploring safeguards to address user privacy, prevent misuse, and ensure AI agents remain reliable. For instance, Project Mariner is designed to prioritise user instructions while resisting malicious prompt injections, preventing threats like phishing or fraudulent transactions. Meanwhile, privacy controls in Project Astra make it easy for users to manage session data and deletion preferences.

Pichai reaffirmed the company’s commitment to responsible development, stating, “We firmly believe that the only way to build AI is to be responsible from the start.”

With the Gemini 2.0 Flash release, Google is edging closer to its vision of building a universal assistant capable of transforming interactions across domains.

See also: Machine unlearning: Researchers make AI models ‘forget’ data

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Gemini 2.0: Google ushers in the agentic AI era  appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/gemini-2-0-google-ushers-in-agentic-ai-era/feed/ 0
Salesforce: UK set to lead agentic AI revolution https://www.artificialintelligence-news.com/news/salesforce-uk-set-lead-agentic-ai-revolution/ https://www.artificialintelligence-news.com/news/salesforce-uk-set-lead-agentic-ai-revolution/#respond Mon, 02 Dec 2024 13:24:31 +0000 https://www.artificialintelligence-news.com/?p=16601 Salesforce has unveiled the findings of its UK AI Readiness Index, signalling the nation is in a position to spearhead the next wave of AI innovation, also known as agentic AI. The report places the UK ahead of its G7 counterparts in terms of AI adoption but also underscores areas ripe for improvement, such as […]

The post Salesforce: UK set to lead agentic AI revolution appeared first on AI News.

]]>
Salesforce has unveiled the findings of its UK AI Readiness Index, signalling the nation is in a position to spearhead the next wave of AI innovation, also known as agentic AI.

The report places the UK ahead of its G7 counterparts in terms of AI adoption but also underscores areas ripe for improvement, such as support for SMEs, fostering cross-sector partnerships, and investing in talent development.

Zahra Bahrololoumi CBE, UKI CEO at Salesforce, commented: “Agentic AI is revolutionising enterprise software by enabling humans and agents to collaborate seamlessly and drive customer success.

“The UK AI Readiness Index positively highlights that the UK has both the vision and infrastructure to be a powerhouse globally in AI, and lead the current third wave of agentic AI.”

UK AI adoption sets the stage for agentic revolution

The Index details how both the public and private sectors in the UK have embraced AI’s transformative potential. With a readiness score of 65.5, surpassing the G7 average of 61.2, the UK is establishing itself as a hub for large-scale AI projects, driven by a robust innovation culture and pragmatic regulatory approaches.

The government has played its part in maintaining a stable and secure environment for tech investment. Initiatives such as the AI Safety Summit at Bletchley Park and risk-oriented AI legislation showcase Britain’s leadership on critical AI issues like transparency and privacy.

Business readiness is equally impressive, with UK industries scoring 52, well above the G7 average of 47.8. SMEs in the UK are increasingly prioritising AI adoption, further bolstering the nation’s stance in the international AI arena.

Adam Evans, EVP & GM of Salesforce AI Platform, is optimistic about the evolution of agentic AI. Evans foresees that, by 2025, these agents will become business-aware—expertly navigating industry-specific challenges to execute meaningful tasks and decisions.

Investments fuelling AI growth

Salesforce is committing $4 billion to the UK’s AI ecosystem over the next five years. Since establishing its UK AI Centre in London, Salesforce says it has engaged over 3,000 stakeholders in AI training and workshops.

Key investment focuses include creating a regulatory bridge between the EU’s rules-based approach and the more relaxed US approach, and ensuring SMEs have the resources to integrate AI. A strong emphasis also lies on enhancing digital skills and centralising training to support the AI workforce of the future.

Feryal Clark, Minister for AI and Digital Government, said: “These findings are further proof the UK is in prime position to take advantage of AI, and highlight our strength in spurring innovation, investment, and collaboration across the public and private sector.

“There is a global race for AI and we’ll be setting out plans for how the UK can use the technology to ramp-up adoption across the economy, kickstart growth, and build an AI sector which can scale and compete on the global stage.”

Antony Walker, Deputy CEO at techUK, added: “To build this progress, government and industry must collaborate to foster innovation, support SMEs, invest in skills, and ensure flexible regulation, cementing the UK’s leadership in the global AI economy.”

Agentic AI boosting UK business productivity 

Capita, Secret Escapes, Heathrow, and Bionic are among the organisations that have adopted Salesforce’s Agentforce to boost their productivity.

Adolfo Hernandez, CEO of Capita, said: “We want to transform Capita’s recruitment process into a fast, seamless and autonomous experience that benefits candidates, our people, and our clients.

“With autonomous agents providing 24/7 support, our goal is to enable candidates to complete the entire recruitment journey within days as opposed to what has historically taken weeks.

Secret Escapes, a curator of luxury travel deals, finds autonomous agents crucial for personalising services to its 60 million European members.

Kate Donaghy, Head of Business Technology at Secret Escapes, added: “Agentforce uses our unified data to automate routine tasks like processing cancellations, updating booking information, or even answering common travel questions about luggage, flight information, and much more—freeing up our customer service agents to handle more complex and last-minute travel needs to better serve our members.”

The UK’s AI readiness is testament to the synergy between government, business, and academia. To maintain its leadership, the UK must sustain its focus on collaboration, skills development, and innovation. 

(Photo by Matthew Wiebe)

See also: Generative AI use soars among Brits, but is it sustainable?

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Salesforce: UK set to lead agentic AI revolution appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/salesforce-uk-set-lead-agentic-ai-revolution/feed/ 0