Stay Ahead, Stay ONMINE

What’s next for agentic AI? LangChain founder looks to ambient agents

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Agentic AI is the latest big trend in generative AI, but what comes after that? While full artificial general intelligence (AGI) is likely still some time in the future, there might well be an intermediate step […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Agentic AI is the latest big trend in generative AI, but what comes after that?

While full artificial general intelligence (AGI) is likely still some time in the future, there might well be an intermediate step with an approach known as ambient agents.

LangChain, the agentic AI pioneer, introduced the term “ambient agents” on January 14. The technology that LangChain develops includes its eponymous open source LangChain framework that enables organizations to chain different large language models (LLMs) together to get a result. LangChain Inc. raised $24 million in funding in February 2024. The company also has a series of commercial products including LangSmith for LLM Ops.

With a traditional AI interface, users typically interact with an LLM via text prompts to initiate an action. Agentic AI generally refers to LLM-powered systems that take actions on the user’s behalf. The concept of ambient agents takes that paradigm a step further.

What are ambient agents?

Ambient agents are AI systems that run in the background, continuously monitoring event streams and then triggered to act when appropriate, according to pre-set instructions and user intent.

While the term “ambient agents” is new, the concept of ambient intelligence, where AI is always listening, is not. Amazon refers to its Alexa personal assistant technology as enabling ambient intelligence.

The goal of ambient agents is to automate repetitive tasks and scale the user’s capabilities by having multiple agents running persistently, rather than the human user having to call them up and interact with each one, one-on-one. This allows the user to focus on higher-level tasks while the agents handle routine work.

To help prove out and advance the concept of ambient agents, LangChain has developed a series of initial use cases, one that monitors emails, the other for social media, to help users manage and respond when needed. 

“I think agents in general are powerful and exciting and cool,” Harrison Chase, cofounder and CEO of LangChain, told VentureBeat. “Ambient agents are way more powerful if there’s a bunch of them doing things in the background, you can just scale yourself way more.”

The tech leverages many open-source solutions, and LangChain did not indicate yet how much it would charge for use of any new tools.

How ambient agents work to improve AI usability

Like many great technology innovations, the original motivation for ambient agents wasn’t to create a new paradigm, but rather to solve a real problem.

For Chase, the problem is one that is all too familiar for many of us: email inbox overload. Chase began his journey to create ambient agents to solve email challenges. Six months ago he started building an ambient agent for his own email.

Chase explained that the email assistant categorizes his emails, handling the triage process automatically. He no longer has to manually sort through his inbox, as the agent takes care of it. Through his own use of the agent inbox over an extended period, Chase was able to refine and improve its capabilities. He noted that it started off imperfect, but by using it regularly and addressing the pain points, he was able to enhance the agent’s performance.

To be clear, the email assistant isn’t some kind of simplistic rules-based system for sorting email. It’s a system that actually understands his email and helps him to decide how to manage it.

The ambient agent architecture for the email assistant use case

The architecture of Chase’s email assistant is quite complex, involving multiple components and language models. 

“It starts off with a triage step that’s kind of like an LLM and a pretty complicated prompt and some few short examples which are retrieved semantically from a vector database,” Chase explained. “Then, if it’s determined that it should try to respond, it goes to a drafting agent.”

Chase further explained that the drafting agent has access to additional tools, including a sub-agent specifically for interacting with the calendar:

“There’s an agent that I have specifically for interacting with the calendar, because actually LLMs kind of suck at dates,” Chase said. “So I had to have a dedicated agent just to interact with the calendar.”

After the draft response is generated, Chase said there’s an additional LLM call that rewrites the response to ensure the correct tone and formatting.

“I found that having the LLM try to call all these tools and construct an email and then also write in the correct tone was really tricky, so I have a step explicitly for tone,” Chase said.

The agent inbox as a way to control and monitor agents

A key part of the ambient agent experience Is having control and visibility into what the agents are doing.

Chase noted that in an initial implementation, he just had agents message via Slack, but that quickly became unwieldy.

Instead, LangChain designed a new user interface, the agent inbox, specifically for interacting with ambient agents.

Screenshot of LangChain agent box. Credit: VentureBeat

The system displays all open lines of communication between users and agents and makes it easy to track outstanding actions.

How to build an ambient agent

LangChain first and foremost is a tool for developers and it’s going to be a tool to help build and deploy ambient agents now too.

Any developer can use the open-source LangChain technology to build an ambient agent, though additional tools can simplify the process. Chase explained that the agent inbox he built is in some respect a view on top of the LangGraph platform. LangGraph is an open-source framework for building agents that provides the infrastructure for operating long-running background jobs.

On top of that, LangChain is using its commercial LangSmith platform, which provides observability and evaluation for agents. This helps developers put agents into production with the necessary monitoring and evaluation tools to ensure they are performing as expected.

Ambient agents: A step toward using generalized intelligence

Chase is optimistic that the concept of ambient agents will catch on with developers in the coming months and years.

Ambient agents bring the prospect of even more autonomy to AI, enabling it to monitor an event stream and take intelligent actions. Chase still expects that there will be a need for keeping humans in the loop as part of the ambient agent experience. But humans need only confirm and validate actions, rather than figure out what needs to be done.

“I think it’s a step towards harnessing and using more generalized intelligence,” Chase said. 

Chase noted that it’s more likely that true AGI will come from improvements in reasoning models. That said, making better use of models is where the concept of ambient agents will bring value.

“There’s still a lot of work to be done to make use of the models, even after they become really intelligent,” Chase said. “I think the ambient agent style of interfacing with them will absolutely be an unlock for using this general form of intelligence.”

An open-source version of the email assistant is currently available. LangChain is releasing a new social media ambient agent today, and will make an open-source version of the agent inbox available on Thursday, January 16.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

TotalEnergies farms out 40% participating interest in certain licenses offshore Nigeria to Chevron

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: var(–color-primary-main); } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style

Read More »

AI-driven network management gains enterprise trust

The way the full process works is that the raw data feed comes in, and machine learning is used to identify an anomaly that could be a possible incident. That’s where the generative AI agents step up. In addition to the history of similar issues, the agents also look for

Read More »

Oil Price Did Not Shift on Fed Cut

In a market update sent to Rigzone by the Rystad Energy team late Wednesday, Rystad highlighted that the price of oil “[did] not shift… on the Fed’s cut”. Rystad pointed out in the update that the Fed lowered its benchmark lending rate by 25 basis points, bringing it to a range of 3.50-3.75 percent, describing the action as “a move that was largely in line with expectations”. “Fundamentals are still the primary drivers of change in commodity markets, with the price of oil not shifting based on the Fed’s cut,” Rystad noted in the statement. “Market participants and investors are paying closer attention to the forward-looking view shared by the central bank,” it added. “The Fed said that uncertainty about the economic outlook remains elevated, and it remains attentive to the risks to its dual mandate of achieving maximum employment and maintaining the inflation rate at two percent,” it continued. In Rystad’s update, Claudio Galimberti, Rystad Energy Chief Economist and Global Director of Market Analysis, stated that “the Federal Reserve’s divided decision to cut rates today [Wednesday] underscores a central bank that is easing cautiously while signaling a potential pause”. “For commodity markets, the message is clear: monetary policy is no longer a dominant driver of price direction. The Fed is cutting, but only reluctantly, and its projections show limited easing ahead despite a still-uncertain labor market and inflation that remains above target,” he added. Galimberti noted in the update that, in the near term, the rate cut modestly loosens financial conditions and may weaken the U.S. dollar at the margin, which he pointed out is typically supportive for crude, metals, and some agricultural commodities. He added, however, that “the signal of a pause tempers that boost, reminding markets that the Fed is unwilling to validate the two-cut easing path

Read More »

Aramco, ExxonMobil Mull Petrochemical Complex at Samref

Exxon Mobil Corp and Saudi Arabian Oil Co (Aramco) have agreed to evaluate upgrading their Samref refinery in Yanbu, Saudi Arabia, with plans to expand the site into an integrated petrochemical complex. The facility currently has a declared oil processing capacity and storage capacity of about 400,000 barrels per day and 13.2 million barrels respectively. It produces mostly gasoline, as well as diesel fuel, heating oil, jet fuel, liquefied petroleum gas and others, the joint venture says on its website. “The companies will explore capital investments to upgrade and diversify production, including high-quality distillates that result in lower emissions and high-performance chemicals, as well as opportunities to improve the refinery’s energy efficiency and reduce emissions from operations through an integrated emissions-reduction strategy”, Aramco said in a press release. Aramco downstream president Mohammed Y. Al Qahtani said, “Designed to increase the conversion of crude oil and petroleum liquids into high-value chemicals, this project reinforces our commitment to advancing downstream value creation and our liquids-to-chemicals strategy. It will also position Samref as a key driver in the growth of the Kingdom’s petrochemical sector”. ExxonMobil senior vice president Jack Williams said, “We look forward to evaluating this project, which aligns with our strategy to focus on investments that allow us to grow high-value products that meet society’s evolving energy needs and contribute to a lower-emission future”. Aramco said, “The companies will commence a preliminary front-end engineering and design phase for the proposed project, which would aim to maximize operational advantages, enhance Samref’s competitiveness and help to meet growing demand for high-quality petrochemical products in the Kingdom”. “Plans are subject to market conditions, regulatory approvals and final investment decisions by Aramco and ExxonMobil”, it said. Samref is equally owned between Aramco and United States energy giant ExxonMobil. In other downstream expansion activities Aramco recently completed

Read More »

Naftogaz Seals New PrivatBank Loan for Winter Gas

Naftogaz Group has secured a fresh UAH 5 billion ($118.82 million) loan from local bank JSC CB PrivatBank to procure more natural gas for Ukraine. “To get through the winter steadily, Naftogaz needs to import 4.4 billion cubic meters of gas [155.38 billion cubic feet], with a total cost of about EUR 1.9 billion [$2.22 billion]”, the state-owned integrated energy company said in an online statement, noting local production has fallen due to war damage. Earlier Naftogaz sealed local loans totaling UAH 9.4 billion from PrivatBank and PJSC JSB Ukrgasbank, each committing UAH 4.7 billion, to buy gas for the heating season, as announced by Naftogaz in July. Chief executive Sergii Koretskyi said then, “At the same time, we continue to work with international financial institutions and partner countries”. Also in 2025 JSC State Saving Bank of Ukraine agreed to lend Naftogaz UAH 3 billion for gas import, as announced by Naftogaz October 9. On November 13 Naftogaz and the European Investment Bank (EIB) announced a grant of EUR 127 million for the procurement of gas for Ukraine, in addition to a EUR 300 million EIB loan to Naftogaz that was disbursed October for the same purpose. The loans are guaranteed under the Ukraine Investment Framework (UIF), part of the European Union’s Ukraine Facility. The facility aims to mobilize up to EUR 50 billion – EUR 33 billion in loans and EUR 17 billion in grants – from 2024 to 2027, according to the European Council, which approved the facility February 2024. The UIF aims to mobilize up to EUR 40 billion of investments for recovery, reconstruction and modernization, according to its implementer the European Commission. On August 13 Naftogaz and the European Bank for Reconstruction and Development (EBRD) said they had agreed on a EUR 270 million loan for gas purchases.

Read More »

CNOOC Ltd Announces 12th Startup in Chinese Waters in 2025

CNOOC Ltd has started production in the Weizhou 11-4 Oilfield Adjustment and Satellite Fields Development Project in the Beibu Gulf Basin of the South China Sea. This is the company’s 12th announced startup offshore China this year, and the sixth in the South China Sea. Including two projects in Brazilian waters and Guyana’s Yellowtail, the publicly listed arm of China National Offshore Oil Corp has now announced 15 upstream startups in 2025. CNOOC Ltd expects the newest startup to reach its capacity of about 16,900 barrels of oil equivalent a day in 2026. The oil is light crude, according to the company. The development plan eyes 35 wells: 28 for production and seven for water injection. “The main production facilities include a newly-built unmanned wellhead platform and a central processing platform, which are connected to an existing platform through a trestle bridge”, CNOOC Ltd said in a press release. “The project has adopted a coordinated development plan of ‘three offshore processing centers + one onshore terminal’, serving as a gathering and transportation hub to release the resource capacity and ensure stable energy supply in the region”, the sole owner said. The project has an average water depth of around 43 meters (141.08 feet), according to CNOOC Ltd. Earlier this year it put online five projects in the South China Sea: the Dongfang 29-1 field, the Panyu 11-12/10-1/10-2 Oilfield Adjustment Joint Development Project, the Weizhou 5-3 field, the Wenchang 16-2 field and phase II of the Wenchang 19-1 field. In the Bohai Sea, CNOOC Ltd put onstream four projects in 2025: phase I of the Bozhong 26-6 field, the Caofeidian 6-4 field adjustment, phase I of the Kenli 10-2 Oilfields Development Project and phase II of the Luda 5-2 North field. In the Yinggehai Basin, two came online this year: the Dongfang 1-1 Gas

Read More »

Crude rises after US seizes Venezuelan tanker

Oil futures erased earlier declines after US forces intercepted and seized a sanctioned oil tanker off the coast of Venezuela, a move that marks a major escalation of tensions between the two countries. West Texas Intermediate traded higher to settle above $58 after earlier dropping as much as 1%. Brent crude settled above $62. The seizure may make it much more difficult for Venezuela to send its oil overseas, as other shippers are now likely to be more reluctant to load its cargoes. Most Venezuelan oil heads to China, usually through intermediaries, at steep discounts due to sanctions risk. US President Donald Trump has suggested numerous times that the US could strike on land in Venezuela and that the country’s President Nicolas Maduro’s “days are numbered.” “Tensions are continuing to move up the escalation ladder and introduce some short-term supply risk,” said Rebecca Babin, a senior energy trader at CIBC Private Wealth Group. “That said, given the administration’s clear desire to keep oil and gasoline prices contained, the market is only assigning a small risk premium. Any potential disruption is still being viewed as short-lived.” Still, oversupply concerns continue to weigh on sentiment. The US said domestic crude production would hit a record 13.6 million barrels a day this year, adding to a flood of supply hitting the global market, while several of India’s largest refiners are buying sanctioned Russian oil, easing the worst fears of a supply threat. Ukraine carried out yet another attack on a Russian shadow-fleet oil tanker as it continues to target Moscow’s vital seaborne petroleum trade. Meantime, data from the US Energy Information Administration on Wednesday showed US inventories declined 1.8 million barrels, the first draw on stocks in around three weeks. Inventories at the Cushing, Oklahoma, storage hub rose. Data also showed a surge

Read More »

South Sudan Oil Exports at Risk

South Sudan’s oil exports faced a new threat after rebels in war-torn Sudan seized facilities key to transporting crude to the Red Sea. Workers fled and operations were halted at Heglig, an oil hub in southern Sudan, as the Rapid Support Forces paramilitary group approached the area, according to people familiar with the events. The RSF, which has been battling Sudan’s army since April 2023, said on Telegram on Monday it had taken control of the “strategic Heglig” area, and was committed to securing the oil facilities. The development raises the prospect of another halt in exports of South Sudan’s Dar Blend after a disruption in mid-November. While Sudan ships little crude of its own, it’s the sole conduit for oil from landlocked South Sudan. Heglig — which lies near their border — plays a crucial role in the pipeline network.  Sudan exported an average of 165,000 barrels a day of its neighbor’s crude in the past three months, according to tanker-tracking data compiled by Bloomberg.  The army-backed government in Sudan didn’t respond to requests for comment. Nor did South Sudanese authorities, who rely on oil exports for the vast majority of state revenue. The RSF’s seizure of Heglig caps a string of recent territorial gains in southern Sudan for the group that’s been accused by the US of genocide in a conflict in which hundreds of thousands of people are thought to have died. The World Health Organization on Monday said drone strikes on a kindergarten and a nearby hospital in South Kordofan on Dec. 4 had killed 114 people, including 63 children. WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate or insulting comments will be removed.

Read More »

New Nvidia software gives data centers deeper visibility into GPU thermals and reliability

Addressing the challenge Modern AI accelerators now draw more than 700W per GPU, and multi-GPU nodes can reach 6kW, creating concentrated heat zones, rapid power swings, and a higher risk of interconnect degradation in dense racks, according to Manish Rawat, semiconductor analyst at TechInsights. Traditional cooling methods and static power planning increasingly struggle to keep pace with these loads. “Rich vendor telemetry covering real-time power draw, bandwidth behavior, interconnect health, and airflow patterns shifts operators from reactive monitoring to proactive design,” Rawat said. “It enables thermally aware workload placement, faster adoption of liquid or hybrid cooling, and smarter network layouts that reduce heat-dense traffic clusters.” Rawat added that the software’s fleet-level configuration insights can also help operators catch silent errors caused by mismatched firmware or driver versions. This can improve training reproducibility and strengthen overall fleet stability. “Real-time error and interconnect health data also significantly accelerates root-cause analysis, reducing MTTR and minimizing cluster fragmentation,” Rawat said. These operational pressures can shape budget decisions and infrastructure strategy at the enterprise level.

Read More »

Arista goes big with campus wireless tech

In a white paper describing how VESPA works, Arista wrote: The first component of VESPA involves Arista access points creating VXLAN tunnels to Arista switches serving as WLAN Gateways…. Second, as device packets arrive via the AP, it dynamically creates an Ethernet Segment Identifier (Type 6 ESI) based on the AP’s VTEP IP address. These dynamically created tunnels can scale to 30K ESI’s spread across paired switches in the cluster which provide active/active load sharing (performance+HA) to the APs. Third, the gateway switches use Type 2 EVPN NLRI (Network Layer Reachability Information) to learn and exchange end point MAC addresses across the cluster. … With this architecture, adding more EVPN WLAN gateways scales both AP and user connections, to tens of thousands of end points. To manage the forwarding information for hundreds of thousands of clients (e.g: FIB next hop and rewrite) would prove very complex and expensive if using conventional networking solutions. Arista’s innovation is to distribute this function across the WiFi access points with a unique MAC Rewrite Offload feature (MRO). With MRO, the access point is responsible for servicing mobile client ARP requests (using its own mac address), building a localized MAC-IP binding table, and forwarding client IP addresses to the WLAN gateways with the APs MAC address. The WLAN Gateways therefore only learns one (MAC) address for all the clients associated with the AP. This improves the gateway’s scaling from 10X to 100X, allowing these cost effective gateways to support hundreds of thousands of clients attached to the APs. AVA system gets a boost In addition to the new wireless technology, Arista is also bolstering the capabilities of its natural-language, generative AI-based Autonomous Virtual Assist (AVA) system for delivering network insights and AIOps.  AVA is aimed at providing an intelligent assistant that’s not there to replace

Read More »

Most significant networking acquisitions of 2025

Cisco makes two AI deals: EzDubs and NeuralFabric Last month Cisco completed its acquisition of EzDubs, a privately held AI software company with speech-to-speech translation technology. EzDubs translates conversations across 31 languages and will accelerate Cisco’s delivery of next-generation features, such as live voice translation that preserves the characteristics of speech, the vendor stated. Cisco plans to incorporate EzDubs’ technology in its Cisco Collaboration portfolio. Also in November, Cisco bought AI platform company NeuralFabric, which offers a generative AI platform that lets organizations develop domain-specific small language models using their own proprietary data. Coreweave buys Core Scientific Nvidia-backed AI cloud provider CoreWeave acquired crypto miner Core Scientific for about $9 billion, giving it access to 1.3 gigawatts of contracted power to support growing demand for AI and high-performance computing workloads. CoreWeave said the deal augments its vertical integration by expanding its owned and operated data center footprint, allowing it to scale GPU-powered services for enterprise and research customers. F5 picks up three: CalypsoAI, Fletch and MantisNet F5 acquired Dublin, Ireland-based CalypsoAI for $180 million. CalypsoAI’s platform creates what the company calls an Inference Perimeter that protects across models, vendors, and environments. F5 says it will integrate CalypsoAI’s adaptive AI security capabilities into its F5 Application Delivery and Security Platform (ADSP). F5’s ADSP also stands to gain from F5’s acquisition of agentic AI and threat management startup Fletch. Fletch’s technology turns external threat intelligence and internal logs into real-time, prioritized insights; its agentic AI capabilities will be integrated into ADSP, according to F5. Lastly, F5 grabbed startup MantisNet to enhance cloud-native observability in F5’s ADSP. MantisNet leverages extended Berkeley Packet Filer (eBPF)-powered, kernel-level telemetry to provide real-time insights into encrypted protocol activity and allow organizations “to gain visibility into even the most elusive traffic, all without performance overhead,” according to an F5 blog

Read More »

Aviz Networks launches enterprise-grade community SONiC distribution

First, the company enabled FRR (Free Range Routing) features that exist in the community code but aren’t consistently implemented across different ASICs. VRRP (Virtual Router Redudancy Protocol) provides router redundancy for high availability. Spanning tree variants prevent network loops in layer 2 topologies. MLAG allows two switches to act as a single logical device for link aggregation. EVPN enhancements support layer 2 and layer 3 VPN services over VXLAN overlays. These protocols work differently depending on the underlying silicon, so Aviz normalized their implementation across Broadcom, Nvidia, Cisco and Marvell chips. Second, Aviz fixed bugs discovered in production deployments. One customer deployed community SONiC with OpenStack and started migrating virtual machines between hosts. The network fabric couldn’t handle the workload and broke. Aviz identified the failure modes and patched them.  Third, Aviz built a software component that normalizes monitoring data across vendors. Broadcom’s Tomahawk ASIC generates different telemetry formats than Nvidia’s Spectrum or Cisco’s Silicon One. Network operators need consistent data for troubleshooting and capacity planning. The software collects ASIC-specific logs and network operating system telemetry, then translates them into a standardized format that works the same way regardless of which silicon vendor’s chips are running in the switches. Validated for enterprise deployment scenarios The distribution supports common enterprise network architectures.  IP CLOS provides the leaf-spine topology used in modern data centers for predictable latency and scalability. EVPN/VXLAN creates layer 2 and layer 3 overlay networks that span physical network boundaries. MLAG configurations provide link redundancy without spanning tree limitations. Aviz provides validated runbooks for these deployments across data center, edge and AI fabric use cases. 

Read More »

US approves Nvidia H200 exports to China, raising questions about enterprise GPU supply

Shifting demand scenarios What remains unclear is how much demand Chinese firms will actually generate, given Beijing’s recent efforts to steer its tech companies away from US chips. Charlie Dai, VP and principal analyst at Forrester, said renewed H200 access is likely to have only a modest impact on global supply, as China is prioritizing domestic AI chips and the H200 remains below Nvidia’s latest Blackwell-class systems in performance and appeal. “While some allocation pressure may emerge, most enterprise customers outside China will see minimal disruption in pricing or lead times over the next few quarters,” Dai added. Neil Shah, VP for research and partner at Counterpoint Research, agreed that demand may not surge, citing structural shifts in China’s AI ecosystem. “The Chinese ecosystem is catching up fast, from semi to stack, with models optimized on the silicon and software,” Shah said. Chinese enterprises might think twice before adopting a US AI server stack, he said. Others caution that even selective demand from China could tighten global allocation at a time when supply of high-end accelerators remains stretched, and data center deployments continue to rise.

Read More »

What does Arm need to do to gain enterprise acceptance?

But in 2017, AMD released the Zen architecture, which was equal if not superior to the Intel architecture. Zen made AMD competitive, and it fueled an explosive rebirth for a company that was near death a few years prior. AMD now has about 30% market share, while Intel suffers from a loss of technology as well as corporate leadership. Now, customers have a choice of Intel or AMD, and they don’t have to worry about porting their applications to a new platform like they would have to do if they switched to Arm. Analysts weigh in on Arm Tim Crawford sees no demand for Arm in the data center. Crawford is president of AVOA, a CIO consultancy. In his role, he talks to IT professionals all the time, but he’s not hearing much interest in Arm. “I don’t see Arm really making a dent, ever, into the general-purpose processor space,” Crawford said. “I think the opportunity for Arm is special applications and special silicon. If you look at the major cloud providers, their custom silicon is specifically built to do training or optimized to do inference. Arm is kind of in the same situation in the sense that it has to be optimized.” “The problem [for Arm] is that there’s not necessarily a need to fulfill at this point in time,” said Rob Enderle, principal analyst with The Enderle Group. “Obviously, there’s always room for other solutions, but Arm is still going to face the challenge of software compatibility.” And therein lies what may be Arm’s greatest challenge: software compatibility. Software doesn’t care (usually) if it’s on Intel or AMD, because both use the x86 architecture, with some differences in extensions. But Arm is a whole new platform, and that requires porting and testing. Enterprises generally don’t like disruption —

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »