Stay Ahead, Stay ONMINE

Study warns of security risks as ‘OS agents’ gain control of computers and phones

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Researchers have published the most comprehensive survey to date of so-called “OS Agents” — artificial intelligence systems that can autonomously control computers, mobile phones and web browsers by directly […]

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now


Researchers have published the most comprehensive survey to date of so-called “OS Agents” — artificial intelligence systems that can autonomously control computers, mobile phones and web browsers by directly interacting with their interfaces. The 30-page academic review, accepted for publication at the prestigious Association for Computational Linguistics conference, maps a rapidly evolving field that has attracted billions in investment from major technology companies.

“The dream to create AI assistants as capable and versatile as the fictional J.A.R.V.I.S from Iron Man has long captivated imaginations,” the researchers write. “With the evolution of (multimodal) large language models ((M)LLMs), this dream is closer to reality.”

The survey, led by researchers from Zhejiang University and OPPO AI Center, comes as major technology companies race to deploy AI agents that can perform complex digital tasks. OpenAI recently launched “Operator,” Anthropic released “Computer Use,” Apple introduced enhanced AI capabilities in “Apple Intelligence,” and Google unveiled “Project Mariner” — all systems designed to automate computer interactions.

The speed at which academic research has transformed into consumer-ready products is unprecedented, even by Silicon Valley standards. The survey reveals a research explosion: over 60 foundation models and 50 agent frameworks developed specifically for computer control, with publication rates accelerating dramatically since 2023.


AI Scaling Hits Its Limits

Power caps, rising token costs, and inference delays are reshaping enterprise AI. Join our exclusive salon to discover how top teams are:

  • Turning energy into a strategic advantage
  • Architecting efficient inference for real throughput gains
  • Unlocking competitive ROI with sustainable AI systems

Secure your spot to stay ahead: https://bit.ly/4mwGngO


This isn’t just incremental progress. We’re witnessing the emergence of AI systems that can genuinely understand and manipulate the digital world the way humans do. Current systems work by taking screenshots of computer screens, using advanced computer vision to understand what’s displayed, then executing precise actions like clicking buttons, filling forms, and navigating between applications.

“OS Agents can complete tasks autonomously and have the potential to significantly enhance the lives of billions of users worldwide,” the researchers note. “Imagine a world where tasks such as online shopping, travel arrangements booking, and other daily activities could be seamlessly performed by these agents.”

The most sophisticated systems can handle complex multi-step workflows that span different applications — booking a restaurant reservation, then automatically adding it to your calendar, then setting a reminder to leave early for traffic. What took humans minutes of clicking and typing can now happen in seconds, without human intervention.

The development of AI agents requires a complex training pipeline that combines multiple approaches, from initial pre-training on screen data to reinforcement learning that optimizes performance through trial and error. (Credit: arxiv.org)

Why security experts are sounding alarms about AI-controlled corporate systems

For enterprise technology leaders, the promise of productivity gains comes with a sobering reality: these systems represent an entirely new attack surface that most organizations aren’t prepared to defend.

The researchers dedicate substantial attention to what they diplomatically term “safety and privacy” concerns, but the implications are more alarming than their academic language suggests. “OS Agents are confronted with these risks, especially considering its wide applications on personal devices with user data,” they write.

The attack methods they document read like a cybersecurity nightmare. “Web Indirect Prompt Injection” allows malicious actors to embed hidden instructions in web pages that can hijack an AI agent’s behavior. Even more concerning are “environmental injection attacks” where seemingly innocuous web content can trick agents into stealing user data or performing unauthorized actions.

Consider the implications: an AI agent with access to your corporate email, financial systems, and customer databases could be manipulated by a carefully crafted web page to exfiltrate sensitive information. Traditional security models, built around human users who can spot obvious phishing attempts, break down when the “user” is an AI system that processes information differently.

The survey reveals a concerning gap in preparedness. While general security frameworks exist for AI agents, “studies on defenses specific to OS Agents remain limited.” This isn’t just an academic concern — it’s an immediate challenge for any organization considering deployment of these systems.

The reality check: Current AI agents still struggle with complex digital tasks

Despite the hype surrounding these systems, the survey’s analysis of performance benchmarks reveals significant limitations that temper expectations for immediate widespread adoption.

Success rates vary dramatically across different tasks and platforms. Some commercial systems achieve success rates above 50% on certain benchmarks — impressive for a nascent technology — but struggle with others. The researchers categorize evaluation tasks into three types: basic “GUI grounding” (understanding interface elements), “information retrieval” (finding and extracting data), and complex “agentic tasks” (multi-step autonomous operations).

The pattern is telling: current systems excel at simple, well-defined tasks but falter when faced with the kind of complex, context-dependent workflows that define much of modern knowledge work. They can reliably click a specific button or fill out a standard form, but struggle with tasks that require sustained reasoning or adaptation to unexpected interface changes.

This performance gap explains why early deployments focus on narrow, high-volume tasks rather than general-purpose automation. The technology isn’t yet ready to replace human judgment in complex scenarios, but it’s increasingly capable of handling routine digital busywork.

OS agents rely on interconnected systems for perception, planning, memory and action execution. The complexity of coordinating these components helps explain why current systems still struggle with sophisticated tasks. (Credit: arxiv.org)

What happens when AI agents learn to customize themselves for every user

Perhaps the most intriguing — and potentially transformative — challenge identified in the survey involves what researchers call “personalization and self-evolution.” Unlike today’s stateless AI assistants that treat every interaction as independent, future OS agents will need to learn from user interactions and adapt to individual preferences over time.

“Developing personalized OS Agents has been a long-standing goal in AI research,” the authors write. “A personal assistant is expected to continuously adapt and provide enhanced experiences based on individual user preferences.”

This capability could fundamentally change how we interact with technology. Imagine an AI agent that learns your email writing style, understands your calendar preferences, knows which restaurants you prefer, and can make increasingly sophisticated decisions on your behalf. The potential productivity gains are enormous, but so are the privacy implications.

The technical challenges are substantial. The survey points to the need for better multimodal memory systems that can handle not just text but images and voice, presenting “significant challenges” for current technology. How do you build a system that remembers your preferences without creating a comprehensive surveillance record of your digital life?

For technology executives evaluating these systems, this personalization challenge represents both the greatest opportunity and the largest risk. The organizations that solve it first will gain significant competitive advantages, but the privacy and security implications could be severe if handled poorly.

The race to build AI assistants that can truly operate like human users is intensifying rapidly. While fundamental challenges around security, reliability, and personalization remain unsolved, the trajectory is clear. The researchers maintain an open-source repository tracking developments, acknowledging that “OS Agents are still in their early stages of development” with “rapid advancements that continue to introduce novel methodologies and applications.”

The question isn’t whether AI agents will transform how we interact with computers — it’s whether we’ll be ready for the consequences when they do. The window for getting the security and privacy frameworks right is narrowing as quickly as the technology is advancing.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Mapping Trump’s tariffs by trade balance and geography

U.S. importers may soon see costs rise for many imported goods, as tariffs on foreign goods are set to rise. On July 31, President Donald Trump announced country-specific reciprocal tariffs would finally be implemented on Aug. 7, after a monthslong pause. The news means more than 90 countries will see

Read More »

JF Expands in Southwest with Maverick Acquisition

The JF Group (JF) has acquired Arizona-based Maverick Petroleum Services. JF, a fueling infrastructure, petroleum equipment distribution, service, general contracting, and construction services provider, said in a media release that Maverick brings expertise in the installation, maintenance, and repair of petroleum handling equipment, Point-of-Sale (POS) systems, and environmental testing. As

Read More »

$1B Qatar Gas Loan Adds to Middle East Funding Frenzy

Qatar Gas Transport Co. is seeking a $1 billion syndicated loan, according to a person familiar with the matter, adding to the flurry of Middle East borrowers that have tapped Asian lenders recently. Mizuho Bank Ltd. is the sole mandated lead arranger and bookrunner of the five-year deal, the person said, who asked not to be identified discussing private matters. The deal carries a greenshoe option, which allows the size to increase by an additional $330 million, the person said.  The loan, which is being syndicated to the broader market, pays an interest margin of 82 basis points over the benchmark Secured Overnight Financing Rate, the person said, adding that the proceeds are for general corporate purposes.  Qatar Gas didn’t immediately respond to a request for comment.  Qatar Gas joins a slew of Middle East borrowers, notably from the Gulf States, keen to tap banks in Asia to diversify fundraising beyond their domestic capital markets. Saudi Investment Bank, for example, just launched a syndicated loan of as much as $750 million, while Saudi Electricity Co. is in the market with a $1 billion facility. Qatar Gas, more commonly known as Nakilat, is expanding its liquefied natural gas fleet as the nation seeks to reinforce its position as a leading global supplier of clean energy, according to local media. The borrower last month launched its first financing package with the Export-Import Bank of Korea to build 25 conventional Korean-built LNG vessels. WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate or insulting comments will be removed.

Read More »

Oil Steadies After Slide Last Week

Oil was little changed after slumping 5% last week as investors await details from US President Donald Trump’s meeting with Russian President Vladimir Putin to gauge the trajectory of supply additions to the global market. West Texas Intermediate settled near $64 a barrel, hovering near the lowest in two months in thin summer trading. On Friday, Trump didn’t reveal additional measures against Moscow or buyers of its crude as he announced a summit with Putin in Alaska. He had earlier declared an Aug. 8 deadline for the Kremlin to reach a ceasefire. Trump later downplayed expectations for his upcoming meeting with Putin, casting it as a “feel-out meeting” and saying he would confer with Ukrainian and European leaders after the sitdown. US and Russian officials are working toward an agreement that would lock in Moscow’s occupation of territory seized during its military invasion, according to people familiar with the matter. The US is working to get buy-in from Ukraine and its European allies on the deal, which is far from certain, they said. Oil has lost more than 10% this year as OPEC+ brings back production faster than initially planned, ending curbs made in 2023, even as slowing economic growth threatens to cut consumption. A peace deal with Ukraine could see an end to sanctions on supply from Russia, removing the risk of disruption to Moscow’s flows after Trump’s comments in recent weeks that he would put measures in place against its biggest buyers. “This is not a deal which will be closed on Friday, but rather the start of a process,” said Bjarne Schieldrop, chief commodities analyst at SEB AB. “Trump is very, very unlikely to slap sanctions on Russian oil while this process is ongoing, i.e. no disruption of Russian oil in sight.” Investors may get further insight

Read More »

Trump to Tap Biden Appointee to Lead FERC

President Donald Trump is preparing to tap David Rosner to be chair of the Federal Energy Regulatory Commission, which oversees key decisions about natural gas export terminals and power lines. The move, described by a White House official who asked for anonymity before a formal announcement, would put a Democrat at the head of an agency central to Trump’s plans to propel American oil, gas and coal.  Rosner, who was appointed to the commission by former President Joe Biden, previously served as an aide to then-Senator Joe Manchin, a Democrat-turned-independent who represented West Virginia and advocated for gas and coal interests in the chamber before his departure in January. Rosner is seen as supportive of the president’s priorities, the official said.  The commission became a political battlefield under the first Trump administration as the president’s appointees unsuccessfully pushed policies promoting natural gas- and coal-fired power. While that effort continues, the data-center boom has ushered in new demand for natural gas and other forms of cheap energy. Last month, an annual power sale by the biggest US electric grid saw prices soar to a record $16.1 billion, underscoring the growing need for megawatts to power artificial intelligence.  The planned elevation of Rosner, which was previously reported by Axios, would put the commissioner in the chair role just vacated by Mark Christie, a Republican.  WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate or insulting comments will be removed.

Read More »

FERC’s Christie steps down, leaving agency with three members

Mark Christie, Federal Energy Regulatory Commission chairman, stepped down from his position at the agency on Friday, leaving it with three sitting commissioners. “I am not retiring but looking forward to whatever comes next,” Christie said on X.  “As I said to staff and colleagues when I left the Virginia [State Corporation Commission], ‘Good luck and I’ll see you on the trail ahead.’” After being nominated by President Donald Trump, Christie, a Republican, joined FERC in January 2021. Trump named him chairman four years later, but decided against tapping him for a second five-year term. Before joining FERC, Christie was a member of the Virginia SCC for nearly 17 years. While at FERC, Christie warned the United States was heading towards a grid reliability crisis, partly driven by retiring power plants. He also sought to trim the financial incentives FERC offers companies that build transmission lines and he was a capacity market skeptic. The White House intends to name FERC Commissioner David Rosner, a Democrat, as interim chairman, according to Axios and E&E News. The agency’s two other commissioners are Lindsay See, a Republican, and Judy Chang, a Democrat. Rosner received an unusually high level of support from Republicans during his Senate confirmation process, according to Devin Hartman, director of energy and environmental policy at the R Street Institute, a free market-oriented think tank. “It is hard to know the specific rationale for this decision, which of course raised some eyebrows, because rarely have we seen President Trump pass the baton of an agency to a Democrat,” Hartman said Monday. “It’s at least in part a sign of their confidence in Rosner as the interim chair, and it makes it even more clear that the administration is thinking that their permanent chair is not yet appointed to FERC.” Former FERC Chairman

Read More »

Oncor has 200 GW of interconnection requests, company officials say

Dive Brief: Oncor Electric Delivery now has some 200 GW of interconnection requests in its queue, including 186 GW of data centers, 7 GW of traditional commercial and industrial customers, 5 GW of crypto currency facilities, and 4 GW of oil and gas operations, Oncor CEO Allen Nye said during Thursday’s earnings call. About 20% of the potential demand has signed contracts or is considered “high-confidence load,” he said. Given the number of interconnection requests, Oncor could add more than $12 billion to its existing $36 billion capital plan when it revises that plan sometime next year, according to Jeffrey Martin, chairman, CEO and president of Oncor parent company Sempra. The sales of Sempra’s Ecogas Mexico and a stake in Sempra Infrastructure and have drawn interest from potential buyers and financiers, Martin said. The sales are intended to raise funds for Oncor’s multibillion dollar expansion. Dive Insight Sempra’s pivot to a more utility-focused business model, inspired in part by the massive growth in electric demand in Oncor’s service territory, should improve earnings and reduce risk for investors, Martin told analysts on Thursday. Martin announced on the call that Sempra Infrastructure, the company’s development arm, has signed a nonbinding letter of intent with global investment firm KKR for the sale of 15% to 30% equity depending on the company’s valuation. The Ecogas sale has also drawn interest, he said, and both sales are expected to close in mid-2026. “It’s also very important … that we’re thoughtful on the timing and use of proceeds.” Martin said. “There’s an opportunity here to improve our balance sheet and put some cushion on the balance sheet.” Sempra should see more of its earnings coming from its regulated utilities, and from Oncor and Texas in particular, in the future, Martin said. Oncor initiated service to 20,000 new

Read More »

Nuclear regulatory approval drives NuScale customer interest, but no deals yet

Dive Brief: Small modular reactor company NuScale Power aims to have “hard contracts” with “two or three major customers” by the end of 2025, CEO John Hopkins told investors and analysts on Thursday.  Hopkins’ comments came as the company reported a significant jump in expenses in its Q2 2025 earnings update. It attributed the change to “higher business development costs associated with NuScale’s transition from a research and development-based company to a commercial company.”  Though NuScale and its developer partner Entra1 have yet to finalize a deal, “we’re getting inundated now” with prospective customers following the Nuclear Regulatory Commission’s May 29 approval of NuScale’s 77-MW power module, Hopkins said. Dive Insight: NuScale’s 77-MW module supplanted an earlier 50-MW design the NRC approved in 2023. Some prospective customers had been in a holding pattern as the commission considered NuScale’s application for the uprated module, Hopkins said, adding, “It was accomplished. We’re there.” NuScale Chief Financial Officer Ramsey Hamady said the May approval puts NuScale in a class by itself among advanced nuclear technology companies.“We’re the only company with two NRC approvals for small modular reactors. There’s no other company with even one … and there [were] a lot of people out there doubting [us], saying, ‘Hey, you’re not gonna get through.’” For NuScale, the NRC decision amplified regulatory tailwinds supporting the wider nuclear industry, Hamady said.  In an investor presentation Thursday, NuScale executives said four executive orders President Donald Trump signed in May would shorten regulatory timelines for new reactor deployments, bolster domestic nuclear supply chains and enable reactor development on military and other government-owned lands. The Trump administration “is pressed to get success stories quickly” on new nuclear deployments, Hopkins said, and it has “a limited period of time to make that happen.” That bodes well for NuScale and other

Read More »

Enterprise tips for cloud success

The remaining tips were cited by roughly two-thirds of the enterprises. Tip number three is to look especially at applications whose users are widely dispersed. And by “widely” here, they mean on different continents, not just different neighborhoods. The reason is that quality of experience and even availability can be compromised when work has to transit a lot of networks just to get to where it’s processed. This can lead to user dissatisfaction, and dispersing resources closer to the users may be the only solution. If an enterprise doesn’t already have their own data center located close to each user concentration, chances are that putting a new hosting point in themselves couldn’t achieve reasonable economy of scale in capex, power and cooling, and operations costs. The cloud would be cheaper. A qualifying comment here is to take great care in evaluating the real impact of dispersion of application users. In some cases, there may not be enough of a difference in QoE or availability to require dispersing hosting points, and in fact it may be that where the application is hosted isn’t even the problem. “The cloud may look like the easy way out,” one enterprise said, “but it may not be the economical way.” See where your QoE issues really lie before you go to the cloud’s distributed hosting to fix them. Tip four is to examine the user-to-application interaction model carefully, to see if there’s a large non-transactional component. Mission-critical business systems, and business core databases, are almost always in the data center. The stuff that changes them are the transactions that add, update, and delete records. If an application’s user interaction is tightly coupled to the creation of transactions, then its processing is tied to those data center resources. That makes it harder to move the user-interface

Read More »

Stargate’s slow start reveals the real bottlenecks in scaling AI infrastructure

The CFO emphasized that SoftBank remains committed to its original target of $346 billion (JPY 500 billion) over 4 years for the Stargate project, noting that major sites have been selected in the US and preparations are taking place simultaneously across multiple fronts. Requests for comment to Stargate partners Nvidia, OpenAI, and Oracle remain unanswered. Infrastructure reality check for CIOs These challenges offer important lessons for enterprise IT leaders facing similar AI infrastructure decisions. Sanchit Vir Gogia, chief analyst and CEO at Greyhound Research, said that Goto’s confirmation of delays “reflects a challenge CIOs see repeatedly” in partner onboarding delays, service activation slips, and revised delivery commitments from cloud and datacenter providers. Oishi Mazumder, senior analyst at Everest Group, noted that “SoftBank’s Stargate delays show that AI infrastructure is not constrained by compute or capital, but by land, energy, and stakeholder alignment.” The analyst emphasized that CIOs must treat AI infrastructure “as a cross-functional transformation, not an IT upgrade, demanding long-term, ecosystem-wide planning.” “Scaling AI infrastructure depends less on the technical readiness of servers or GPUs and more on the orchestration of distributed stakeholders — utilities, regulators, construction partners, hardware suppliers, and service providers — each with their own cadence and constraints,” Gogia said.

Read More »

Incentivizing the Digital Future: Inside America’s Race to Attract Data Centers

Across the United States, states are rolling out a wave of new tax incentives aimed squarely at attracting data centers, one of the country’s fastest-growing industries. Once clustered in only a handful of industry-friendly regions, today’s data-center boom is rapidly spreading, pushed along by profound shifts in federal policy, surging demand for artificial intelligence, and the drive toward digital transformation across every sector of the economy. Nowhere is this transformation more visible than in the intensifying state-by-state competition to land massive infrastructure investments, advanced technology jobs, and the alluring prospect of long-term economic growth. The past year alone has seen a record number of states introducing or expanding incentives for data centers, from tax credits to expedited permitting, reflecting a new era of proactive, tech-focused economic development policy. Behind these moves, federal initiatives and funding packages underscore the essential role of digital infrastructure as a national priority, encouraging states to lower barriers for data center construction and operation. As states watch their neighbors reap direct investment and job creation benefits, a real “domino effect” emerges: one state’s success becomes another’s blueprint, heightening the pressure and urgency to compete. Yet, this wave of incentives also exposes deeper questions about the local impact, community costs, and the evolving relationship between public policy and the tech industry. From federal levels to town halls, there are notable shifts in both opportunities and challenges shaping the landscape of digital infrastructure advancement. Industry Drivers: the Federal Push and Growth of AI The past year has witnessed a profound federal policy shift aimed squarely at accelerating U.S. digital infrastructure, especially for data centers in direct response both to the explosive growth of artificial intelligence and to intensifying international competition. In July 2025, the administration unveiled “America’s AI Action Plan,” accompanied by multiple executive orders that collectively redefined

Read More »

AI Supercharges Hyperscale: Capacity, Geography, and Design Are Being Redrawn

From Cloud to GenAI, Hyperscalers Cement Role as Backbone of Global Infrastructure Data center capacity is undergoing a major shift toward hyperscale operators, which now control 44 percent of global capacity, according to Synergy Research Group. Non-hyperscale colocations account for another 22 percent of capacity and is expected to continue, but hyperscalers projected to hold 61 percent of the capacity by 2030. That swing also reflects the dominance of hyperscalers geographically. In a separate Synergy study revealing the world’s top 20 hyperscale data center locations, just 20 U.S. state or metro markets account for 62 percent of the world’s hyperscale capacity.  Northern Virginia and the Greater Beijing areas alone make up 20 percent of the total. They’re followed by the U.S. states of Oregon and Iowa, Dublin, the U.S. state of Ohio, Dallas, and then Shanghai. Of the top 20 markets, 14 are in the U.S., five in APAC region, and only one is in Europe. This rapid shift is fueled by the explosive growth of cloud computing, artificial intelligence (AI), and especially generative AI (GenAI)—power-intensive technologies that demand the scale, efficiency, and specialized infrastructure only hyperscalers can deliver. What’s Coming for Capacity The capacity research shows on-premises data centers with 34 percent of the total capacity, a significant drop from the 56 percent capacity they accounted for just six years ago.  Synergy projects that by 2030, hyperscale operators such as Google Cloud, Amazon Web Services, and Microsoft Azure will claim 61 percent of all capacity, while on-premises share will drop to just 22 percent. So, it appears on-premises data centers are both increasing and decreasing. That’s one way to put it, but it’s about perspective. Synergy’s capacity study indicates they’re growing as the volume of enterprise GPU servers increases. The shrinkage refers to share of the market: Hyperscalers are growing

Read More »

In crowded observability market, Gartner calls out AI capabilities, cost optimization, DevOps integration

Support for OpenTelemetry and open standards is another differentiator for Gartner. Vendors that embrace these frameworks are better positioned to offer extensibility, avoid vendor lock-in, and enable broader ecosystem integration. This openness is paired with a growing focus on cost optimization—an increasingly important concern as telemetry data volumes increase. Leaders offer granular data retention controls, tiered storage, and usage-based pricing models to help customers Gartner also highlights the importance of the developer experience and DevOps integration. Observability leaders provide “integration with other operations, service management, and software development technologies, such as IT service management (ITSM), configuration management databases (CMDB), event and incident response management, orchestration and automation, and DevOps tools.” On the automation front, observability platforms should support initiating changes to application and infrastructure code to optimize cost, capacity or performance—or to take corrective action to mitigate failures, Gartner says. Leaders must also include application security functionality to identify known vulnerabilities and block attempts to exploit them. Gartner identifies observability leaders This year’s report highlights eight vendors in the leaders category, all of which have demonstrated strong product capabilities, solid technology execution, and innovative strategic vision. Read on to learn what Gartner thinks makes these eight vendors (listed in alphabetical order) stand out as leaders in observability: Chronosphere: Strengths include cost optimization capabilities with its control plane that closely manages the ingestion, storage, and retention of incoming telemetry using granular policy controls. The platform requires no agents and relies largely on open protocols such as OpenTelemetry and Prometheus. Gartner cautions that Chronosphere has not emphasized AI capabilities in its observability platform and currently offers digital experience monitoring via partnerships. Datadog: Strengths include extensive capabilities for managing service-level objectives across data types and providing deep visibility into system and application behavior without the need for instrumentation. Gartner notes the vendor’s licensing

Read More »

LiquidStack CEO Joe Capes on GigaModular, Direct-to-Chip Cooling, and AI’s Thermal Future

In this episode of the Data Center Frontier Show, Editor-in-Chief Matt Vincent speaks with LiquidStack CEO Joe Capes about the company’s breakthrough GigaModular platform — the industry’s first scalable, modular Coolant Distribution Unit (CDU) purpose-built for direct-to-chip liquid cooling. With rack densities accelerating beyond 120 kW and headed toward 600 kW, LiquidStack is targeting the real-world requirements of AI data centers while streamlining complexity and future-proofing thermal design. “AI will keep pushing thermal output to new extremes,” Capes tells DCF. “Data centers need cooling systems that can be easily deployed, managed, and scaled to match heat rejection demands as they rise.” LiquidStack’s new GigaModular CDU, unveiled at the 2025 Datacloud Global Congress in Cannes, delivers up to 10 MW of scalable cooling capacity. It’s designed to support single-phase direct-to-chip liquid cooling — a shift from the company’s earlier two-phase immersion roots — via a skidded modular design with a pay-as-you-grow approach. The platform’s flexibility enables deployments at N, N+1, or N+2 resiliency. “We designed it to be the only CDU our customers will ever need,” Capes says. From Immersion to Direct-to-Chip LiquidStack first built its reputation on two-phase immersion cooling, which Joe Capes describes as “the highest performing, most sustainable cooling technology on Earth.” But with the launch of GigaModular, the company is now expanding into high-density, direct-to-chip cooling, helping hyperscale and colocation providers upgrade their thermal strategies without overhauling entire facilities. “What we’re trying to do with GigaModular is simplify the deployment of liquid cooling at scale — especially for direct-to-chip,” Capes explains. “It’s not just about immersion anymore. The flexibility to support future AI workloads and grow from 2.5 MW to 10 MW of capacity in a modular way is absolutely critical.” GigaModular’s components — including IE5 pump modules, dual BPHx heat exchangers, and intelligent control systems —

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »