Stay Ahead, Stay ONMINE

Out-analyzing analysts: OpenAI’s Deep Research pairs reasoning LLMs with agentic RAG to automate work — and replace jobs

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Enterprise companies need to take note of OpenAI’s Deep Research. It provides a powerful product based on new capabilities, and is so good that it could put a lot of people out of jobs. Deep Research […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Enterprise companies need to take note of OpenAI’s Deep Research. It provides a powerful product based on new capabilities, and is so good that it could put a lot of people out of jobs.

Deep Research is on the bleeding edge of a growing trend: integrating large language models (LLMs) with search engines and other tools to greatly expand their capabilities. (Just as this article was being reported, for example, Elon Musk’s xAI unveiled Grok 3, which claims similar capabilities, including a Deep Search product. However, it’s too early to assess Grok 3’s real-world performance, since most subscribers haven’t actually gotten their hands on it yet.)

OpenAI’s Deep Research, released on February 3, requires a Pro account with OpenAI, costing $200 per month, and is currently available only to U.S. users. So far, this restriction may have limited early feedback from the global developer community, which is typically quick to dissect new AI advancements.

With Deep Research mode, users can ask OpenAI’s leading o3 model any question. The result? A report often superior to what human analysts produce, delivered faster and at a fraction of the cost.

How Deep Research works

While Deep Research has been widely discussed, its broader implications have yet to fully register. Initial reactions praised its impressive research capabilities, despite its occasional hallucinations in its citations. There was the guy who said he used it to help his wife who had breast cancer. It provided deeper analysis than what her oncologists provided on how radiation therapy was the right course of action, he said. The consensus, summarized by Wharton AI professor Ethan Mollick, is that its advantages far outweigh occasional inaccuracies, as fact-checking takes less time than what the AI saves overall. This is something I agree with, based on my own usage.

Financial institutions are already exploring applications. BNY Mellon, for instance, sees potential in using Deep Research for credit risk assessments. Its impact will extend across industries, from healthcare to retail, manufacturing, and supply chain management — virtually any field that relies on knowledge work.

A smarter research agent

Unlike traditional AI models that attempt one-shot answers, Deep Research first asks clarifying questions. It might ask four or more questions to make sure it understands exactly what you want. It then develops a structured research plan, conducts multiple searches, revises its plan based on new insights, and iterates in a loop until it compiles a comprehensive, well-formatted report. This can take between a few minutes and half an hour. Reports range from 1,500 to 20,000 words, and typically include citations from 15 to 30 sources with exact URLs, at least according to my usage over the past week and a half.

The technology behind Deep Research: reasoning LLMs and agentic RAG

Deep Research does this by merging two technologies in a way we haven’t seen before in a mass-market product. 

Reasoning LLMs: The first is OpenAI’s cutting-edge model, o3, which leads in logical reasoning and extended chain-of-thought processes. When it was announced in December 2024, o3 scored an unprecedented 87.5% on the super-difficult ARC-AGI benchmark designed to test novel problem-solving abilities. What’s interesting is that o3 hasn’t been released as a standalone model for developers to use. Indeed, OpenAI’s CEO Sam Altman announced last week that the model instead would be wrapped into a “unified intelligence” system, which would unite models with agentic tools like search, coding agents and more. Deep Research is an example of such a product. And while competitors like DeepSeek-R1 have approached o3’s capabilities (one of the reasons why there was so much excitement a few weeks ago), OpenAI is still widely considered to be slightly ahead.

Agentic RAG: The second, agentic RAG, is a technology that has been around for about a year now. It uses agents ​​to autonomously seek out information and context from other sources, including searching the internet. This can include other tool-calling agents to find non-web information via APIs; coding agents that can complete complex sequences more efficiently; and database searches. Initially, OpenAI’s Deep Research is primarily searching the open web, but company leaders have suggested it would be able to search more sources over time.

OpenAI’s competitive edge (and its limits)

While these technologies are not entirely new, OpenAI’s refinements — enabled by things like its jump-start on working on these technologies, massive funding, and its closed-source development model — have taken Deep Research to a new level. It can work behind closed doors, and leverage feedback from the more than 300 million active users of OpenAI’s popular ChatGPT product. OpenAI has led in research in these areas, for example in how to do verification step by step to get better results. And it has clearly implemented search in an interesting way, perhaps borrowing from Microsoft’s Bing and other technologies.

While it is still hallucinating some results from its searches, it’s doing so less than competitors, perhaps in part because the underlying o3 model itself has set an industry low for these hallucinations at 8%. And there are ways to reduce mistakes still further, by using mechanisms like confidence thresholds, citation requirements and other sophisticated credibility checks

At the same time, there are limits to OpenAI’s lead and capabilities. Within two days of Deep Research’s launch, HuggingFace introduced an open-source AI research agent called Open Deep Research that got results that weren’t too far off of OpenAI’s — similarly merging leading models and freely available agentic capabilities. There are few moats. Open-source competitors like DeepSeek appear set to stay close in the area of reasoning models, and Microsoft’s Magentic-One offers a framework for most of OpenAI’s agentic capabilities, to name just two more examples. 

Furthermore, Deep Research has limitations. The product is really efficient at researching obscure information that can be found on the web. But in areas where there is not much online and where domain expertise is largely private — whether in peoples’ heads or in private databases — it doesn’t work at all. So this isn’t going to threaten the jobs of high-end hedge-fund researchers, for example, who are paid to go talk with real experts in an industry to find out otherwise very hard-to-obtain information, as Ben Thompson argued in a recent post (see graphic below). In most cases, OpenAI’s Deep Research is going to affect lower-skilled analyst jobs. 

Deep Research’s value first increases as information online gets scarce, then drops off when it gets really scarce. Source: Stratechery.

The most intelligent product yet

When you merge top-tier reasoning with agentic retrieval, it’s not really surprising that you get such a powerful product. OpenAI’s Deep Research achieved 26.6% on Humanity’s Last Exam, arguably the best benchmark for intelligence. This is a relatively new AI benchmark designed to be the most difficult for any AI model to complete, covering 3,000 questions across 100 different subjects. On this benchmark, OpenAI’s Deep Research significantly outperforms Perplexity’s Deep Research (20.5%) and earlier models like o3-mini (13%) and DeepSeek-R1 (9.4%) that weren’t hooked up with agentic RAG. But early reviews suggest OpenAI leads in both quality and depth. Google’s Deep Research has yet to be tested against this benchmark, but early reviews suggest OpenAI leads in both quality and depth.

How it’s different: the first mass-market AI that could displace jobs

What’s different with this product is its potential to eliminate jobs. Sam Witteveen, cofounder of Red Dragon and a developer of AI agents, observed in a deep-dive video discussion with me that a lot of people are going to say: “Holy crap, I can get these reports for $200 that I could get from some top-4 consulting company that would cost me $20,000.” This, he said, is going to cause some real changes, including likely putting people out of jobs.

Which brings me back to my interview last week with Sarthak Pattanaik, head of engineering and AI at BNY Mellon, a major U.S. bank.

To be sure, Pattanaik didn’t say anything about the product’s ramifications for actual job counts at his bank. That’s going to be a particularly sensitive topic that any enterprise is probably going to shy away from addressing publicly. But he said he could see OpenAI’s Deep Research being used for credit underwriting reports and other “topline” activities, and having significant impact on a variety of jobs: “Now that doesn’t impact every job, but that does impact a set of jobs around strategy [and] research, like comparison vendor management, comparison of product A versus product B.” He added: “So I think everything which is more on system two thinking — more exploratory, where it may not have a right answer, because the right answer can be mounted once you have that scenario definition — I think that’s an opportunity.”

A historical perspective: job loss and job creation

Technological revolutions have historically displaced workers in the short term while creating new industries in the long run. From automobiles replacing horse-drawn carriages to computers automating clerical work, job markets evolve. New opportunities created by the disruptive technologies tend to spawn new hiring. Companies that fail to embrace these advances will fall behind their competitors.

OpenAI’s Altman acknowledged the link, even if indirect, between Deep Research and labor. At the AI Summit in Paris last week, he was asked about his vision for artificial general intelligence (AGI), or the stage at which AI can perform pretty much any task that a human can. As he answered, his first reference was to Deep Research: “It’s a model I think is capable of doing like a low-single-digit percentage of all the tasks in the economy in the world right now, which is a crazy statement, and a year ago I don’t think something that people thought is going to be coming.” (See minute three of this video). He continued: “For 50 cents of compute, you can do like $500 or $5,000 of work. Companies are implementing that to just be way more efficient.” 

The takeaway: a new era for knowledge work

Deep Research represents a watershed moment for AI in knowledge-based industries. By integrating cutting-edge reasoning with autonomous research capabilities, OpenAI has created a tool that is smarter, faster and significantly more cost-effective than human analysts.

The implications are vast, from financial services to healthcare to enterprise decision-making. Organizations that leverage this technology effectively will gain a significant competitive edge. Those that ignore it do so at their peril.

For a deeper discussion on how OpenAI’s Deep Research works, and how it is reshaping knowledge work, check out my in-depth conversation with Sam Witteveen in our latest video:

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

ExxonMobil bumps up 2030 target for Permian production

ExxonMobil Corp., Houston, is looking to grow production in the Permian basin to about 2.5 MMboe/d by 2030, an increase of 200,000 boe/d from executives’ previous forecasts and a jump of more than 45% from this year’s output. Helping drive that higher target is an expected 2030 cost profile that

Read More »

US Oil Slides to Four Year Low

The US oil benchmark fell to its lowest level since February 2021, with traders weighing renewed signs of optimism surrounding a deal to end the war in Ukraine and mixed economic data from China. West Texas Intermediate settled below $57 a barrel in thin trading ahead of the Christmas and New Year holidays, sliding as stocks wavered. US negotiators offered more substantial security guarantees to Kyiv in a renewed bid to clinch a deal, though the effort still appeared part of a bid to pressure Ukrainian President Volodymyr Zelenskiy on territory. An agreement to end the conflict could lift restrictions on the flows of Russian oil, limiting disruptions in an already well-supplied market. The potentially positive developments in the talks added to earlier bearish momentum on signs of weakness in China’s economy that could limit a key source of demand for crude, outweighing news that the country’s apparent oil demand and refining activity increased in November. Oil is set for an annual loss, with supply set to exceed demand this year and next. Concerns about a glut are showing up in the key Middle Eastern crude market, and trend-following commodity trading advisers were 100% short in both Brent and WTI on Monday, according to data from Bridgeton Research Group. “Crude continues to trade heavy as headlines this morning suggest there’s growing consensus around elements of a potential Russia-Ukraine ceasefire,” said Rebecca Babin, a senior energy trader at CIBC Private Wealth Group. “While a ceasefire wouldn’t trigger a sudden wave of Russian barrels returning to market, it would materially reduce the risk of future supply disruptions.” Still, the fact that some details of a peace deal remain unclear could offer support for prices, Babin said. And there are other geopolitical inputs at play. Even as US-Ukraine talks advanced, Ukraine has intensified

Read More »

Tokyo Gas to Invest in USA Downstream Assets

Tokyo Gas Co., Japan’s biggest distributor of the fuel, plans to invest in US downstream assets to lift earnings and reinforce the last leg of its energy supply chain. The company is looking to deploy capital in assets like liquefaction plants, export terminals and the energy services sector, said Tokyo Gas President Shinichi Sasayama. “We’ve already made investments in midstream, downstream areas such as marketing and trading, and we intend to raise profitability,” he said in an interview.  Tokyo Gas shares rose as much as 2.3% during Monday morning trading hours, while the broader Topix index fell as much as 0.4%.  The firm’s planned expansion in the US comes as President Donald Trump rolls back climate commitments and elevates fossil fuels in national security planning. A surge in power use from artificial intelligence and data centers is boosting demand for gas-fired electricity, creating favorable conditions for energy producers.  Tokyo Gas has allocated 350 billion yen ($2.2 billion) for overseas investments for the next three years starting from fiscal 2026, according to a strategy document released in October. However, a spokesperson declined to say on Friday how much the company has earmarked for downstream expansion in the US. Sasayama said much of that money will go toward developing and making the company’s shale gas assets profitable. Any decision to increase spending on upstream assets will depend on circumstances at the time, he added. Tokyo Gas’ US subsidiary bought Rockcliff Energy II LLC, a Texas natural-gas producer, in late 2023 for about $2.7 billion. It also acquired a stake in gas marketing and trading firm Arm Energy Trading LLC in 2024.  The Japanese utility drew attention last year after activist Elliott Investment Management disclosed a 5% stake. Elliott initially pressured Tokyo Gas to divest parts of its multibillion dollar real estate portfolio and boost shareholder value.  Sasayama said the

Read More »

Finnish Refiner Says It Won’t Meet 2035 Oil Exit Goal

Finland’s Neste Oyj is scaling back some of its climate targets and doesn’t expect to stop using crude at its only oil refinery by 2035 as previously guided.  Neste had pledged a “very ambitious” target for achieve carbon neutral production by that year, but it now expects to reduce greenhouse gas emissions in its own operations — known as Scope 1 and 2 — by 80% by 2040, according to a statement on its website. Neste’s share rose.  Reaching the original climate targets and schedule “would have required significant investments that are currently not realistic,” the company said. “The timeline for transitioning from crude oil to processing renewable and circular raw materials will be determined in line with the actual fuel market demand.” The bulk of Neste’s climate goals were put in place under the company’s previous chief executive officer. Heikki Malinen, who took over in 2024, had already said that a target on using waste plastics in the Porvoo refinery had been put on hold.  Neste’s share advanced as much as 3.9% to €19.03 as of 12:36 p.m. local time. Neste is set to benefit from decisions taken last week by the German government that are expected to boost demand for renewable diesel, one of its main products.  WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate or insulting comments will be removed.

Read More »

Tribal nations regroup after loss of federal funding for clean energy

Tribal nations looking to build clean energy projects are exploring new funding pathways after the Trump administration’s cuts to clean energy grants like Solar for All, which earmarked more than $500 million for solar development on tribal lands.  The Tribal Renewable Energy Coalition, an organization of 14 tribes which received a grant for more than $135 million in now-cancelled Solar for All funding, is continuing to use the solar project plans it developed but will seek funding sources like loans and philanthropy, Indigenized Energy CEO Cody Two Bears said during a November press call. “I think that’s the positive side of what we get out of Solar for All, is we had over a year to plan this out, and now we have some things teed up for tribes to go after and start to leverage some of these dollars,” Two Bears said. “And I think that’s the plan for what we’re going to do moving forward.” After the Trump administration took office and began to cancel grants associated with the Inflation Reduction Act, “the large utility scale projects that we were working on suddenly became a lot more difficult to finance,” said Chéri Smith, CEO of the Alliance for Tribal Clean Energy. “And yet, there are still opportunities there.” Earlier this year, the alliance launched a suite of project finance and planning tools called the Weaver Platform, and is helping tribes “to use that tool to build their capital stacks,” at the same time that they’re “engaging with networks of investors of all flavors to try and take projects and fund them with private dollars instead of public,” Smith said. The Trump administration’s grant clawback has rescinded a great deal of federal funding for clean energy from tribes, Smith said, noting that the alliance’s grant-writing team got $484 million

Read More »

Texas Oil, Gas Upstream Jobs Fall in September

Texas oil and gas upstream jobs fell by 1,300 in September, industry body the Texas Oil & Gas Association (TXOGA) said in a statement sent to Rigzone recently. “The Texas Workforce Commission, having skipped data releases during the federal shutdown that included the Bureau of Labor Statistics, has resumed job data publication,” TXOGA noted in the statement. “Today [December 11] the Commission released September 2025 data, indicating that upstream oil and gas employment fell by 1,300 in September compared to August,” it added. In the statement, TXOGA noted that, “despite recent flat performance”, growth for this year through September “remains a positive 3,900 upstream jobs”. “At 204,800 upstream jobs, compared to the same month in the prior year, September 2025 jobs were up by 1,900, or 0.9 percent,” TXOGA highlighted. A chart included in the statement, which displayed Texas oil and gas upstream job figures from January 2021 to September 2025, showed that, despite the monthly dip from August to September 2025, these job figures in September still stood well above the figures in January 2021.  “The recent downward cycle of the upstream job count confirms Texas is not immune to circumstances facing global oil markets,” TXOGA President Todd Staples warned in the statement. “As a major oil exporter for the United States, the Lone Star State must remain competitive on the worldwide stage,” he added. “To remain the global leader, our industry depends on Texas legislative, regulatory, and business climate certainty that is favorable to investment and job creation even when supply and demand factors present uncertainty and instability,” he continued. The White House website highlights that the U.S. government was recently shut down “for a record 43 days”. August Figures, 2025 Analysis In a statement posted on its site back in September, TXOGA stated that data from the

Read More »

Strategists Say Oil’s Fermi Paradox Nearing an End

In an oil and gas report sent to Rigzone recently by the Macquarie team, Macquarie strategists, including Vikas Dwivedi, noted that oil’s “Fermi Paradox [is] nearing an end”, adding that “onshore stocks [are] starting to build”. “We continue to expect a heavily oversupplied market,” the strategists said in the report. “We estimate a 1Q26 peak supply-demand surplus of over four million barrels per day. Signs of the surplus are showing with continued offshore builds, increasing onshore builds, and extremely strong freight rates,” they added. “We estimate that approximately one-third of the offshore build is long-haul shipments from the Americas to Asia,” they continued. In the report, the strategists revealed that they expect onshore builds to accelerate through year-end 2025 and into early 2026, a process which they said “should drive Brent towards the low $50 range, with a possibility of reaching $45 per barrel”. “Since the end of August, offshore inventories have increased by roughly 250 million barrels and onshore storage up by ~30 million barrels,” the strategists highlighted in the report. “In the past month, the trend has accelerated with onshore … [plus] offshore stocks building by ~ three million barrels per day. Yet, structure remains backwardated, as AB barrels continued clearing East,” they added. A separate report sent to Rigzone by the Macquarie team on December 5 showed that Macquarie was projecting that the Brent price will average $68.21 per barrel overall in 2025 and $60.75 per barrel overall in 2026. According to that report, Macquarie expects the Brent price to average $63.00 per barrel in the fourth quarter of this year, $57.00 per barrel in the first quarter of 2026, $59.00 per barrel in the second quarter, $60.00 per barrel in the third quarter, and $67.00 per barrel in the fourth quarter.   In that report, Macquarie

Read More »

Executive Roundtable: Converging Disciplines in the AI Buildout

At Data Center Frontier, we rely on industry leaders to help us understand the most urgent challenges facing digital infrastructure. And in the fourth quarter of 2025, the data center industry is adjusting to a new kind of complexity.  AI-scale infrastructure is redefining what “mission critical” means, from megawatt density and modular delivery to the chemistry of cooling fluids and the automation of energy systems. Every project has arguably in effect now become an ecosystem challenge, demanding that electrical, mechanical, construction, and environmental disciplines act as one.  For this quarter’s Executive Roundtable, DCF convened subject matter experts from Ecolab, EdgeConneX, Rehlko and Schneider Electric – leaders spanning the full chain of facilities design, deployment, and operation. Their insights illuminate how liquid cooling, energy management, and sustainable process design in data centers are now converging to set the pace for the AI era. Our distinguished executive panelists for this quarter include: Rob Lowe, Director RD&E – Global High Tech, Ecolab Phillip Marangella, Chief Marketing and Product Officer, EdgeConneX Ben Rapp, Manager, Strategic Project Development, Rehlko Joe Reele, Vice President, Datacenter Solution Architects, Schneider Electric Today: Engineering the New Normal – Liquid Cooling at Scale  Today’s kickoff article grapples with how, as liquid cooling technology transitions to default hyperscale design, the challenge is no longer if, but how to scale builds safely, repeatably, and globally.  Cold plates, immersion, dielectric fluids, and liquid-to-chip loops are converging into factory-integrated building blocks, yet variability in chemistry, serviceability, materials, commissioning practices, and long-term maintenance threatens to fragment adoption just as demand accelerates.  Success now hinges on shared standards and tighter collaboration across OEMs, builders, and process specialists worldwide. So how do developers coordinate across the ecosystem to make liquid cooling a safe, maintainable global default? What’s Ahead in the Roundtable Over the coming days, our panel

Read More »

DCF Trends Summit 2025: AI for Good – How Operators, Vendors and Cooling Specialists See the Next Phase of AI Data Centers

At the 2025 Data Center Frontier Trends Summit (Aug. 26-28) in Reston, Va., the conversation around AI and infrastructure moved well past the hype. In a panel sponsored by Schneider Electric—“AI for Good: Building for AI Workloads and Using AI for Smarter Data Centers”—three industry leaders explored what it really means to design, cool and operate the new class of AI “factories,” while also turning AI inward to run those facilities more intelligently. Moderated by Data Center Frontier Editor in Chief Matt Vincent, the session brought together: Steve Carlini, VP, Innovation and Data Center Energy Management Business, Schneider Electric Sudhir Kalra, Chief Data Center Operations Officer, Compass Datacenters Andrew Whitmore, VP of Sales, Motivair Together, they traced both sides of the “AI for Good” equation: building for AI workloads at densities that would have sounded impossible just a few years ago, and using AI itself to reduce risk, improve efficiency and minimize environmental impact. From Bubble Talk to “AI Factories” Carlini opened by acknowledging the volatility surrounding AI investments, citing recent headlines and even Sam Altman’s public use of the word “bubble” to describe the current phase of exuberance. “It’s moving at an incredible pace,” Carlini noted, pointing out that roughly half of all VC money this year has flowed into AI, with more already spent than in all of the previous year. Not every investor will win, he said, and some companies pouring in hundreds of billions may not recoup their capital. But for infrastructure, the signal is clear: the trajectory is up and to the right. GPU generations are cycling faster than ever. Densities are climbing from high double-digits per rack toward hundreds of kilowatts. The hyperscale “AI factories,” as NVIDIA calls them, are scaling to campus capacities measured in gigawatts. Carlini reminded the audience that in 2024,

Read More »

FinOps Foundation sharpens FOCUS to reduce cloud cost chaos

“The big change that’s really started to happen in late 2024 early 2025 is that the FinOps practice started to expand past the cloud,” Storment said. “A lot of organizations got really good at using FinOps to manage the value of cloud, and then their organizations went, ‘oh, hey, we’re living in this happily hybrid state now where we’ve got cloud, SaaS, data center. Can you also apply the FinOps practice to our SaaS? Or can you apply it to our Snowflake? Can you apply it to our data center?’” The FinOps Foundation’s community has grown to approximately 100,000 practitioners. The organization now includes major cloud vendors, hardware providers like Nvidia and AMD, data center operators and data cloud platforms like Snowflake and Databricks. Some 96 of the Fortune 100 now participate in FinOps Foundation programs. The practice itself has shifted in two directions. It has moved left into earlier architectural and design processes, becoming more proactive rather than reactive. It has also moved up organizationally, from director-level cloud management roles to SVP and COO positions managing converged technology portfolios spanning multiple infrastructure types. This expansion has driven the evolution of FOCUS beyond its original cloud billing focus. Enterprises are implementing FOCUS as an internal standard for chargeback reporting even when their providers don’t generate native FOCUS data. Some newer cloud providers, particularly those focused on AI infrastructure, are using the FOCUS specification to define their billing data structures from the ground up rather than retrofitting existing systems. The FOCUS 1.3 release reflects this maturation, addressing technical gaps that have emerged as organizations apply cost management practices across increasingly complex hybrid environments. FOCUS 1.3 exposes cost allocation logic for shared infrastructure The most significant technical enhancement in FOCUS 1.3 addresses a gap in how shared infrastructure costs are allocated and

Read More »

Aetherflux joins the race to launch orbital data centers by 2027

Enterprises will connect to and manage orbital workloads “the same way they manage cloud workloads today,” using optical links, the spokesperson added. The company’s approach is to “continuously launch new hardware and quickly integrate the latest architectures,” with older systems running lower-priority tasks to serve out the full useful lifetime of their high-end GPUs. The company declined to disclose pricing. Aetherflux plans to launch about 30 satellites at a time on SpaceX Falcon 9 rockets. Before the data center launch, the company will launch a power-beaming demonstration satellite in 2026 to test transmission of one kilowatt of energy from orbit to ground stations, using infrared lasers. Competition in the sector has intensified in recent months. In November, Starcloud launched its Starcloud-1 satellite carrying an Nvidia H100 GPU, which is 100 times more powerful than any previous GPU flown in space, according to the company, and demonstrated running Google’s Gemma AI model in orbit. In the same month, Google announced Project Suncatcher, with a 2027 demonstration mission planned. Analysts see limited near-term applications Despite the competitive activity, orbital data centers won’t replace terrestrial cloud regions for general hosting through 2030, said Ashish Banerjee, senior principal analyst at Gartner. Instead, they suit specific workloads, including meeting data sovereignty requirements for jurisdictionally complex scenarios, offering disaster recovery immune to terrestrial risks, and providing asynchronous high-performance computing, he said. “Orbital centers are ideal for high-compute, low-I/O batch jobs,” Banerjee said. “Think molecular folding simulations for pharma, massive Monte Carlo financial simulations, or training specific AI model weights. If the job takes 48 hours, the 500ms latency penalty of LEO is irrelevant.” One immediate application involves processing satellite-generated data in orbit, he said. Earth observation satellites using synthetic aperture radar generate roughly 10 gigabytes per second, but limited downlink bandwidth creates bottlenecks. Processing data in

Read More »

Here’s what Oracle’s soaring infrastructure spend could mean for enterprises

He said he had earlier told analysts in a separate call that margins for AI workloads in these data centers would be in the 30% to 40% range over the life of a customer contract. Kehring reassured that there would be demand for the data centers when they were completed, pointing to Oracle’s increasing remaining performance obligations, or services contracted but not yet delivered, up $68 billion on the previous quarter, saying that Oracle has been seeing unprecedented demand for AI workloads driven by the likes of Meta and Nvidia. Rising debt and margin risks raise flags for CIOs For analysts, though, the swelling debt load is hard to dismiss, even with Oracle’s attempts to de-risk its spend and squeeze more efficiency out of its buildouts. Gogia sees Oracle already under pressure, with the financial ecosystem around the company pricing the risk — one of the largest debts in corporate history, crossing $100 billion even before the capex spend this quarter — evident in the rising cost of insuring the debt and the shift in credit outlook. “The combination of heavy capex, negative free cash flow, increasing financing cost and long-dated revenue commitments forms a structural pressure that will invariably finds its way into the commercial posture of the vendor,” Gogia said, hinting at an “eventual” increase in pricing of the company’s offerings. He was equally unconvinced by Magouyrk’s assurances about the margin profile of AI workloads as he believes that AI infrastructure, particularly GPU-heavy clusters, delivers significantly lower margins in the early years because utilisation takes time to ramp.

Read More »

New Nvidia software gives data centers deeper visibility into GPU thermals and reliability

Addressing the challenge Modern AI accelerators now draw more than 700W per GPU, and multi-GPU nodes can reach 6kW, creating concentrated heat zones, rapid power swings, and a higher risk of interconnect degradation in dense racks, according to Manish Rawat, semiconductor analyst at TechInsights. Traditional cooling methods and static power planning increasingly struggle to keep pace with these loads. “Rich vendor telemetry covering real-time power draw, bandwidth behavior, interconnect health, and airflow patterns shifts operators from reactive monitoring to proactive design,” Rawat said. “It enables thermally aware workload placement, faster adoption of liquid or hybrid cooling, and smarter network layouts that reduce heat-dense traffic clusters.” Rawat added that the software’s fleet-level configuration insights can also help operators catch silent errors caused by mismatched firmware or driver versions. This can improve training reproducibility and strengthen overall fleet stability. “Real-time error and interconnect health data also significantly accelerates root-cause analysis, reducing MTTR and minimizing cluster fragmentation,” Rawat said. These operational pressures can shape budget decisions and infrastructure strategy at the enterprise level.

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »