Stay Ahead, Stay ONMINE

OpenAI returns to open source roots with new models gpt-oss-120b and gpt-oss-20b 

OpenAI is getting back to its roots as an open source AI company with today’s announcement and release of two new, open source, frontier large language models (LLMs): gpt-oss-120b and gpt-oss-20b.The former is a 120-billion parameter model as the name would suggest, capable of running on a single Nvidia H100 graphics processing unit (GPU) and the latter is only 20 billion, small enough to run locally on a consumer laptop or desktop PC. Both are text-only language models, which means unlike the multimodal AI that we’ve had for nearly two years that allows users to upload files and images and have the AI analyze them, users will be confined to only inputting text messages to the models and receiving text back out. However, they can still of course write code and provide math problems and numerics, and in terms of their performance on tasks, they rank above some of OpenAI’s paid models and much of the competition globally.

OpenAI is getting back to its roots as an open source AI company with today’s announcement and release of two new, open source, frontier large language models (LLMs): gpt-oss-120b and gpt-oss-20b.

The former is a 120-billion parameter model as the name would suggest, capable of running on a single Nvidia H100 graphics processing unit (GPU) and the latter is only 20 billion, small enough to run locally on a consumer laptop or desktop PC.

Both are text-only language models, which means unlike the multimodal AI that we’ve had for nearly two years that allows users to upload files and images and have the AI analyze them, users will be confined to only inputting text messages to the models and receiving text back out.

However, they can still of course write code and provide math problems and numerics, and in terms of their performance on tasks, they rank above some of OpenAI’s paid models and much of the competition globally.


The AI Impact Series Returns to San Francisco – August 5

The next phase of AI is here – are you ready? Join leaders from Block, GSK, and SAP for an exclusive look at how autonomous agents are reshaping enterprise workflows – from real-time decision-making to end-to-end automation.

Secure your spot now – space is limited: https://bit.ly/3GuuPLF


They can also be connected to external tools including web search to perform research on behalf of the user. More on this below.

Most importantly: they’re free, they’re available for enterprises and indie developers to download the code and use right now, modifying according to their needs, and can be run locally without a web connection, ensuring maximum privacy, unlike the other top OpenAI models and those from leading U.S.-based rivals Google and Anthropic.

The models can be downloaded today with full weights (the settings guiding its behavior) on the AI code sharing community Hugging Face and GitHub.

High benchmark scores

According to OpenAI, gpt-oss-120b matches or exceeds its proprietary o4-mini model on reasoning and tool-use benchmarks, including competition mathematics (AIME 2024 & 2025), general problem solving (MMLU and HLE), agentic evaluations (TauBench), and health-specific evaluations (HealthBench). The smaller gpt-oss-20b model is comparable to o3-mini and even surpasses it in some benchmarks.

The models are multilingual and perform well across a variety of non-English languages, though OpenAI declined to specify which and how many.

While these capabilities are available out of the box, OpenAI notes that localized fine-tuning — such as an ongoing collaboration with the Swedish government to produce a version fine-tuned on the country’s language —can still meaningfully enhance performance for specific regional or linguistic contexts.

A hugely advantageous license for enterprises and privacy-minded users

But the biggest feature is the licensing terms for both: Apache 2.0, the same as the wave of Chinese open source models that have been released over the last several weeks, and a more enterprise-friendly license than Meta’s trickier and more nuanced open-ish Llama license, which requires that users who operate a service with more than 700 million monthly active users obtain a paid license to keep using the company’s family of LLMs.

By contrast, OpenAI’s new gpt-oss series of models offer no such restrictions. In keeping with Chinese competitors and counterparts, any consumer, developer, independent entrepreneur or enterprise large and small is empowered by the Apache 2.0 license to be able to download the new gpt-oss models at will, fine-tune and alter them to fit their specific needs, and use them to generate revenue or operate paid services, all without paying OpenAI a dime (or anything!).

This also means enterprises can use a powerful, near topline OpenAI model on their own hardware totally privately and securely, without sending any data up to the cloud, on web servers, or anywhere else. For highly regulated industries like finance, healthcare, and legal services, not to mention organizations in military, intelligence, and government, this may be a requirement.

Before today, anyone using ChatGPT or its application programming interface (API) — the service that acts like a switching board and allows third-party software developers to connect their own apps and services to these OpenAI’s proprietary/paid models like GPT-4o and o3 — was sending data up to OpenAI servers that could technically be subpoenaed by government agencies and accessed without a user’s knowledge. That’s still the case for anyone using ChatGPT or the API going forward, as OpenAI co-founder and Sam Altman recently warned.

And while running the new gpt-oss models locally on a user’s own hardware disconnected from the web would allow for maximum privacy, as soon as the user decides to connect it to external web search or other web enabled tools, some of the same privacy risks and issues would then arise — through any third-party web services the user or developer was relying on when hooking the models up to said tools.

The last OpenAI open source language model was released more than six years ago

“This is the first time we’re releasing an open-weight language model in a long time… We view this as complementary to our other products,” said OpenAI co-founder and president Greg Brockman on an embargoed press video call with VentureBeat and other journalists last night.

The last time OpenAI released a fully open source language model was GPT-2 in 2019, more than six years ago, and three years before the release of ChatGPT.

This fact has sparked the ire of — and resulted in several lawsuits fromformer OpenAI co-founder and backer turned rival Elon Musk, who, along with many other critics, have spent the last several years accusing OpenAI of betraying its mission and founding principles and namesake by eschewing open source AI releases in favor of paid proprietary models available only to customers of OpenAI’s API or paying ChatGPT subscribers (though there is a free tier for the latter).

OpenAI co-founder CEO Sam Altman did express regret about being on the “wrong side of history” but not releasing more open source AI sooner in a Reddit AMA (ask me anything) QA with users in February of this year, and Altman committed to releasing a new open source model back in March, but ultimately the company delayed its release from a planned July date until now.

Now OpenAI is tacking back toward open source, and the question is, why?

Why would OpenAI release a set of free open source models that it makes no money from?

To paraphrase Jesse Plemons’ character’s memorable line from the film Game Night: “How can that be profitable for OpenAI?”

After all, business to OpenAI’s paid offerings appears to be booming.

Revenue has skyrocketed alongside the rapid expansion of its ChatGPT user base, now at 700 million weekly active users. As of August 2025, OpenAI reported $13 billion in annual recurring revenue, up from $10 billion in June. That growth is driven by a sharp rise in paying business customers — now 5 million, up from 3 million just two months earlier — and surging daily engagement, with over 3 billion user messages sent every day.

The financial momentum follows an $8.3 billion funding round that valued OpenAI at $300 billion and provides the foundation for the company’s aggressive infrastructure expansion and global ambitions.

Compare that to closed/proprietary rival AI startup Anthropic’s reported $5 billion in total annual recurring revenue, but interestingly, Anthropic is said to be getting more money from its API, $3.1 billion in revenue compared to OpenAI’s $2.9 billion, according to The Information.

So, given how well the paid AI business is already doing, the business strategy behind these open source offerings is less clear — especially since the new OpenAI gpt-oss models will almost certainly cut into some (perhaps a lot of) usage of OpenAI’s paid models. Why go back to offering open source LLMs now when so much money is flowing into paid and none will, by virtue of its very intent, go directly toward open source models?

Put simply: because open source competitors, beginning with the release of the impressively efficient DeepSeek R1 by the Chinese AI division of the same name in January 2025, are offering near parity on performance benchmarks to paid proprietary models, for free, with fewer (basically zero) implementation restrictions for enterprises and end users. And increasingly, enterprises are adopting these open source models in production.

As OpenAI executives and team members revealed to VentureBeat and many other journalists on an embargoed video call last night about the new models that when it comes to OpenAI’s API, the majority of customers are using a mix of paid OpenAI models and open source models from other providers. (I asked, but OpenAI declined to specify what percentage or total number of API customers are using open source models and which ones).

At least, until now. OpenAI clearly hopes these new gpt-oss offerings will get more of these users to switch away from competing open source offerings and back into OpenAI’s ecosystem, even if OpenAI doesn’t see any direct revenue or data from that usage.

On a grander scale, it seems OpenAI wants to be a full-service, full-stack, one-stop shop AI offering for all of an enterprise, indie developer’s, or regular consumer’s machine intelligence needs — from a clean chatbot interface to an API to build services and apps atop of to agent frameworks for building AI agents through said API to an image generation model (gpt-4o native image generation), video model (Sora), audio transcription model (gpt-4o-transcribe), and now, open source offerings as well. Can a music generation and “world model” be far behind?

OpenAI seeks to span the AI market, propriety and open source alike, even if the latter is worth nothing in terms of actual, direct dollars and cents.

Training and architecture

Feedback from developers directly influenced gpt-oss’s design. OpenAI says the top request was for a permissive license, which led to the adoption of Apache 2.0 for both models. Both models use a Mixture-of-Experts (MoE) architecture with a Transformer backbone.

The larger gpt-oss-120b activates 5.1 billion parameters per token (out of 117 billion total), and gpt-oss-20b activates 3.6 billion (out of 21 billion total).

Both support 128,000 token context length (about 300-400 pages of a novel’s worth of text a user can upload at once), and employ locally banded sparse attention and use Rotary Positional Embeddings for encoding.

The tokenizer — the program that converts words and chunks of words into the numerical tokens that the LLMs can understand, dubbed “o200k_harmony — is also being open-sourced.

Developers can select among low, medium, or high reasoning effort settings based on latency and performance needs. While these models can reason across complex agentic tasks, OpenAI emphasizes they were not trained with direct supervision of CoT outputs, to preserve the observability of reasoning behavior—an approach OpenAI considers important for safety monitoring.

Another common request from OpenAI’s developer community was for strong support for function calling, particularly for agentic workloads, which OpenAI believes gpt-oss now delivers.

The models are engineered for chain-of-thought reasoning, tool use, and few-shot function calling, and are compatible with OpenAI’s Responses API introduced back in March, which allows developers to augment their apps by connecting an OpenAI LLM of their choice to three powerful built-in tools — web search, file search, and computer use — within a single API call.

But for the new gpt-oss models, tool use capabilities — including web search and code execution — are not tied to OpenAI infrastructure. OpenAI provides the schemas and examples used during training, such as a basic browser implementation using the Exa API and a Python interpreter that operates in a Docker container.

It is up to individual inference providers or developers to define how tools are implemented. Providers like vLLM, for instance, allow users to configure their own MCP (Model-Controller-Proxy) server to specify the browser backend.

While these models can reason across complex agentic tasks, OpenAI emphasizes they were not trained with direct supervision of CoT outputs, to preserve the observability of reasoning behavior—an approach OpenAI considers important for safety monitoring.

Safety evaluations and measures

OpenAI conducted safety training using its Preparedness Framework, a document that outlines the procedural commitments, risk‑assessment criteria, capability categories, thresholds, evaluations, and governance mechanisms OpenAI uses to monitor, evaluate, and mitigate frontier AI risks.

These included filtering chemical, biological, radiological, and nuclear threat (CBRN) related data out during pretraining, and applying advanced post-training safety methods such as deliberative alignment and an instruction hierarchy to enforce refusal behavior on harmful prompts.

To test worst-case misuse potential, OpenAI adversarially fine-tuned gpt-oss-120b on sensitive biology and cybersecurity data using its internal RL training stack. These malicious fine-tuning (MFT) scenarios—one of the most sophisticated evaluations of this kind to date—included enabling browsing and disabling refusal behavior, simulating real-world attack potential.

The resulting models were benchmarked against both open and proprietary LLMs, including DeepSeek R1-0528, Qwen 3 Thinking, Kimi K2, and OpenAI’s o3. Despite enhanced access to tools and targeted training, OpenAI found that even the fine-tuned gpt-oss models remained below the “High” capability threshold for frontier risk domains such as biorisk and cybersecurity. These conclusions were reviewed by three independent expert groups, whose recommendations were incorporated into the final methodology.

In parallel, OpenAI partnered with SecureBio to run external evaluations on biology-focused benchmarks like Human Pathogen Capabilities Test (HPCT), Molecular Biology Capabilities Test (MBCT), and others. Results showed that gpt-oss’s fine-tuned models performed close to OpenAI’s o3 model, which is not classified as frontier-high under OpenAI’s safety definitions.

According to OpenAI, these findings contributed directly to the decision to release gpt-oss openly. The release is also intended to support safety research, especially around monitoring and controlling open-weight models in complex domains.

Availability and ecosystem support

The gpt-oss models are now available on Hugging Face, with pre-built support through major deployment platforms including Azure, AWS, Databricks, Cloudflare, Vercel, Together AI, OpenRouter, and others. Hardware partners include NVIDIA, AMD, and Cerebras, and Microsoft is making GPU-optimized builds available on Windows via ONNX Runtime.

OpenAI has also announced a $500,000 Red Teaming Challenge hosted on Kaggle, inviting researchers and developers to probe the limits of gpt-oss and identify novel misuse pathways. A public report and an open-source evaluation dataset will follow, aiming to accelerate open model safety research across the AI community.

Early adopters such as AI Sweden, Orange, and Snowflake have collaborated with OpenAI to explore deployments ranging from localized fine-tuning to secure on-premise use cases. OpenAI characterizes the launch as an invitation for developers, enterprises, and governments to run state-of-the-art language models on their own terms.

While OpenAI has not committed to a fixed cadence for future open-weight releases, it signals that gpt-oss represents a strategic expansion of its approach — balancing openness with aligned safety methodologies to shape how large models are shared and governed in the years ahead.

The big question: with so much competition in open source AI, will OpenAI’s own efforts pay off?

OpenAI re-enters the open source model market in the most competitive moment yet.

At the top of public AI benchmarking leaderboards, U.S. frontier models remain proprietaryOpenAI (GPT-4o/o3), Google (Gemini), and Anthropic (Claude).

But they now compete directly with a surge of open-weights contenders. From China: DeepSeek-R1 (open source, MIT) and DeepSeek-V3 (open-weights under a DeepSeek Model License that permits commercial use); Alibaba’s Qwen 3 (open-weights, Apache-2.0); MoonshotAI’s Kimi K2 (open-weights; public repo and model cards); and Z.ai’s GLM-4.5 (also Apache 2.0 licensed).

Europe’s Mistral (Mixtral/Mistral, open-weights, Apache-2.0) anchors the EU push; the UAE’s Falcon 2/3 publish open-weights under TII’s Apache-based license. In the U.S. open-weights camp, Meta’s Llama 3.1 ships under a community (source-available) license, Google’s Gemma under Gemma terms (open weights with use restrictions), and Microsoft’s Phi-3.5 under MIT.

Developer pull mirrors that split. On Hugging Face, Qwen2.5-7B-Instruct (open-weights, Apache-2.0) sits near the top by “downloads last month,” while DeepSeek-R1 (MIT) and DeepSeek-V3 (model-licensed open weights) also post heavy traction. Open-weights stalwarts Mistral-7B / Mixtral (Apache-2.0), Llama-3.1-8B/70B (Meta community license), Gemma-2 (Gemma terms), Phi-3.5 (MIT), GLM-4.5 (open-weights), and Falcon-2-11B (TII Falcon License 2.0) round out the most-pulled families —underscoring that the open ecosystem spans the U.S., Europe, the Middle East, and China. Hugging Face signals adoption, not market share, but they show where builders are experimenting and deploying today. 

Consumer usage remains concentrated in proprietary apps even as weights open up. ChatGPT still drives the largest engagement globally (about 2.5 billion prompts/day, proprietary service), while in China the leading assistants — ByteDance’s Doubao, DeepSeek’s app, Moonshot’s Kimi, and Baidu’s ERNIE Bot — are delivered as proprietary products, even as several base models (GLM-4.5, ERNIE 4.5 variants) now ship as open-weights.

But now that a range of powerful open source models are available to businesses and consumers — all nearing one another in terms of performance — and can be downloaded on consumer hardware, the big question facing OpenAI is: who will pay for intelligence at all? Will the convenience of the web-based chatbot interface, multimodal capabilities, and more powerful performance be enough to keep the dollars flowing? Or has machine intelligence already become, in the words of Atlman himself, “too cheap to meter”? And if so, how to build a successful business atop it, especially with OpenAI and other AI firms’ sky-high valuations and expenditures.

One clue: OpenAI is already said to be offering in-house engineers to help its enterprise customers customize and deploy fine-tuned models, similar to Palantir’s “forward deployed” software engineers (SWEs), essentially charging for experts to come in, set up the models correctly, and train employees how to use them for best results.

Perhaps the world will migrate toward a majority of AI usage going to open source models, or a sizeable minority, with OpenAI and other AI model providers offering experts to help install said models into enterprises. Is that enough of a service to build a multi-billion dollar business upon? Or will enough people continue paying $20, $200 or more each month to have access to even more powerful proprietary models?

I don’t envy the folks at OpenAI figuring out all the business calculations — despite what I assume to be hefty compensation as a result, at least for now. But for end users and enterprises, the release of the gpt-oss series is undoubtedly compelling.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Mapping Trump’s tariffs by trade balance and geography

U.S. importers may soon see costs rise for many imported goods, as tariffs on foreign goods are set to rise. On July 31, President Donald Trump announced country-specific reciprocal tariffs would finally be implemented on Aug. 7, after a monthslong pause. The news means more than 90 countries will see

Read More »

JF Expands in Southwest with Maverick Acquisition

The JF Group (JF) has acquired Arizona-based Maverick Petroleum Services. JF, a fueling infrastructure, petroleum equipment distribution, service, general contracting, and construction services provider, said in a media release that Maverick brings expertise in the installation, maintenance, and repair of petroleum handling equipment, Point-of-Sale (POS) systems, and environmental testing. As

Read More »

HPE unveils AI-powered network security and data protection technology

Also announced at Black Hat is the HPE Zerto integration hub, which the company says will simplify system security and disaster recovery. The integration hub connects HPE Zerto with cybersecurity software and enterprise networking devices to generate insights, automate workflows, and simplify data protections, HPE says. “HPE Zerto will streamline

Read More »

Energy Transfer to Build $5.3B TX to AZ Natural Gas Pipeline

Energy Transfer LP is moving forward with plans to build a $5.3 billion natural gas pipeline from the Permian Basin to markets in New Mexico and Arizona.  The expansion of the company’s Transwestern Pipeline networks that’s slated to begin operations in late 2029 will provide an outlet for Permian gas producers that have struggled with a dearth of shipping capacity. The project, which Energy Transfer said already has lined up shipping commitments from unidentified companies, may be a death knell for a Kinder Morgan Inc. proposal to build a conduit along a similar route. Kinder Morgan fell as much as 4.6%, making it the worst performer of the day in the S&P 500 Energy Index. Energy Transfer rose 1.8% to $18.06 at 10:01 a.m. in New York.    The boom in Permian Basin shale drilling that unleashed vast outflows of crude oil has also resulted in gluts of so-called associated gas so severe that local prices for the fuel have on occasion turned negative.  Energy Transfer’s project will involve more than 500 miles (800 kilometers) of 42-inch pipe with a capacity of 1.5 billion cubic feet per day, the company said in a statement Wednesday. It also will include nine compressor stations in Arizona, New Mexico and Texas. WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate or insulting comments will be removed.

Read More »

Diamondback takes another $100 million out of 2025’s capex plan

Executives at Diamondback Energy Inc., Midland, have cut the company’s 2025 capital spending plans by about $100 million, building on a $400 million cut three months ago as they prioritize keeping oil volumes flat in an environment that continues to be volatile. Newly named chief executive officer Kaes Van’t Hof and his team are now forecasting that Diamondback will spend between $3.4 billion and $3.6 billion this year. The lower end of that range is the same as in May but the higher bound is $200 million lower, which has trimmed the midpoint of the range to $3.5 billion from $3.6 billion (and $4.0 billion early this year). “It’s certainly hard for me to get extremely bullish today,” Van’t Hof told analysts and investors about crude oil prices on an Aug. 5 conference call discussing the company’s second-quarter results. “That’s why I think 2025 for us is a year of debt reduction and share count reduction—waiting for that spring to coil when commodity prices do rally.” Diamondback produced 495,700 b/d of oil (and 919,900 boe/d in total) in the second quarter from its operations in the Midland and Delaware basins. Its teams completed 116 gross wells (109 net) during the quarter, down about 5% from its Q1 pace, as they reduced their rig count to 13 from 17. Thanks to higher production and capture of NGLs, executives have lifted the company’s full-year total production forecast by about 2% to between 890,000 and 910,000 boe/d. Despite the further retrenching on capex, Van’t Hof said there’s less uncertainty in the market today than in early May and added that “it seems that the double whammy of a demand shock and a supply shock has dissipated for now.” He also said Diamondback’s drilling and completion work is, like many of its peers, becoming

Read More »

Energy Transfer to build 1.5-bcfd Desert Southwest pipeline

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: #c19a06; } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style { font-size: 0.75rem; opacity: .6; } #onetrust-pc-sdk [id*=btn-handler], #onetrust-pc-sdk [class*=btn-handler] { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-policy a, #onetrust-pc-sdk a, #ot-pc-content a { color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-pc-sdk .ot-active-menu { border-color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-accept-btn-handler, #onetrust-banner-sdk #onetrust-reject-all-handler, #onetrust-consent-sdk #onetrust-pc-btn-handler.cookie-setting-link { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-consent-sdk .onetrust-pc-btn-handler { color: #c19a06 !important; border-color: #c19a06 !important; background-color: undefined !important; }

Read More »

Devon again trims ’25 capex plan by $100 million

Devon Energy Corp., Oklahoma City, has reduced its planned capital spending by $100 million for the second consecutive quarter, with executives joining several of their peers in saying that their teams are working ever more efficiently. Devon’s assets in the Delaware, Rockies, Eagle Ford and Anadarko basins required $932 million in capital during second-quarter 2025, which was 7% less than president and chief executive Clay Gaspar and his team had planned for. Over the past 2 years, executives pointed out, Devon’s average drilling cost in the Delaware has fallen 12% and its average completion spending is down 15%. “We are optimizing well performance, reducing cycle times, and streamlining field operations all while delivering production performance,” Gaspar said on an Aug. 6 conference call. “These are sustainable, structural gains.” For the year, Devon’s leaders now expect the company’s capital spending to be between $3.6 billion and $3.8 billion. When they first laid out their 2025 plans last November, the midpoint of their range was $4.1 billion. The capex reduction—which is of the same magnitude as recent moves by both Diamondback Energy Corp. and Expand Energy Corp.—comes after a second quarter in which Devon averaged 387,000 b/d of oil production and 841,000 boe/d of total production while averaging 21 rigs and six completion crews. The oil output was up 15% from last year’s spring quarter and oil-equivalent production was up 19% thanks in part to sizable natural gas and natural gas liquids increases from the company’s Permian operations.

Read More »

Imperial’s Strathcona refinery begins renewable fuels production

ExxonMobil Corp.’s majority owned affiliate Imperial Oil Ltd. has formally commissioned its long-planned  grassroots renewable diesel production complex at the operator’s 196,000-b/d Strathcona refinery near Edmonton, Alta., in western Canada. Following recent completion of construction and initial start of operations in July 2025, the Strathcona refinery’s new plant is now producing renewable diesel, Imperial said on Aug. 5. First announced in 2021 and approved for final investment decision in January 2023, the Strathcona renewable diesel complex will become the largest of its kind in Canada upon reaching full-nameplate production capacity of 20,000 b/d, the operator said. While Imperial did not reveal details regarding the complex’s current production rates, the company did confirm it is already supplying renewable diesel from the plant to customers in western Canada and Imperial’s own operations in northern Alberta. Part of Imperial’s commitment to bringing lower-emission fuels to market and making them accessible to Canada’s hard-to-decarbonize industries, the renewable diesel complex uses an unidentified proprietary technology that combines a mix of locally sourced renewable feedstocks such as canola oil and blue hydrogen (hydrogen produced from natural gas with carbon capture and storage technology) received from Air Products Inc. in the production process (OGJ Online, Sept. 6, 2022). Upon revealing the project in 2021, Imperial said production from the renewable diesel complex at full operation would help reduce greenhouse gas emissions from Canada’s transportation sector by about 3 million tonnes/year, the equivalent of removing more than 650,000 passenger vehicles/year from the road (OGJ Online, Aug. 25, 2021). Imperial reiterated the new complex was made possible because of beneficial government policy and support, which included incentives from Strathcona County and the governments of Alberta and British Columbia, the latter of which supported the project in the form of credits under its provincial low carbon fuel standard (BC LCFS).

Read More »

Oil Slips as Putin Talks Loom

Oil fell for the fifth straight session as traders waited to see whether US President Donald Trump would impose more severe measures to restrict Russian energy flows. Crude extended losses in late trading on a report that Trump plans to meet with Russian President Vladimir Putin soon. West Texas Intermediate dipped 1.2% to settle just above $64 a barrel, notching the longest daily losing streak since September, as traders positioned for the possibility of a softer stance on Russia than the White House previously telegraphed. Futures dipped further after the New York Times reported that Trump plans to meet with Putin in person as soon as next week and follow that up with a meeting between himself, Putin and Ukrainian President Volodymyr Zelenskiy. Trump earlier highlighted “great progress made” in a meeting between his envoy and Putin, but didn’t rule out the possibility of further penalties on Moscow’s oil revenues. The developments injected further uncertainty into a market that already was grappling with mixed signals on supply and demand. On the one hand, US crude inventories fell as domestic refiners run at the highest levels since 2019 seasonally. On the other, Trump has said repeatedly that he wants to see lower oil prices and an end to the war in Ukraine. In a CNBC interview Tuesday, he said that “if energy goes down low enough, Putin’s going to stop killing people.” “If you get energy down another $10 a barrel, he’s going to have no choice, because his economy stinks,” Trump added. The White House announced an additional 25% tariff on India over its purchases of Russian energy this week. Still, the levies on India won’t come into effect for another 21 days, meaning traders will remain on edge in case Trump makes further moves. US special envoy Steve Witkoff

Read More »

Incentivizing the Digital Future: Inside America’s Race to Attract Data Centers

Across the United States, states are rolling out a wave of new tax incentives aimed squarely at attracting data centers, one of the country’s fastest-growing industries. Once clustered in only a handful of industry-friendly regions, today’s data-center boom is rapidly spreading, pushed along by profound shifts in federal policy, surging demand for artificial intelligence, and the drive toward digital transformation across every sector of the economy. Nowhere is this transformation more visible than in the intensifying state-by-state competition to land massive infrastructure investments, advanced technology jobs, and the alluring prospect of long-term economic growth. The past year alone has seen a record number of states introducing or expanding incentives for data centers, from tax credits to expedited permitting, reflecting a new era of proactive, tech-focused economic development policy. Behind these moves, federal initiatives and funding packages underscore the essential role of digital infrastructure as a national priority, encouraging states to lower barriers for data center construction and operation. As states watch their neighbors reap direct investment and job creation benefits, a real “domino effect” emerges: one state’s success becomes another’s blueprint, heightening the pressure and urgency to compete. Yet, this wave of incentives also exposes deeper questions about the local impact, community costs, and the evolving relationship between public policy and the tech industry. From federal levels to town halls, there are notable shifts in both opportunities and challenges shaping the landscape of digital infrastructure advancement. Industry Drivers: the Federal Push and Growth of AI The past year has witnessed a profound federal policy shift aimed squarely at accelerating U.S. digital infrastructure, especially for data centers in direct response both to the explosive growth of artificial intelligence and to intensifying international competition. In July 2025, the administration unveiled “America’s AI Action Plan,” accompanied by multiple executive orders that collectively redefined

Read More »

AI Supercharges Hyperscale: Capacity, Geography, and Design Are Being Redrawn

From Cloud to GenAI, Hyperscalers Cement Role as Backbone of Global Infrastructure Data center capacity is undergoing a major shift toward hyperscale operators, which now control 44 percent of global capacity, according to Synergy Research Group. Non-hyperscale colocations account for another 22 percent of capacity and is expected to continue, but hyperscalers projected to hold 61 percent of the capacity by 2030. That swing also reflects the dominance of hyperscalers geographically. In a separate Synergy study revealing the world’s top 20 hyperscale data center locations, just 20 U.S. state or metro markets account for 62 percent of the world’s hyperscale capacity.  Northern Virginia and the Greater Beijing areas alone make up 20 percent of the total. They’re followed by the U.S. states of Oregon and Iowa, Dublin, the U.S. state of Ohio, Dallas, and then Shanghai. Of the top 20 markets, 14 are in the U.S., five in APAC region, and only one is in Europe. This rapid shift is fueled by the explosive growth of cloud computing, artificial intelligence (AI), and especially generative AI (GenAI)—power-intensive technologies that demand the scale, efficiency, and specialized infrastructure only hyperscalers can deliver. What’s Coming for Capacity The capacity research shows on-premises data centers with 34 percent of the total capacity, a significant drop from the 56 percent capacity they accounted for just six years ago.  Synergy projects that by 2030, hyperscale operators such as Google Cloud, Amazon Web Services, and Microsoft Azure will claim 61 percent of all capacity, while on-premises share will drop to just 22 percent. So, it appears on-premises data centers are both increasing and decreasing. That’s one way to put it, but it’s about perspective. Synergy’s capacity study indicates they’re growing as the volume of enterprise GPU servers increases. The shrinkage refers to share of the market: Hyperscalers are growing

Read More »

In crowded observability market, Gartner calls out AI capabilities, cost optimization, DevOps integration

Support for OpenTelemetry and open standards is another differentiator for Gartner. Vendors that embrace these frameworks are better positioned to offer extensibility, avoid vendor lock-in, and enable broader ecosystem integration. This openness is paired with a growing focus on cost optimization—an increasingly important concern as telemetry data volumes increase. Leaders offer granular data retention controls, tiered storage, and usage-based pricing models to help customers Gartner also highlights the importance of the developer experience and DevOps integration. Observability leaders provide “integration with other operations, service management, and software development technologies, such as IT service management (ITSM), configuration management databases (CMDB), event and incident response management, orchestration and automation, and DevOps tools.” On the automation front, observability platforms should support initiating changes to application and infrastructure code to optimize cost, capacity or performance—or to take corrective action to mitigate failures, Gartner says. Leaders must also include application security functionality to identify known vulnerabilities and block attempts to exploit them. Gartner identifies observability leaders This year’s report highlights eight vendors in the leaders category, all of which have demonstrated strong product capabilities, solid technology execution, and innovative strategic vision. Read on to learn what Gartner thinks makes these eight vendors (listed in alphabetical order) stand out as leaders in observability: Chronosphere: Strengths include cost optimization capabilities with its control plane that closely manages the ingestion, storage, and retention of incoming telemetry using granular policy controls. The platform requires no agents and relies largely on open protocols such as OpenTelemetry and Prometheus. Gartner cautions that Chronosphere has not emphasized AI capabilities in its observability platform and currently offers digital experience monitoring via partnerships. Datadog: Strengths include extensive capabilities for managing service-level objectives across data types and providing deep visibility into system and application behavior without the need for instrumentation. Gartner notes the vendor’s licensing

Read More »

LiquidStack CEO Joe Capes on GigaModular, Direct-to-Chip Cooling, and AI’s Thermal Future

In this episode of the Data Center Frontier Show, Editor-in-Chief Matt Vincent speaks with LiquidStack CEO Joe Capes about the company’s breakthrough GigaModular platform — the industry’s first scalable, modular Coolant Distribution Unit (CDU) purpose-built for direct-to-chip liquid cooling. With rack densities accelerating beyond 120 kW and headed toward 600 kW, LiquidStack is targeting the real-world requirements of AI data centers while streamlining complexity and future-proofing thermal design. “AI will keep pushing thermal output to new extremes,” Capes tells DCF. “Data centers need cooling systems that can be easily deployed, managed, and scaled to match heat rejection demands as they rise.” LiquidStack’s new GigaModular CDU, unveiled at the 2025 Datacloud Global Congress in Cannes, delivers up to 10 MW of scalable cooling capacity. It’s designed to support single-phase direct-to-chip liquid cooling — a shift from the company’s earlier two-phase immersion roots — via a skidded modular design with a pay-as-you-grow approach. The platform’s flexibility enables deployments at N, N+1, or N+2 resiliency. “We designed it to be the only CDU our customers will ever need,” Capes says. From Immersion to Direct-to-Chip LiquidStack first built its reputation on two-phase immersion cooling, which Joe Capes describes as “the highest performing, most sustainable cooling technology on Earth.” But with the launch of GigaModular, the company is now expanding into high-density, direct-to-chip cooling, helping hyperscale and colocation providers upgrade their thermal strategies without overhauling entire facilities. “What we’re trying to do with GigaModular is simplify the deployment of liquid cooling at scale — especially for direct-to-chip,” Capes explains. “It’s not just about immersion anymore. The flexibility to support future AI workloads and grow from 2.5 MW to 10 MW of capacity in a modular way is absolutely critical.” GigaModular’s components — including IE5 pump modules, dual BPHx heat exchangers, and intelligent control systems —

Read More »

Oracle’s Global AI Infrastructure Strategy Takes Shape with Bloom Energy and Digital Realty

Bloom Energy: A Leading Force in On-Site Power As of mid‑2025, Bloom Energy has deployed over 400 MW of capacity at data centers worldwide, working with partners including Equinix, American Electric Power (AEP), and Quanta Computing. In total, Bloom has delivered more than 1.5 GW of power across 1,200+ global installations, a tripling of its customer base in recent years. Several key partnerships have driven this rapid adoption. A decade-long collaboration with Equinix, for instance, began with a 1 MW pilot in 2015 and has since expanded to more than 100 MW deployed across 19 IBX data centers in six U.S. states, providing supplemental power at scale. Even public utilities are leaning in: in late 2024, AEP signed a deal to procure up to 1 GW of Bloom’s solid oxide fuel cell (SOFC) systems for fast-track deployments aimed at large data centers and commercial users facing grid connection delays. More recently, on July 24, 2025, Bloom and Oracle Cloud Infrastructure (OCI) announced a strategic partnership to deploy SOFC systems at select U.S. Oracle data centers. The deployments are designed to support OCI’s gigawatt-scale AI infrastructure, delivering clean, uninterrupted electricity for high-density compute workloads. Bloom has committed to providing sufficient on-site power to fully support an entire data center within 90 days of contract signing. With scalable, modular, and low-emissions energy solutions, Bloom Energy has emerged as a key enabler of next-generation data center growth. Through its strategic partnerships with Oracle, Equinix, and AEP, and backed by a rapidly expanding global footprint, Bloom is well-positioned to meet the escalating demand for multi-gigawatt on-site generation as the AI era accelerates. Oracle and Digital Realty: Accelerating the AI Stack Oracle, which continues to trail hyperscale cloud providers like Google, AWS, and Microsoft in overall market share, is clearly betting big on AI to drive its next phase of infrastructure growth.

Read More »

From Brownfield to Breakthrough: Aligned Data Centers Extends Its AI-First Infrastructure Vision from Ohio to the Edge of Innovation

In an AI-driven world of exponential compute demand, Aligned Data Centers is meeting the moment not just with scale, but with intent. The company’s recent blitz of strategic announcements, led by plans for a transformative new campus on legacy industrial land in Ohio, offers a composite image of what it means to build data center infrastructure for the AI era: rapid, resilient, regionally targeted, and relentlessly sustainable. From converting a former coal power plant site into a hub for digital progress in Coshocton County, to achieving new heights of energy efficiency in Phoenix, to enabling liquid-cooled, NVIDIA-accelerated AI deployments with Lambda in Dallas, Aligned is assembling a modular, AI-optimized framework designed to meet both today’s and tomorrow’s computational extremes. Ohio Expansion: A New Chapter for Conesville, and for Aligned Announced July 24, Aligned’s newest mega-scale data center campus in Central Ohio will rise on a 197-acre parcel adjacent to the retired AEP Conesville coal-fired power plant, a brownfield site that once symbolized legacy energy and is now poised to power the future of digital infrastructure. As noted by Andrew Schaap, CEO of Aligned Data Centers: “Through this strategic expansion, Aligned not only reinforces its commitment to providing future-ready digital infrastructure in vital growth markets but also directly catalyzes billions of dollars in investment for the state of Ohio and the Coshocton County community.” It’s a project with deep regional implications. The phased, multi-billion dollar development is expected to create thousands of construction jobs and hundreds of high-quality, long-term operational roles, while generating significant tax revenues that will support local services and infrastructure improvements. The campus has already secured a foundational customer, with the first facility targeting initial capacity delivery in mid-2026. This marks Aligned’s third campus in Ohio, a clear indication that the company sees the Buckeye State, with its

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »