Stay Ahead, Stay ONMINE

Deep Data Center: Neoclouds as the ‘Picks and Shovels’ of the AI Gold Rush

In 1849, the discovery of gold in California ignited a frenzy, drawing prospectors from around the world in pursuit of quick fortune. While few struck it rich digging and sifting dirt, a different class of entrepreneurs quietly prospered: those who supplied the miners with the tools of the trade. From picks and shovels to tents […]

In 1849, the discovery of gold in California ignited a frenzy, drawing prospectors from around the world in pursuit of quick fortune. While few struck it rich digging and sifting dirt, a different class of entrepreneurs quietly prospered: those who supplied the miners with the tools of the trade. From picks and shovels to tents and provisions, these providers became indispensable to the gold rush, profiting handsomely regardless of who found gold.

Today, a new gold rush is underway, in pursuit of artificial intelligence. And just like the days of yore, the real fortunes may lie not in the gold itself, but in the infrastructure and equipment that enable its extraction. This is where neocloud players and chipmakers are positioned, representing themselves as the fundamental enablers of the AI revolution.

Neoclouds: The Essential Tools and Implements of AI Innovation

The AI boom has sparked a frenzy of innovation, investment, and competition. From generative AI applications like ChatGPT to autonomous systems and personalized recommendations, AI is rapidly transforming industries. Yet, behind every groundbreaking AI model lies an unsung hero: the infrastructure powering it. Enter neocloud providers—the specialized cloud platforms delivering the GPU horsepower that fuels AI’s meteoric rise. Let’s examine how neoclouds represent the “picks and shovels” of the AI gold rush, used for extracting the essential backbone of AI innovation.

Neoclouds are emerging as indispensable players in the AI ecosystem, offering tailored solutions for compute-intensive workloads such as training large language models (LLMs) and performing high-speed inference. Unlike traditional hyperscalers (e.g., AWS, Azure, Google Cloud), which cater to a broad range of use cases, neoclouds focus exclusively on optimizing infrastructure for AI and machine learning applications. This specialization allows them to deliver superior performance at a lower cost, making them the go-to choice for startups, enterprises, and research institutions alike.

The analogy to historical gold rushes is striking: just as miners relied on suppliers of picks and shovels to extract value from the earth, today’s AI pioneers depend on neocloud providers and chipmakers to access the computational resources needed to unlock insights and drive innovation. Neoclouds don’t compete in creating AI applications themselves; instead, they profit by enabling others to do so. This business model positions them as foundational players in the AI economy—profiting regardless of which companies or applications ultimately dominate the market.

Why Neoclouds Are Surging: Four Key Advantages

As AI development accelerates, the demand for compute is outpacing even the hyperscale cloud’s capacity to deliver. Training a foundation model like GPT-4 can require tens of thousands of GPUs running continuously for weeks—something traditional providers weren’t architected to support at scale. Enter the neoclouds: leaner, purpose-built platforms designed to meet the needs of modern AI workloads with greater precision and speed.

Here’s how they’re closing the gap:

1. Specialized Hardware

Neocloud providers are laser-focused on providing access to the newest and most powerful GPUs—often before hyperscalers can make them widely available. NVIDIA’s H100 and A100 accelerators, crucial for training and inference, are the cornerstone of these platforms. Many neoclouds go a step further, offering liquid-cooled racks, ultra-low-latency interconnects, and AI-specific storage tiers designed to keep pace with multi-petabyte datasets. For cutting-edge AI labs and fast-moving startups, this means the difference between weeks and months in development timelines.

2. Bare-Metal Performance

By eliminating the virtualization layers common in general-purpose clouds, neoclouds give users direct access to raw compute power. This bare-metal approach reduces latency and avoids the “noisy neighbor” problem, enabling highly deterministic performance—crucial when fine-tuning large language models or orchestrating tightly coupled GPU workloads. For teams pushing the edge of performance, every clock cycle matters, and neoclouds are delivering those cycles unfiltered.

3. Scalability on Demand

AI R&D is rarely linear. One month you’re iterating on small models, and the next you’re scaling to train a 70-billion-parameter transformer. Neocloud infrastructure is designed to expand and contract with those demands—supporting everything from a few nodes to full-scale superclusters. Unlike traditional clouds, which often impose capacity planning constraints or quotas, neoclouds thrive on elasticity, provisioning capacity dynamically and often within hours rather than weeks.

4. Cost Efficiency with Purpose-Built Pricing

Where hyperscalers often price GPU instances at a premium—factoring in legacy overhead and multi-tenant complexity—neoclouds keep things lean. Many operate with thinner margins and lower operational complexity, translating to significantly lower costs per training hour. Providers like Lambda, Run:ai, and Voltage Park offer transparent, workload-specific pricing that aligns with actual usage. High utilization rates and tailored provisioning models keep costs in check, making neoclouds especially appealing for startups and research groups running on grant cycles or VC runway.

These advantages make neoclouds invaluable not only for startups with limited budgets but also for established enterprises seeking to accelerate their AI initiatives.

Resilience at Scale: Why Neoclouds May Outlast the AI Hype Cycle

Investing in neoclouds offers a unique opportunity to participate in the AI boom without betting on specific applications or platforms. The rapid pace of innovation means that today’s leading AI models could be eclipsed by new breakthroughs tomorrow. However, regardless of which technologies prevail, the need for robust infrastructure will remain constant.

This dynamic mirrors historical gold rushes, where equipment suppliers thrived even as individual miners faced uncertainty. By providing essential tools for AI development, neocloud providers are positioned to benefit from sustained demand across diverse industries—from healthcare and finance to entertainment and logistics.

As the AI gold rush continues, neoclouds are poised to play an increasingly central role in shaping its trajectory. Their ability to deliver cost-effective, high-performance infrastructure makes them critical enablers of innovation. At the same time, their business model—focused on empowering others rather than competing directly—ensures they remain indispensable partners in the AI ecosystem.

Looking forward, neoclouds face challenges such as supply chain constraints for GPUs and competition from hyperscalers attempting to close the performance gap. However, their agility and specialization give them a distinct edge in navigating these hurdles. In many ways, they represent the future of cloud computing: leaner, faster, and more focused on solving specific problems.

As investors and enterprises seek ways to capitalize on AI’s transformative potential, neoclouds offer a compelling proposition—one that promises steady growth amid the chaos of rapid technological change.

Economic Disruption: How Neoclouds Are Redefining Cost and Performance

The emergence of neoclouds is causing a significant economic disruption in the AI infrastructure landscape. Unlike traditional hyperscalers that offer a broad range of services, neocloud providers concentrate on delivering optimized price-performance specifically for AI workloads. This specialization translates into several key advantages: higher GPU utilization rates, bare-metal access, and the application of deep, specialist expertise. These elements combine to create a compelling economic proposition for AI developers and enterprises.

The numbers speak for themselves. Neoclouds are achieving significant cost reductions, with reports from Uptime Institute indicating as much as 66% savings on GPU instances when compared to major hyperscalers. This substantial difference stems from the ability to maximize the use of expensive GPU resources and minimize overhead. For organizations running large-scale AI training or inference tasks, this can lead to considerable savings in operational expenses.

The efficiencies introduced by neoclouds are reshaping the overall economics of AI development. As models grow in complexity and require more compute power, the cost of training and deploying them has become a major barrier. By lowering these costs, neoclouds make it feasible for a wider array of organizations – from startups to established corporations – to engage in AI initiatives. This democratization of AI resources has the potential to accelerate innovation across diverse sectors, allowing more companies to harness the power of AI without breaking the bank.

Advancing Coopetition Between Neocloud Providers and Hyperscalers

The relationship between neocloud providers and traditional hyperscalers is increasingly defined by a complex blend of competition and collaboration. As the demand for AI infrastructure surges, both camps are vying for dominance in the lucrative GPU cloud market, yet their interactions are far from zero-sum, resulting in a competitive, yet symbiotic, market landscape.

Neoclouds have carved out a niche by specializing in GPU-accelerated infrastructure tailored for AI and machine learning workloads. Their agility, focus, and deep understanding of AI developers’ needs allow them to offer cost-effective, high-performance solutions that challenge the broader, premium-priced offerings of hyperscalers. While hyperscalers benefit from vast economies of scale and integrated ecosystems, their diversified business models and high-margin pricing strategies on their cloud businesses often result in higher prices for AI-specific resources.

Despite this competition, the two groups are increasingly intertwined. Neoclouds often position themselves not as direct competitors, but as complementary partners within enterprise multi-cloud strategies. For example, an organization might use a neocloud to train a large language model, then deploy it on a hyperscaler’s platform for inference and integration with other services. This approach allows enterprises to optimize for both performance and cost, leveraging the strengths of each provider.

Investment, Partnership, and Multi-Cloud Integration

The interplay between neoclouds and hyperscalers is further complicated by growing investment and partnership activity. Hyperscalers are not only competing with neoclouds but also investing in them and, in some cases, becoming their customers. A prominent example is Microsoft’s $10 billion commitment to CoreWeave to secure access to specialized GPU infrastructure through 2029. Such deals highlight the recognition by hyperscalers of the unique value neoclouds bring to the AI infrastructure ecosystem.

For enterprises, this dynamic is accelerating the adoption of multi-cloud strategies. By integrating neoclouds into their cloud portfolios, organizations can avoid vendor lock-in, optimize for specific workloads, and ensure access to scarce GPU resources. However, this also introduces new complexities, as enterprises must now manage interoperability and data movement across increasingly fragmented cloud environments.

Looking forward, the evolving relationship between neoclouds and hyperscalers increases the prospects (and pricing) of further M&A activity across key players. As neoclouds grow in scale and strategic importance, it is likely that some will be acquired by hyperscalers seeking to bolster their AI infrastructure capabilities and maintain competitive advantage. Such consolidation will reshape the market’s structure, potentially accelerating innovation through deeper integration, but also raising questions about pricing power and the pace of future disruption.

High Stakes Require High Investment

The capital intensity of neoclouds is staggering. Building and maintaining these specialized infrastructure platforms requires massive investments in GPUs, networking equipment, and data center facilities. To finance this expansion, neoclouds have tapped into both equity and debt markets, with some pioneering the use of GPU assets as collateral for loans. This approach allows them to leverage their hardware investments more efficiently, but also exposes them to the risk of depreciating GPU values and limited liquidity.

The market is also seeing large supply chain vulnerabilities, and rapid equipment obsolescence, increasing the risk of relying too heavily on any one provider. The rapid pace of innovation in chip design presents both opportunities and challenges. While new generations of GPUs promise increased performance and efficiency, they also render older hardware obsolete on extremely short timelines.

This creates a constant pressure to upgrade infrastructure, potentially straining finances and exacerbating supply chain vulnerabilities. Geopolitical factors and manufacturing bottlenecks can further disrupt the supply of GPUs, as we are seeing with recent tariffs affecting NVIDIA, thus impacting neocloud providers’ abilities to meet growing demand.

The Chipmaker Landscape: Navigating the Silicon Battleground of the Neocloud Era

The neocloud boom is redefining the data center ecosystem, driven by billions in venture capital and a red-hot GPU-backed debt market that’s rewriting the rules of infrastructure financing. But while the capital is flowing fast, the underlying hardware story is anything but straightforward. Supply chain constraints, hardware churn, and the specter of over-leverage hang over the sector as neocloud builders sprint to scale. At the heart of this high-stakes race sit the chipmakers — the true power brokers of the AI infrastructure gold rush.

NVIDIA holds the pole position, having effectively set the standard for AI compute with its high-performance GPUs and the proprietary CUDA software stack that developers now treat as foundational. The company’s dominance isn’t just about raw silicon; it’s about controlling the ecosystem. But as demand for AI infrastructure skyrockets, so do concerns about supply bottlenecks, pricing leverage, and the systemic risks of depending too heavily on a single vendor.

Enter AMD and Intel, both aggressively positioning themselves as viable alternatives. AMD’s Instinct accelerators have made meaningful headway, especially among hyperscalers and research labs looking for more open, programmable environments. Its embrace of open-source software and its tight integration across CPU and GPU workloads gives it an edge in environments where flexibility and long-term value matter.

Intel, meanwhile, is betting on a diversified portfolio and a vertically integrated approach. With CPUs, discrete GPUs, and dedicated AI accelerators (via Habana Labs), Intel is aiming to meet the market wherever the workload lands — from training massive models to powering real-time inference at the edge. Its growing software investments signal a deeper push to win developer mindshare, not just silicon sockets.

What’s increasingly clear is that chipmakers now play a strategic role that transcends component supply. Their influence touches everything from AI model optimization to deployment timelines and total cost of ownership. In a capital-intensive market where seconds of inference time and watts per rack can make or break a business model, silicon choices aren’t just technical — they’re existential.

For neocloud operators, the path forward demands architectural agility. Betting solely on NVIDIA may win short-term performance, but long-term resilience will require multi-vendor strategies that hedge against market shocks, broaden workload compatibility, and enhance buyer leverage. The most successful neocloud platforms will be those that understand chipmakers not just as suppliers, but as strategic partners — and occasionally, as competitive threats.

As the AI era matures, the chipmaker battleground will increasingly shape who wins the neocloud race — and who gets left behind in the silicon stampede.

Which Provides the Winning Competitive Edge: Specialization or Scale?

The rise of neoclouds has been driven by their ability to deliver specialized, high-performance infrastructure tailored for AI workloads. However, the question remains: are these advantages sustainable in the long term, or do they merely represent a transitional phase before hyperscalers catch up? For neoclouds to maintain a competitive edge, there are many potential scenarios to consider.

The Bull Case: Persistent Barriers to Entry

The argument for neoclouds’ long-term viability rests on several key factors:

Technical Complexity: Building and managing AI-optimized infrastructure requires deep expertise in GPU architecture, networking, and software. Neoclouds have cultivated this expertise over time, creating a barrier to entry that is difficult for hyperscalers to replicate quickly.

Specialization: Neoclouds focus solely on AI workloads, allowing them to optimize their infrastructure and services for the specific needs of AI developers. This specialization translates into superior performance and cost-efficiency compared to the more generalized offerings of hyperscalers.

Agility: Neoclouds tend to be smaller and more agile than hyperscalers, enabling them to adapt quickly to changing market conditions and emerging technologies. This agility is particularly valuable in the rapidly evolving field of AI.

The Bear Case: Margin Compression and Scale Advantages

Despite many advantages, neoclouds face significant challenges from hyperscalers:

Margin Compression: As hyperscalers invest more heavily in AI infrastructure and refine their offerings, they may be able to erode the price advantage currently enjoyed by neoclouds. Hyperscalers’ scale economies and ability to cross-subsidize AI services with other cloud offerings could put significant pressure on neocloud margins.

Scale Advantages: Hyperscalers possess massive economies of scale, allowing them to procure hardware at lower prices and invest more heavily in R&D. This scale advantage could enable them to leapfrog neoclouds in terms of performance and innovation.

Ecosystem Integration: Hyperscalers offer tightly integrated ecosystems of cloud services, making it easier for customers to build and deploy AI applications. Neoclouds may struggle to match this level of integration, particularly for enterprises that rely on a wide range of cloud services.

Scenarios for the Future

The future of the neocloud market is uncertain, but several scenarios are possible:

Coexistence: Neoclouds and hyperscalers coexist, with each catering to different segments of the market. Neoclouds focus on specialized AI workloads and customers who prioritize performance and cost-efficiency, while hyperscalers cater to enterprises seeking a broader range of cloud services and ecosystem integration.

Market Consolidation: Hyperscalers acquire leading neoclouds to bolster their AI infrastructure capabilities and gain access to specialized expertise. This scenario could lead to greater integration and innovation, but also raise concerns about pricing power and market competition.

Disruption: Neoclouds continue to innovate and disrupt the market, challenging the dominance of hyperscalers and attracting a growing share of AI workloads. This scenario would require neoclouds to overcome challenges related to scale, ecosystem integration, and capital access.

Ultimately, the long-term success of neoclouds will depend on their ability to differentiate themselves from hyperscalers, innovate continuously, and adapt to the rapidly evolving needs of the AI community.

Making AI More Accessible: Who’s Using Neoclouds—and Why It Matters

The neocloud market represents a dynamic and essential force driving innovation in the age of AI. By specializing in high-performance, cost-effective infrastructure, neocloud providers are not only enabling the AI revolution, but also reshaping the economics of cloud computing.

Whether as “picks and shovels” suppliers to the AI gold rush, or as competitive partners alongside hyperscalers, neoclouds are proving their enduring value in a rapidly evolving technological landscape. And as they navigate the challenges of scale, financing, and chipmaker dependencies, the neoclouds are poised to continue pushing the boundaries of what’s possible in AI, fostering a more accessible and innovative future for all.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

IBM proposes unified architecture for hybrid quantum-classical computing

Quantum computers and classical HPC are traditionally “disparate systems [that] operate in isolation,” IBM researchers explain in a new paper. This can be “cumbersome,” because users have to manually orchestrate workflows, coordinate scheduling, and transfer data between systems, thus hindering productivity and “severely” limiting algorithmic exploration. But a hybrid approach

Read More »

F5 brings new visibility and AI controls to Big-IP, NGINX

The demand came from a gap that general-purpose observability tools were not filling. Customers running tools like Datadog and New Relic told F5 they needed something different.  F5 Insight pulls from technology acquired through the Threat Stack and Fletch acquisitions and runs on F5’s AI data fabric. It includes an

Read More »

Tech layoffs surpass 45,000 in early 2026

Layoffs spread across tech sectors Beyond Amazon, Meta, and Block, several technology vendors and platform companies have also announced sizable layoffs this year. According to the RationalFX report: Semiconductor and electronics company ams OSRAM has announced 2,000 layoffs. Telecommunications vendor Ericsson has announced 1,900 job cuts. Semiconductor equipment manufacturer ASML

Read More »

Energy Department Announces $1.9B Investment in Critical Grid Infrastructure to Reduce Electricity Costs

WASHINGTON—The U.S. Department of Energy’s Office of Electricity (OE) today announced an approximately $1.9 billion funding opportunity to accelerate urgently needed upgrades to the nation’s power grid. These investments will meet rising electricity demand and resource adequacy needs, while lowering electricity costs for American households and businesses. Projects selected through the Speed to Power through Accelerated Reconductoring and other Key Advanced Transmission Technology Upgrades (SPARK) funding opportunity will deliver fast and durable upgrades to the grid with real results. In line with President Trump’s Executive Order, Unleashing American Energy, selected projects will demonstrate how reconductoring—replacing existing power lines with higher‑capacity conductors—paired with other Advanced Transmission Technologies (ATTs) can expand grid capacity, increase operational efficiency, lower prices for consumers, and improve overall system reliability and security of the nation’s electric grid. “For too long, important grid modernization and energy addition efforts were not prioritized by past leaders,” said U.S. Secretary of Energy Chris Wright. “Thanks to President Trump, we are doing the important work of modernizing our grid so electricity costs will be lowered for American families and businesses.” “The United States must increase grid capacity to meet demand, and ensure the grid provides reliable power—day-in and day-out,” said OE Assistant Secretary Katie Jereza. “Through this SPARK funding opportunity, we will stabilize and optimize grid operations to strengthen it for rapid growth.” The SPARK opportunity builds on the Grid Resilience and Innovation Partnerships (GRIP) Program, which provided up to $10.5 billion in competitive funding over five years to states, tribes, electric utilities, and other eligible recipients to strengthen grid resilience and innovation. The previous two GRIP funding rounds covered FY 2022-2023 and FY 2023-2024 funding. Today’s announcement continues the mission of the GRIP Program under the SPARK funding opportunity, focusing on the rapid deployment of reconductoring and other ATTs that expand transfer capability, strengthen reliability

Read More »

United States to Release 172 Million Barrels of Oil From the Strategic Petroleum Reserve

WASHINGTON—U.S. Secretary of Energy Chris Wright released the following statement regarding the International Energy Agency (IEA) and the U.S. Strategic Petroleum Reserve (SPR): “Earlier today, 32 member nations of the International Energy Agency unanimously agreed to President Trump’s request to lower energy prices with a coordinated release of 400 million barrels of oil and refined products from their respective reserves.  “As part of this effort, President Trump authorized the Department of Energy to release 172 million barrels from the Strategic Petroleum Reserve, beginning next week. This will take approximately 120 days to deliver based on planned discharge rates.  “President Trump promised to protect America’s energy security by managing the Strategic Petroleum Reserve responsibly and this action demonstrates his commitment to that promise. Unlike the previous administration, which left America’s oil reserves drained and damaged, the United States has arranged to more than replace these strategic reserves with approximately 200 million barrels within the next year—20% more barrels than will be drawn down—and at no cost to the taxpayer.  “For 47 years, Iran and its terrorist proxies have been intent on killing Americans. They have manipulated and threatened the energy security of America and its allies. Under President Trump, those days are coming to an end.  “Rest assured, America’s energy security is as strong as ever.”                                                                                         ###

Read More »

Occidental Petroleum, 1PointFive STRATOS DAC plant nears startup in Texas Permian basin

Occidental Petroleum Corp. and its subsidiary 1PointFive expect Phase 1 of the STRATOS direct air capture (DAC) plant in Texas’ Permian basin to come online in this year’s second quarter. In a post to LinkedIn, 1PointFive said Phase 1 “is in the final stage of startup” and that Phase 2, which incorporates learnings from research and development and Phase 1 construction activities, “will also begin commissioning in Q2, with operational ramp-up continuing through the rest of the year.” Once fully operational, STRATOS is designed to capture up to 500,000 tonnes/year (tpy) of CO2. As part of the US Environmental Protection Agency (EPA) Class VI permitting process and approval, it was reported that STRATOS is expected to include three wells to store about 722,000 tpy of CO2 in saline formations at a depth of about 4,400 ft. The company said a few activities before start-up remain, including ramping up remaining pellet reactors, completing calciner final commissioning in parallel, and beginning CO2 injection. Start-up milestones achieved include: Completed wet commissioning with water circulation. Received Class VI permits to sequester CO2. Ran CO2 compression system at design pressure. Added potassium hydroxide (KOH) to capture CO2 from the atmosphere. Building pellet inventory. Burners tested on calciner.  

Read More »

Brava Energia weighs Phase 3 at Atlanta to extend production plateau

Just 2 months after bringing its flagship Atlanta field onstream with the new FPSO Atlanta, Brazil’s independent operator Brava Energia SA is evaluating a potential third development phase that could add roughly 25 million bbl of reserves and help sustain peak production longer than originally planned. The Phase 3 project, still at an early technical and economic evaluation stage, focuses on the Atlanta Nordeste area; a separate, shallower reservoir discovered in 2006 by Shell’s 9-SHEL-19D-RJS well. According to André Fagundes, vice-president of research (Brazil) at Welligence Energy Analytics, Phase 2 has four wells still to be developed: two expected in 2027 and two in 2029. Phase 3 would involve drilling two additional wells in 2031, bringing total development to 12 producing wells. Until recently, full-field development was understood to comprise 10 wells, but Brava has since updated guidance to reflect a 12-well development concept. Atlanta field upside The primary objective is clear. “We believe its main objective is to extend the production plateau,” Fagundes said. Welligence estimates incremental recovery could reach 25 MMbbl, increasing the field’s overall recovery factor by roughly 1.5%. Lying outside Atlanta’s main Cretaceous reservoir, Atlanta Nordeste represents a genuine upside opportunity, Fagundes explained. The field benefits from strong natural aquifer support, and no water or gas injection is anticipated. Water-handling constraints that affected early production using the Petrojarl I—limited to 11,500 b/d of water treatment—are no longer a bottleneck. FPSO Atlanta can process up to 140,000 b/d of water. Reservoir performance to date has been solid, albeit with difficulties. Recurrent electric submersible pump (ESP) failures and processing limits on the previous FPSO complicated full validation of original reservoir models. With the new 50,000-b/d FPSO in operation since late 2024, reservoir deliverability has become the main constraint. Phase 3 wells would also use ESPs and require additional subsea

Read More »

California Resources eyes ‘measured’ capex ramp on way to 12% production growth thanks to Berry buy

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: var(–color-primary-main); } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style { font-size: 0.75rem; opacity: .6; } #onetrust-pc-sdk [id*=btn-handler], #onetrust-pc-sdk [class*=btn-handler] { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-policy a, #onetrust-pc-sdk a, #ot-pc-content a { color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-pc-sdk .ot-active-menu { border-color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-accept-btn-handler, #onetrust-banner-sdk #onetrust-reject-all-handler, #onetrust-consent-sdk #onetrust-pc-btn-handler.cookie-setting-link { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-consent-sdk .onetrust-pc-btn-handler { color: #c19a06 !important; border-color: #c19a06 !important; } The leaders of California Resources Corp., Long Beach, plan to have the company’s total production average 152,000-157,000 boe/d in 2026, with each quarter expected to be in that range. That output would equate to an increase of more than 12% from the operator’s 137,000 boe/d during fourth-quarter 2025, due mostly to the mid-December acquisition of Berry Corp. Fourth-quarter results folded in 14 days of Berry production and included 109,000 b/d of oil, with the company’s assets in the San Joaquin and Los Angeles basins accounting for 99,000 b/d of that total. The company dilled 31 new wells during the quarter and 76 in all of 2025—all in the San Joaquin—but that number will grow significantly to about 260 this year as state officials have resumed issuing permits following the passage last fall of a bill focused on Kern County production. Speaking to analysts after CRC reported fourth-quarter net income of $12 million on $924 million in revenues, president and chief executive officer Francisco Leon and chief financial officer Clio Crespy said the goal is to manage 2026 output decline to roughly 0.5% per quarter while operating four rigs and

Read More »

Petro-Victory Energy spuds São João well in Brazil

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: var(–color-primary-main); } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style { font-size: 0.75rem; opacity: .6; } #onetrust-pc-sdk [id*=btn-handler], #onetrust-pc-sdk [class*=btn-handler] { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-policy a, #onetrust-pc-sdk a, #ot-pc-content a { color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-pc-sdk .ot-active-menu { border-color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-accept-btn-handler, #onetrust-banner-sdk #onetrust-reject-all-handler, #onetrust-consent-sdk #onetrust-pc-btn-handler.cookie-setting-link { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-consent-sdk .onetrust-pc-btn-handler { color: #c19a06 !important; border-color: #c19a06 !important; } Petro-Victory Energy Corp. has spudded the SJ‑12 well at São João field in Barreirinhas basin, on the Brazilian equatorial margin, Maranhão.  Drilling and testing SJ‑12 is aimed at proving enough gas can be produced to sell locally. The well forms part of the single non‑associated gas well commitment under a memorandum of understanding signed in 2024 with Enava. São João contains 50.1 bcf (1.4 billion cu m) non‑associated gas resources. Petro‑Victory 100% owns and operates São João field.

Read More »

Cisco grows high-end optical support for AI clusters

Cisco has also upgraded its Network Conversion System (NCS) with a 1RU, 800GE line card offering 12.8T capacity, with 32 OSFP-based ports for 100GE, 400GE, and 800GE clients and 800ZR/ZR+ WDM trunks. The NCS 1014  doubles the density of previous-generation NCS versions and now includes MACsec encryption (IEEE 802.1AE) to secure point-to-point links with hardware-based encryption, data integrity, and authentication for Ethernet traffic, Ghioni stated. It supports enhanced capacity and performance with C&L-band support and NCS 1014 systems with the 2.4T WDM line card based on the Coherent Interconnect Module 8 and now supports 800 GE clients, which can be mapped directly to a wavelength or inverse multiplexed across two wavelengths to maximize reach, Ghioni wrote.  In the pluggable optic arena, Cisco is now offering a Quad Small Form Factor Pluggable Double Density (QSFP-DD) Pluggable Protection Switch Module that can monitor the optical link and switch traffic if it detects a fault in less than 50 milliseconds. The module occupies a quarter of the rack space compared to traditional protection devices—offering 90% rack space saving over available options, Ghioni wrote.  It is aimed at Metro and DCI network customers where sub-50 ms failure recovery is essential and data centers needing fiber protection without bulky hardware, Ghioni stated.  Cisco also added its Acacia developed Bright QSFP28 100ZR 0 dBm coherent optical pluggable in a standard QSFP28 form factor.  It is aimed at edge, access, enterprise, and campus network deployment. Cisco has been actively growing its optical portfolio  recently adding the Cisco Silicon One G300, which powers 102.4T N9000 and Cisco 8000 systems, as well as advanced 1.6T OSFP optics and 800G Linear Pluggable Optics. 

Read More »

Datalec targets rapid infrastructure deployment with new modular data centers

“We are engineering the data center with a new lens bringing pre-engineered system designs that are flexible and adaptable that enables a tailored solution for clients,” said John Lever, director of modular solutions at Datalec. The systems are flexible enough that these solutions cater for all types of data center, from standard server technology to AI and high-density compute. Datalec also provides “bolt-on” solutions, including a ‘digital wrapper’ including digital twinning and lifecycle and global support, Lever says. Another way Datalec says it differentiates from competing modular designs is a larger share of work is done offsite in a controlled manufacturing environment, which cuts onsite construction time, improves safety and limits disruption to live facilities, Lever says. The company competes with other modular data center vendors including Schneider Electric, Vertiv, Flex many others. DPI’s says its services are aimed at colocation providers, hyperscale and AI infrastructure teams, and large enterprises that need to add capacity quickly, safely and cost effectively across multiple regions.

Read More »

Study finds significant savings from direct current power for AI workloads

The result is a 50% to 80% reduction in copper usage, due to fewer conductors and less parallel cabling, and an 8% to 12% reduction in annual energy-related OpEx through lower conversion and distribution losses. By reducing conductor count, cabling, and redundant power components, 800VDC enables meaningful savings at both build-out and operational stages. AI-first facilities can see a $4 million to $8 million in CapEx savings per 10 MW build by reducing upstream AC. For a one-gigawatt data center, you’re saving a couple million pounds of copper wire, he said. Burke says an all-DC data center is best done with a whole new facility rather than retrofitting old facilities. “[DC] is going to be in a lot of greenfield data centers that are going to be built, and data centers that are going to go to higher compute power are also going to DC,” he said. He did recommend all-DC retrofits for existing data centers that are going to employ high power computing with GPUs. Enteligent’s unnamed and as yet unreleased product is a converter that takes 800 volts and partitions it to 50 volts for the computing servers. The company will provide a new power supply, power shelf that converts 800 volts DC to 50 volts DC much more efficiently than any current power supplies. Burke said the company is doing NDA level testing and pilot programs now with its product, but it will be making a formal announcement within the next few weeks. There are a number of players in the DC arena focusing on different parts of the power supply market including Vertiv, Rutherford, Siemens, Eaton and many more.

Read More »

Cisco blends Splunk analytics, security with core data center management

With the integration, data center teams can gather and act on events, alarms, health scores, and inventory through open APIs, Cisco stated. It also offers pre-built and customizable dashboards for inventory, health, fabric state, anomalies, and advisories as well as correlates telemetry across fabrics and technology tiers for actionable insights, according to Cisco. “This isn’t just another connector or API call. This is an embedded, architectural integration designed to transform how you monitor, troubleshoot, and secure your data center fabric. By bringing the power of Splunk directly into the Data Center Networking environment, we are enabling teams to solve complex problems faster, maintain strict data sovereignty, and dramatically reduce operational costs,” wrote Usha Andra is a senior product marketing leader and Anant Shah, senior product manager, both with Cisco Data Center Networking in a blog about the integration.  “Traditionally, network monitoring involves a trade-off. You either send massive amounts of raw logs to a centralized data lake, incurring high ingress and storage costs. Or you rely on sampled data that misses critical microbursts and anomalies,” Andra and Shah wrote.  “Native Splunk integration changes the paradigm by running Splunk capabilities directly within the Cisco Nexus Dashboard. This allows for the streaming of high-fidelity telemetry, including anomalies, advisories, and audit logs, directly to Splunk analytics.”

Read More »

Execution, Power, and Public Trust: Rich Miller on 2026’s Data Center Reality and Why He Built Data Center Richness

DCF founder Rich Miller has spent much of his career explaining how the data center industry works. Now, with his latest venture, Data Center Richness, he’s also examining how the industry learns. That thread provided the opening for the latest episode of The DCF Show Podcast, where Miller joined present Data Center Frontier Editor in Chief Matt Vincent and Senior Editor David Chernicoff for a wide-ranging discussion that ultimately landed on a simple conclusion: after two years of unprecedented AI-driven announcements, 2026 will be the year reality asserts itself. Projects will either get built, or they won’t. Power will either materialize, or it won’t. Communities will either accept data center expansion – or they’ll stop it. In other words, the industry is entering its execution phase. Why Data Center Richness Matters Now Miller launched Data Center Richness as both a podcast and a Substack publication, an effort to experiment with formats and better understand how professionals now consume industry information. Podcasts have become a primary way many practitioners follow the business, while YouTube’s discovery advantages increasingly make video versions essential. At the same time, Miller remains committed to written analysis, using Substack as a venue for deeper dives and format experimentation. One example is his weekly newsletter distilling key industry developments into just a handful of essential links rather than overwhelming readers with volume. The approach reflects a broader recognition: the pace of change has accelerated so much that clarity matters more than quantity. The topic of how people learn about data centers isn’t separate from the industry’s trajectory; it’s becoming part of it. Public perception, regulatory scrutiny, and investor expectations are now shaped by how stories are told as much as by how facilities are built. That context sets the stage for the conversation’s core theme. Execution Defines 2026 After

Read More »

Nomads at the Frontier: PTC 2026 Signals the Digital Infrastructure Industry’s Moment of Execution

Each January, the Pacific Telecommunications Council conference serves as a barometer for where digital infrastructure is headed next. And according to Nomad Futurist founders Nabeel Mahmood and Phillip Koblence, the message from PTC 2026 was unmistakable: The industry has moved beyond hype. The hard work has begun. In the latest episode of The DCF Show Podcast, part of our ongoing ‘Nomads at the Frontier’ series, Mahmood and Koblence joined Data Center Frontier to unpack the tone shift emerging across the AI and data center ecosystem. Attendance continues to grow year over year. Conversations remain energetic. But the character of those conversations has changed. As Mahmood put it: “The hype that the market started to see is actually resulting a bit more into actions now, and those conversations are resulting into some good progress.” The difference from prior years? Less speculation. More execution. From Data Center Cowboys to Real Deployments Koblence offered perhaps the sharpest contrast between PTC conversations in 2024 and those in 2026. Two years ago, many projects felt speculative. Today, developers are arriving with secured power, customers, and construction underway. “If 2024’s PTC was data center cowboys — sites that in someone’s mind could be a data center — this year was: show me the money, show me the power, give me accurate timelines.” In other words, the market is no longer rewarding hypothetical capacity. It is demanding delivered capacity. Operators now speak in terms of deployments already underway, not aspirational campuses still waiting on permits and power commitments. And behind nearly every conversation sits the same gating factor. Power. Power Has Become the Industry’s Defining Constraint Whether discussions centered on AI factories, investment capital, or campus expansion, Mahmood and Koblence noted that every conversation eventually returned to energy availability. “All of those questions are power,” Koblence said.

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »