Stay Ahead, Stay ONMINE

Eight Trends That Will Shape the Data Center Industry in 2026

For much of the past decade, the data center industry has been able to speak in broad strokes. Growth was strong. Demand was durable. Power was assumed to arrive eventually. And “the data center” could still be discussed as a single, increasingly important, but largely invisible, piece of digital infrastructure. That era is ending. As […]

For much of the past decade, the data center industry has been able to speak in broad strokes. Growth was strong. Demand was durable. Power was assumed to arrive eventually. And “the data center” could still be discussed as a single, increasingly important, but largely invisible, piece of digital infrastructure.

That era is ending.

As the industry heads into 2026, the dominant forces shaping data center development are no longer additive. They are interlocking and increasingly unforgiving. AI drives density. Density drives cooling. Cooling and density drive power. Power drives site selection, timelines, capital structure, and public response. And once those forces converge, they pull the industry into places it has not always had to operate comfortably: utility planning rooms, regulatory hearings, capital committee debates, and community negotiations.

The throughline of this year’s forecast is clarity:

  • Clarity about workload classes.
  • Clarity about physics.
  • Clarity about risk.
  • And clarity about where the industry’s assumptions may no longer hold.

One of the most important shifts entering 2026 is that it may increasingly no longer be accurate, or useful, to talk about “data centers” as a single category. What public discourse often lumps together now conceals two very different realities: AI factories built around sustained, power-dense GPU utilization, and general-purpose data centers supporting a far more elastic mix of cloud, enterprise, storage, and interconnection workloads. That distinction is no longer academic. It is shaping how projects are financed, how power is delivered, how facilities are cooled, and how communities respond.

It’s also worth qualifying a line we’ve used before, and still stand by in spirit: that every data center is becoming an AI data center.

In 2026, we feel that statement is best understood more as a trajectory, and less a design brief. AI is now embedded across the data center stack: in operations, in customer workloads, in planning assumptions, and in the economics of capacity. But that does not mean every facility behaves the same way, bears the same risks, or demands the same infrastructure response. Some data centers absorb AI as one workload among many. Others are purpose-built around it, with consequences that ripple through power, cooling, capital structure, and public visibility.

The industry’s challenge and opportunity heading into 2026 is learning to operate with both truths at once: AI is everywhere, but AI does not make all data centers alike.

At the same time, the industry is becoming more self-aware. The AI buildout has not exactly slowed, but it has matured. Capital is still available, but it is no longer blind. Utilities are no longer downstream service providers; they are co-architects. Liquid cooling has moved from experiment to industrial system; but only where density demands it. Modularization and standardization are no longer about speed alone; they are about survivability. And for the first time in years, the industry is apparently beginning to plan not just for growth, but for volatility.

That does not signal retreat. It signals experience.

The eight trends that follow reflect an industry still expanding, still innovating, and still under pressure; but increasingly disciplined about where ambition meets reality. They trace how AI reshapes infrastructure without consuming it entirely; how power becomes the central organizing constraint; how execution overtakes novelty as the competitive advantage; and how a sector long accustomed to momentum begins designing for durability.

If there is a single lesson embedded in this year’s forecast, it is this: the frontier has moved inward. The challenges shaping 2026 are less about discovering what’s next than about managing what scale actually demands.

What follows is our view of the eight trends that will define that reckoning.

1. AI Energy Demand Defines a Split Between AI Factories and General-Purpose Data Centers and Pulls the Industry Into the Public Arena

In 2026, it may no longer be accurate or useful to talk about “data centers” as a single category.

What are often grouped together in public debate are, in practice, two very different classes of infrastructure:

  • AI factories, designed around sustained GPU utilization, extreme rack densities, and tightly coupled power-and-cooling systems.
  • General-purpose data centers, supporting cloud, enterprise, storage, interconnection, and mixed workloads with far more elastic demand profiles.

This distinction matters because the most acute energy stress and, arguably, the bulk of public scrutiny has concentrated around AI factories, not the broader data center ecosystem.

AI factories behave differently. Their loads are larger, less interruptible, and more visible. Campuses routinely plan for hundreds of megawatts of firm power, with timelines that compress utility planning cycles and push infrastructure decisions upstream. Their tolerance for curtailment is low. Their demand profiles look less like traditional commercial load and more like industrial infrastructure.

General-purpose data centers remain constrained, but more adaptable. They can phase capacity, diversify workloads, and absorb efficiency gains more gradually. In many cases, they are now being pulled into debates sparked by AI factories that operate at an entirely different scale.

This divergence has a direct social consequence.

As AI factories grow larger and more concentrated, they increasingly lose the industry’s long-standing advantage of invisibility. Utilities, regulators, local governments, and communities are no longer reacting to “data centers” in the abstract; they are responding to very specific projects with very specific impacts on grid capacity, land use, water, and long-term energy planning.

By 2026, data centers, particularly AI factories, are being treated less as optional economic development projects and more as critical infrastructure, alongside transportation, energy, and water systems. That reframing brings stature, but also scrutiny in the form of:

  • More formal load-prioritization debates.
  • Greater regulatory visibility at the state and regional level.
  • Expectations around resilience, transparency, and continuity planning.
  • Public questions about who benefits, who pays, and who bears risk.

Importantly, this scrutiny is not evenly distributed. It follows density and power. The sharper the load profile, the brighter the spotlight.

The result is a more precise, but more demanding, conversation. In 2026, the question is no longer just whether “data centers” are straining the grid. It is which class of data center, under what assumptions, and at what scale.

Under this framing, the industry gains clarity, but loses the comfort of being misunderstood.

Onsite Power Moves From Contingency to Architecture

One of the clearest signals embedded in the AI factory versus general-purpose data center split is how power is being sourced.

For decades, behind-the-meter generation was treated more as a contingency: backup power, resilience insurance, and/or a niche solution for remote sites. By 2026, that framing no longer holds for AI factory campuses. In these environments, onsite power is increasingly part of the primary architecture.

The reason is straightforward. AI factories combine three characteristics that strain traditional grid-only models:

  • Scale – Single-tenant or tightly clustered loads routinely planning for hundreds of megawatts.
  • Continuity – Sustained, non-interruptible utilization profiles with low tolerance for curtailment.
  • Speed – Development timelines that move faster than transmission upgrades and interconnection queues.

Utilities are not failing. They are simply being asked to do something grids were not designed to do quickly: absorb massive, fast-ramping industrial-scale loads on compressed timelines.

The industry’s response has been pragmatic rather than ideological. Natural gas has emerged as the near-term bridge not because it is fashionable, but because it is dispatchable, scalable, and deployable within realistic time horizons. For many AI factory projects, gas-fired generation, sometimes paired with carbon capture planning or hydrogen-blending roadmaps, is the only way to reconcile density with delivery schedules.

At the same time, long-cycle power strategies are being layered in. Nuclear, particularly small modular reactors (SMR) and microreactors, has now obviously moved from theoretical alignment to strategic positioning. While few operators expect meaningful nuclear capacity to materialize this decade, the planning assumptions are already influencing site selection, land control, and partnership structures.

What is striking is how asymmetric this shift remains.

General-purpose data centers continue to rely primarily on the grid, augmented by efficiency gains, phased delivery, and demand flexibility. AI factories, by contrast, are forcing the issue. Their power requirements are so concentrated, and frankly so visible, that they are accelerating new models of generation ownership, co-investment, and hybrid supply.

The result is not grid abandonment by any means, but grid re-negotiation. Onsite power does not replace utilities; it reshapes the relationship. Developers are increasingly by necessity co-planning load, generation, and phasing from day one, rather than treating power as a downstream procurement exercise.

In 2026, behind-the-meter power is no longer a signal of exceptionalism. It’s a marker of workload class. Where density demands it, onsite generation moves from the margins to the blueprint.

2. The AI Infrastructure “Bubble” Debate Moves Inside the Industry as Power Becomes a Negotiated Relationship

By 2026, concerns about whether parts of the AI infrastructure buildout are running ahead of sustainable demand are no longer whispered. They are increasingly debated in plain view inside boardrooms, earnings calls, utility planning sessions, and capital committees.

This is not a blanket bubble call. Demand for AI compute remains real and growing. But the shape of the buildout, particularly GPU-dense capacity optimized for large-scale training workloads, has introduced new questions about utilization durability, asset lifecycles, and financial exposure. Rapid hardware iteration, short depreciation curves, and the narrowing reuse profile of AI-factory infrastructure have made “build it and they will come” a harder assumption to defend.

What has sharpened that debate is power.

In earlier eras, power was something data center developers procured after selecting a site. In 2026, power is increasingly something that must be co-designed from the outset, and utilities are asserting themselves accordingly. Large AI-oriented projects are being asked to phase load, accept curtailment provisions, co-invest in generation or transmission upgrades, and, in some cases, relocate altogether to align with grid realities.

That shift has important implications for the bubble conversation. Projects that once penciled out on paper can stall when power delivery timelines stretch, interconnection conditions tighten, or utilities impose behavioral constraints on load. In that environment, speculative capacity becomes riskier not because AI demand disappears, but because power delivery, not capital availability, becomes the gating variable.

Developers and investors are responding with greater discipline. Phased campuses, modular delivery, optional expansion rights, and diversified workload strategies are increasingly favored over monolithic, all-at-once builds. Capital remains available, but it is more selective: rewarding projects that demonstrate power certainty, flexibility, and credible paths to sustained utilization.

The result in 2026 is not an industry in retreat, but one in recalibration. The AI infrastructure boom continues, but it does so under tighter scrutiny from utilities, regulators, and investors alike. Power is no longer a background assumption. It is an active participant in shaping which projects move forward, at what scale, and on what terms.

In that sense, the bubble debate is less about whether AI demand is real, and more about who bears the risk when infrastructure ambitions collide with grid constraints.

3. AI Integration Deepens Even as GPU-Centric Economics Remain Unsettled and Capital Discipline Sharpens

As the industry heads into 2026, artificial intelligence is no longer an overlay on big-picture data center strategy. It is embedded across the stack; from facility design assumptions and cooling architectures to operational tooling, workload scheduling, and infrastructure planning.

At the same time, the economics of accelerated compute remain unsettled.

The industry spent much of 2024 and 2025 building GPU-centric capacity at unprecedented speed, driven by hyperscale demand, competitive pressure, and the fear of missing the AI moment. What emerged alongside that buildout was a more sober understanding of the tradeoffs involved. Rapid hardware iteration, high capital intensity, and uncertain utilization curves have introduced meaningful balance-sheet and lifecycle risk; particularly for assets optimized narrowly around specific generations of GPUs or training workloads.

As a result, 2026 opens with an industry that is far more fluent in AI infrastructure, but less naïve about its financial implications. This is where capital discipline begins to matter more than capital abundance.

Money remains available for data center and AI infrastructure projects, but it is no longer indiscriminate. Investors, lenders, and partners are increasingly separating AI factories from general-purpose data centers, short-cycle compute risk from long-cycle real estate risk, and speculative capacity from projects anchored by durable contracts or diversified workloads. Coverage from The Wall Street Journal, Financial Times, and Bloomberg over the past year has consistently highlighted this shift, noting greater scrutiny of utilization assumptions, depreciation schedules, and exit flexibility tied to GPU-heavy builds. 

The consequence is a quieter but significant change in who wins deals. Speed alone is no longer the decisive advantage. The developers and operators that attract capital in 2026 are those that can clearly articulate risk: how assets can be reused, how density can be dialed up or down, how power and cooling investments remain valuable across multiple compute generations, and how exposure is phased rather than front-loaded.

AI integration continues to deepen across the industry but it does so within a more disciplined capital framework. Facilities are still being built. AI capacity is still expanding. What changes is the tone of the conversation: from growth-at-any-cost to growth-with-options.

In that sense, the maturation of AI infrastructure in 2026 is not marked by slower adoption, but by better questions being asked earlier; by boards, investors, and operators who now understand that AI fluency does not eliminate risk, it simply makes it easier to see.

4. The MegaCampus Becomes an AI Factory and Utilities Become Co-Architects

In 2026, the megacampus is no longer a generic hyperscale construct. It is increasingly an AI factory campus, and that shift has permanently altered the relationship between data center developers and utilities.

This transition is being driven most forcefully by the major hyperscalers (AWS, Microsoft, Google, Meta, and Oracle) whose AI factory requirements compress timelines and concentrate demand at a scale utilities cannot treat as incremental. These companies are no longer adding load at the margins. They are reshaping regional power planning assumptions.

In earlier eras, utilities were service providers. Power was requested, modeled, queued, and delivered, eventually. That model breaks down when campuses plan for hundreds of megawatts of sustained, non-interruptible load on timelines that outpace transmission upgrades and generation build-outs.

As a result, utilities are no longer downstream participants in AI factory development. They are effectively becoming co-architects, engaged before site plans are finalized and often before land is fully controlled. Power strategy now gates everything that follows: campus layout, phasing, cooling architecture, capital structure, and even customer mix.

This co-architect role shows up in several concrete ways:

  • Phased load agreements that tie capacity delivery to infrastructure milestones.
  • Co-investment models in substations, generation, or transmission.
  • Load shaping and curtailment frameworks negotiated up front, not imposed later.
  • Locational discipline, with utilities increasingly steering projects toward grid-advantaged zones rather than reacting to developer preference.

In this environment, the megacampus era does not simply expand. It specializes.

Mapping the Power Stack: Gas, Nuclear, and Hybrid Models

What has also become clearer by 2026 is how different power sources align to different time horizons, and how utilities and hyperscalers are orchestrating those layers together.

Natural gas occupies the near-term execution layer. It is dispatchable, scalable, and deployable within timelines that match AI factory demand. For hyperscalers seeking speed-to-market, gas-fired generation (whether utility-owned, developer-owned, or structured through long-term supply agreements) has become the default bridge where grid capacity lags load growth. This is not a philosophical choice. It is a scheduling one.

Nuclear, particularly small modular reactors and microreactors, occupies the long-cycle planning layer. Hyperscalers and utilities alike are increasingly aligning land control, permitting strategy, and partnership structures around future nuclear potential, even as most acknowledge that meaningful capacity will arrive later in the decade. Nuclear is shaping where megacampuses are planned, even if it does not yet power them.

Hybrid models now define the middle ground. These combine grid supply, onsite generation, phased delivery, storage, and future-proofing assumptions into a single integrated plan. In practice, this is where utility co-architecture is most visible. Hyperscalers, in particular, are increasingly willing to engage utilities directly on generation strategy; co-planning gas capacity in the near term, reserving nuclear-adjacent land for the long term, and structuring hybrid delivery models that smaller operators cannot replicate.

What emerges is a more explicit division of labor. Utilities retain their central role in grid reliability and long-term planning. Developers translate hyperscale requirements into buildable form. Hyperscalers accept that power can no longer be abstracted away from design.

The megacampus era did not arrive in 2026. It consolidated – and learned to speak fluently with the grid.

5. Pricing Pressure Pushes Demand Outward but Power Remains the Constraint, Forcing a Turn Toward Standardization

By 2026, rising pricing and limited deliverable supply in core data center markets continue to push demand outward. Larger continuous requirements (particularly blocks of 10 megawatts and above) face the sharpest pressure, driven by hyperscale absorption, constrained power availability, and persistently elevated construction costs.

But this was never simply a geography story.

Even as developers and hyperscale partners widen their search radius into secondary and tertiary markets, the same constraint follows them: power. In many emerging regions, land is available, entitlements are workable, and local officials are receptive…but utility headroom remains limited. Interconnection timelines, substation capacity, and grid upgrade requirements may increasingly prove more decisive than zoning, tax incentives, or political goodwill.

Industry analysis from CBRE, JLL, Cushman & Wakefield, and Reuters over the past year has consistently shown that new markets do not eliminate constraints so much as reintroduce them under different utility footprints. The outward push continues, but it is bounded by the same physical realities.

As a result, the industry’s response in 2026 is probably not just geographic expansion, but also operational compression; and that is where standardization reasserts itself.

After years of bespoke design driven by hyperscale customization and early AI experimentation, data center development is swinging back toward standardization: not for elegance or theoretical efficiency, but for survivability. Repeatable power blocks, modular cooling architectures, factory-built subsystems, and standardized electrical rooms increasingly replace one-off designs that slow delivery, complicate commissioning, and collide with utility timelines.

This shift is particularly visible in projects tied to large, continuous loads. Hyperscalers still drive requirements; but even they are showing greater tolerance for standardized delivery models when speed, phasing, and power coordination matter more than architectural novelty. Vendor lock-in, once resisted, is becoming more quietly accepted as the price of execution certainty.

Standardization also serves a financial purpose. When pricing pressure is high and power delivery uncertain, developers need to reduce the variables they can control. Standardized designs shorten timelines, lower construction risk, and make it easier to phase capacity in alignment with utility delivery schedules. They also allow operators to make clearer, more defensible commitments to customers at a time when overpromising has become increasingly risky.

In that sense, standardization becomes the industry’s coping mechanism for constraint. It is how operators keep promises in markets where supply is tight, pricing is unforgiving, and power cannot be taken for granted. The outward push continues in 2026, but it does so with fewer design experiments, tighter playbooks, and a growing recognition that repeatability is now a competitive advantage.

6. Liquid Cooling Will Become Table Stakes, But Mainly Where Density Demands It

By 2026, liquid cooling is obviously now very far from a speculative technology. But neither is it a universal mandate.

What the industry has learned, often through hard experience, is that deploying liquid cooling at scale is less about thermodynamics than execution. The real inflection point is not whether liquid cooling works. It does. The question is where it can be deployed repeatably, operated safely, and maintained without friction as fleets scale from pilots to production.

The Physical Reality: Designing for Scale, Not Demos

On a physical track, direct-to-chip liquid cooling becomes standard in AI factory environments where sustained GPU utilization and extreme rack densities overwhelm the limits of air. These facilities are designed from inception around liquid loops, higher inlet temperatures, and thermal architectures optimized for continuous, high-intensity workloads. Cooling is no longer an accessory system layered onto the building. It is a defining design constraint that shapes floor loading, piping routes, redundancy models, and commissioning timelines.

As AI factories scale, operators increasingly standardize around repeatable cooling blocks rather than bespoke hall-by-hall designs. Manifold layouts, CDU placement, leak-detection systems, and maintenance access are engineered for replication, not experimentation. The priority shifts from achieving maximum theoretical efficiency to ensuring predictable performance across hundreds or thousands of racks.

General-purpose data centers follow a different path. Rather than liquid-first designs, most continue adopting hybrid approaches: rear-door heat exchangers, localized liquid-cooled zones, and selective support for high-density clusters. This allows operators to support AI workloads without committing entire facilities to liquid infrastructure that may limit future reuse. In these environments, liquid cooling is an overlay, not the foundation.

Immersion: Operationally Viable, Strategically Selective

Immersion cooling continues to advance, validate, and professionalize – but selectively. By 2026, it has proven itself operationally viable in specific AI factory use cases where density, space efficiency, or thermal headroom justify the added complexity. However, immersion remains uneven in its operational footprint.

The barriers are not technical so much as logistical and organizational: fluid handling, component compatibility, maintenance workflows, vendor coordination, and regulatory familiarity. Immersion systems demand new service models, retraining of technicians, and tighter integration between IT and facilities teams. For many operators, those tradeoffs remain acceptable only in tightly controlled, purpose-built environments.

As a result, in 2026 immersion probably does not flip into a mainstream default across hyperscale or colocation design. It matures, but remains intentional, not ubiquitous.

The Structural Shift: Cooling Becomes an O&M Discipline

On a structural track, cooling strategy becomes explicitly workload-specific rather than aspirational. The industry moves away from one-size-fits-all narratives toward pragmatic segmentation:

  • AI factories optimize for sustained thermal performance and continuous utilization.
  • General-purpose facilities optimize for adaptability, serviceability, and long-term reuse.
  • Hybrid designs bridge the two where economics and customer mix demand it.

As fleets scale, operations and maintenance move to the center of cooling decisions. Leak management, spare-parts logistics, service intervals, technician training, and failure isolation increasingly outweigh marginal gains in efficiency. Designs that are repeatable, serviceable, and compatible with evolving hardware roadmaps gain favor over more exotic configurations that introduce operational risk.

This operational reality reinforces the turn toward standardization seen elsewhere in the industry. Liquid cooling systems are increasingly specified, installed, and maintained as industrial infrastructure: less bespoke, more modular, and tightly integrated with power and monitoring systems.

The 2026 Reality Check

The net effect in 2026 is clarity. Liquid cooling becomes essential where density demands it but optional where it does not. The industry stops arguing whether liquid cooling is “the future” and starts deciding precisely where, how, and for which workloads it belongs.

The winners are not the operators with the most aggressive cooling concepts, but those who can deploy liquid cooling at scale: reliably, repeatably, and without disrupting the rest of the facility. In that sense, liquid cooling’s maturation mirrors the broader trajectory of the industry itself: fewer experiments, tighter playbooks, and a growing emphasis on execution over ambition.

7. Speed Meets Gravity: Accelerated Deployment Tests the Limits of Edge Scale-Out

As the data center industry moves into 2026, the impulse to build faster is unmistakable. Modular construction, prefabrication, standardized power blocks, and repeatable designs are in no way experimental techniques; they are becoming default responses to compressed timelines, constrained power availability, and hyperscale demand.

What remains in flux is where that acceleration ultimately expresses itself.

For years, edge computing has been positioned as a counterweight to hyperscale concentration. The logic is straightforward: latency-sensitive workloads, distributed inference, and data-local processing should push compute outward, closer to users, devices, and data sources. In 2026, those use cases continue to grow, but they coexist with a persistent gravitational pull toward centralized infrastructure.

Power and Physics Still Favor the Core

The most demanding AI workloads (training, large-scale inference, and sustained GPU utilization) continue to favor centralized environments. AI factories require firm power, dense interconnection, liquid cooling at scale, and operational maturity that remains difficult to replicate economically at the edge.

As a result, even as edge deployments expand, the largest capital commitments remain anchored to megacampuses that can support utility coordination, onsite or hybrid power strategies, and standardized delivery at scale. The industry’s fastest-growing workloads still pull infrastructure inward, not outward.

Where Edge Expansion Is Likely to Materialize

That does not mean edge infrastructure stalls. Instead, it becomes more targeted.

In 2026, edge-leaning deployments are most likely to gain traction in vertical-specific use cases rather than as a generalized alternative to centralized AI infrastructure. These include healthcare imaging and diagnostics, manufacturing and logistics automation, autonomous and transportation systems, and retail or municipal analytics where latency, data locality, or regulatory constraints justify localized compute.

Geographically, this expansion favors secondary markets that combine strong fiber connectivity, moderate power availability, and proximity to population centers; markets such as Raleigh-Durham, Minneapolis, Salt Lake City, Denver, and Columbus fit this bill. These locations sit close enough to users to matter, but are large enough to support repeatable deployment models.

Modular and Prefabricated By Design

Modular and prefabricated data center designs play a central role in this evolution. In edge contexts, these approaches are not about maximizing density; they are about bounding complexity. Factory-built power and cooling systems, containerized enclosures, and pre-engineered modules shorten deployment timelines and reduce execution risk.

These designs emphasize scale-out rather than scale-up. They trade peak density for predictability, accepting smaller footprints and lower per-site capacity in exchange for faster delivery and clearer operational limits. In doing so, they make edge deployments viable where bespoke builds would struggle to pencil.

A More Disciplined Topology Emerges

By 2026, the industry is converging on a more nuanced infrastructure topology. Dense, power-intensive AI factories anchor the core. Lighter, purpose-built facilities extend compute outward where workloads demand proximity rather than sheer scale.

Accelerated deployment strategies are central to both; but they express themselves differently depending on physics, power, and workload profile. Speed matters. So do boundaries.

Bottom line: The edge scale-out continues in 2026, but within limits defined by gravity.

8. The Data Center Industry Begins Planning for Volatility, Not Just Growth

This may be the most under-discussed shift heading into 2026.

After several years of relentless expansion driven first by cloud, then by AI, the data center industry now begins to quietly ask different questions. Not about how fast it can build, but about how its assets behave if conditions change.

In 2026, forward-looking operators, developers, and investors may increasingly ask:

  • What happens if AI demand moderates or fragments?
  • How reusable are these facilities beyond their first workload?
  • How do you pause, phase, or mothball capacity without writing it off?
  • What does a soft landing look like in an industry built for acceleration?

This line of inquiry is not pessimism. It indicates institutional maturity.

Designing for Optionality, Not Just Peak Demand

On the physical track, facilities are increasingly designed with reuse, reconfiguration, and staged expansion in mind. The industry moves away from single-purpose, all-or-nothing builds toward layouts that can absorb change.

That shows up in more flexible power distribution, modular cooling zones, convertible halls, and campus designs that allow capacity to be added (or deferred) without stranding sunk costs. Even AI factory projects may begin incorporating assumptions about secondary uses, phased densification, or partial repurposing over time.

The goal is not to dilute performance at peak demand. It is to preserve value if demand curves flatten, shift, or bifurcate.

Capital Structures Begin to Reflect Cycles

On the structural track, capital discipline deepens. In 2026, the industry begins planning explicitly for variability rather than assuming uninterrupted growth.

Developers and investors pay closer attention to duration mismatch, utilization risk, and depreciation cycles, particularly in GPU-dense environments where hardware refresh timelines move faster than real estate amortization. Lease structures, joint ventures, and financing models increasingly reflect phased delivery, optional expansion, and downside protection.

This trend does not signal retreat from AI infrastructure. It signals a more realistic understanding of how technology cycles behave over time.

From Expansion Mindset to Resilience Mindset

What changes most in 2026 is tone.

The industry does not stop building. It does not pull back from AI. But it begins to acknowledge that infrastructure built at this scale must endure more than one market condition. Designing for volatility – operational, financial, and technological – becomes part of responsible planning rather than an admission of doubt.

In that sense, this forecast point is less about preparing for decline than about earning durability. The data center industry enters 2026 still expanding, but no longer pretending that growth is the only state it needs to survive.

Honorable Mentions: 7 Additional Pressure Points the Data Center Industry Can’t Ignore in 2026

Beyond these eight defining trends, a set of quieter shifts is reshaping how the industry plans, staffs, and operates at scale. These themes may not define the center of gravity in 2026, but they increasingly influence how the industry executes against its core challenges.

In many cases, they function less as standalone trends than as pressure points: areas where technology, operations, and institutional behavior are quietly evolving in response to scale.

1. Digital Twins Move From Planning Tool to Operational Infrastructure

Digital twins are shedding their early identity as design-time visualization aids and moving toward something more consequential: an operational abstraction layer for complex infrastructure. In 2026, their value increasingly lies in real-time modeling of power flows, thermal behavior, equipment stress, and failure propagation; particularly in AI-dense environments where margins for error are thin.

What is changing is not the concept, but the use case. As facilities grow larger and more interdependent, operators need ways to simulate decisions before executing them: i.e. how a cooling adjustment affects power draw, how phased expansion alters redundancy, how maintenance schedules intersect with utilization peaks. Digital twins offer a way to make those tradeoffs explicit.

Adoption remains uneven. Integration with legacy systems is complex, data fidelity varies, and organizational ownership is often unclear. But where scale and density converge, digital twins are becoming increasingly less optional, and more infrastructural.

2. Workforce Pressures Shift From Hiring to Retention and Specialization

The workforce challenge facing the data center industry does not lessen in 2026. It matures.

The most acute constraint is no longer raw headcount, but the availability of highly specific skill combinations: technicians fluent in liquid cooling systems, engineers comfortable operating at higher voltages, managers capable of bridging IT, facilities, and energy disciplines. As systems become more integrated, the cost of turnover rises.

This shifts the focus from hiring to retention, training, and institutional continuity. Knowledge transfer, career pathways, and operational resilience increasingly matter as much as staffing numbers. In an environment where execution risk is high, experience becomes an asset that compounds over time.

3. Energy Storage Becomes a Conditional Infrastructure Lever

By 2026, battery and energy storage systems are no longer theoretical in the data center industry, but neither are they universal. Storage is increasingly evaluated as a situational tool for load shaping, interconnection timing, and resilience, particularly in regions with constrained grids or volatile delivery schedules.

What limits broader adoption is not relevance, but variability. The value of storage differs dramatically by geography, regulatory framework, utility posture, and workload profile. In some markets, it materially alters execution risk or accelerates delivery. In others, it adds cost without unlocking meaningful capacity.

The result is not neglect, but selectivity. Energy storage is becoming an important part of the power toolkit; deployed deliberately where it changes outcomes, rather than assumed as a default layer of every data center design.

4. Cybersecurity Expands From IT Concern to Infrastructure Reality

As data centers become more software-defined, energy-integrated, and operationally interconnected, cybersecurity expands beyond the traditional IT perimeter. In 2026, attention increasingly turns to operational technology (OT), energy interfaces, building management systems, and the control layers that tie physical infrastructure together.

This shift is still early. Many organizations remain structured around legacy distinctions between IT and facilities. But the direction is clear: as infrastructure becomes programmable, it also becomes addressable, and therefore vulnerable.

Cybersecurity is beginning to be discussed not just as a compliance requirement, but as an element of infrastructure resilience. That conversation is only starting, but it will grow louder as systems converge.

5. Sustainability Becomes a Tradeoff Exercise, Not a Slogan

By 2026, sustainability discussions in the data center industry are more pragmatic and more constrained.

Ambitious targets remain, but the industry increasingly acknowledges the tradeoffs involved in meeting them under real-world conditions. Speed-to-market, grid reliability, power availability, and regional constraints all shape what is achievable. The result is a shift from aspirational framing toward explicit prioritization.

This does not represent abandonment of sustainability goals. It reflects a more honest reckoning with scale. Sustainability becomes something that must be negotiated between power sources, timelines, and stakeholders – rather than assumed as a default outcome.

6. Nuclear Power Shapes Planning Long Before It Shapes Power Bills

Nuclear energy continues to exert outsized influence on data center planning despite limited near-term deployment. Small modular reactors and microreactors increasingly inform site selection, land control strategies, and long-term utility relationships, even as most operators acknowledge that meaningful capacity remains years away.

In 2026, nuclear functions less as an execution tool than as a strategic signal. It shapes where campuses are planned, how partnerships are structured, and which regions are considered viable for long-duration growth. Its impact is real, but temporal.

7. Onsite and Hybrid Power Models Multiply Without Converging

Beyond the headline narratives of gas and nuclear, the industry experiments with an expanding array of hybrid power models: partial generation ownership, utility co-investment, phased interconnection, storage overlays, and future-fuel optionality.

What defines 2026 is not convergence, but diversity. Power strategies are increasingly bespoke, shaped by geography, utility posture, regulatory frameworks, and workload class. This fragmentation reflects adaptation, not confusion. It is the natural outcome of an industry operating under constraint.

Over time, patterns may emerge. In the near term, the power stack remains situational.

Why These Forces Matter

Individually, none of these “pressure point” themes defines the data center industry in 2026. Collectively, they explain how the industry is learning to operate at scale: absorbing complexity, accepting tradeoffs, and prioritizing execution over novelty.

They are the connective tissue beneath the headline trends. And they suggest that the most important changes underway are not always the loudest ones, but the ones quietly reshaping how decisions get made.

Why These Trends—and Why Now

Taken together, these trends are less a forecast of disruption than a portrait of an industry growing into its own consequences.

What distinguishes 2026 from earlier cycles is not the emergence of any single technology or business model, but the convergence of scale, visibility, and constraint. AI did not merely increase demand; it exposed the limits of existing assumptions about power, cooling, capital, and execution. Growth did not slow: but it became heavier, more physical, and harder to abstract away.

That is why so many of this year’s defining forces are not about invention, but about translation: translating AI ambition into buildable infrastructure, translating utility realities into development strategy, translating capital availability into disciplined deployment, and translating public scrutiny into operational legitimacy.

In earlier eras, the data center industry could afford to treat friction as temporary. Power would arrive. Permits would clear. Communities would acclimate. Capital would follow growth. In 2026, those assumptions no longer hold uniformly, and the industry knows it.

What replaces them is not retrenchment, but realism.

The most telling shift running through this forecast is the industry’s growing comfort with specificity. Not every data center is the same. Not every workload justifies the same density. Not every market can absorb the same scale. And not every year will reward speed over durability. These distinctions, once glossed over, are now central to how projects are conceived, financed, and delivered.

That is why the “frontier” in this forecast looks different than it once did. It is less about what comes next and more about how well the industry operates at the scale it has already reached. Execution, coordination, and resilience have become as strategic as technology choice.

If there is a throughline to 2026, it is this: the data center industry is no longer building toward inevitability. It is building toward sustainability in the broadest sense: technical, financial, operational, and social.

These trends matter now because the industry has reached a point where momentum alone is no longer sufficient. The next phase will be defined not by who builds the most, but by who builds with the clearest understanding of risk, responsibility, and reuse.

That doesn’t mean the end of growth. It may denote the beginning of a new kind of durability.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

How AWS is reinventing the telco revenue model

Consider what that means for the mobile operator and its relationship with its customers. Instead of selling a generic 5G pipe with a static SLA, a telco can now sell a dynamic, guaranteed slice for a specific use case—say, a remote robotic surgery setup or a high-density, low-latency industrial IoT

Read More »

What’s the biggest barrier to AI success?

AI’s challenge starts with definition. We hear all the time about how AI raises productivity, and many have experienced that themselves. But what, exactly, does “productivity” mean? To the average person, it means they can do things with less effort, which they like, so it generates a lot of favorable

Read More »

IBM proposes unified architecture for hybrid quantum-classical computing

Quantum computers and classical HPC are traditionally “disparate systems [that] operate in isolation,” IBM researchers explain in a new paper. This can be “cumbersome,” because users have to manually orchestrate workflows, coordinate scheduling, and transfer data between systems, thus hindering productivity and “severely” limiting algorithmic exploration. But a hybrid approach

Read More »

Brent retreats from highs after Trump signals Iran war nearing end

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: var(–color-primary-main); } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style { font-size: 0.75rem; opacity: .6; } #onetrust-pc-sdk [id*=btn-handler], #onetrust-pc-sdk [class*=btn-handler] { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-policy a, #onetrust-pc-sdk a, #ot-pc-content a { color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-pc-sdk .ot-active-menu { border-color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-accept-btn-handler, #onetrust-banner-sdk #onetrust-reject-all-handler, #onetrust-consent-sdk #onetrust-pc-btn-handler.cookie-setting-link { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-consent-sdk .onetrust-pc-btn-handler { color: #c19a06 !important; border-color: #c19a06 !important; } Oil futures eased from recent highs Tuesday as markets reacted to comments from US President Donald Trump suggesting the war with Iran may be nearing its conclusion, easing concerns about prolonged disruptions to Middle East crude supplies. Brent crude had climbed above $100/bbl amid escalating tensions in the region and fears that the war could prolong disruptions to shipments through the Strait of Hormuz—one of the world’s most critical energy chokepoints and a transit route for roughly one-fifth of global oil supply. Prices pulled back after Pres. Trump said the war was “almost done,” prompting traders to reassess the risk premium that had built into crude markets during the latest escalation. The earlier gains were driven by the fact that the war had disrupted tanker traffic in the Strait of Hormuz, raising concerns about wider supply disruptions from major Gulf oil producers. While the latest remarks helped calm markets, analysts note that geopolitical risks remain elevated and price volatility is likely to persist as traders monitor developments in the region. Any renewed escalation could quickly send crude prices higher again.

Read More »

Southwest Arkansas lithium project moves toward FID with 10-year offtake deal

Smackover Lithium, a joint venture between Standard Lithium Ltd. and Equinor, through subsidiaries of Equinor ASA, signed the first commercial offtake agreement for the South West Arkansas Project (SWA Project) with commodities group Trafigura Trading LLC. Under the terms of a binding take-or-pay offtake agreement, the JV will supply Trafigura with 8,000 metric tonnes/year (tpy) of battery-quality lithium carbonate (Li2CO3) over a 10-year period, beginning at the start of commercial production. Smackover Lithium is expected to achieve final investment decision (FID) for the project, which aims to use direct lithium extraction technology to produce lithium from brine resources in the Smackover formation in southern Arkansas, in 2026, with first production anticipated in 2028. The project encompasses about 30,000 acres of brine leases in the region, with the initial phase of project development focused on production from the 20,854-acre Reynolds Brine Unit.   Front-end engineering design was completed in support of a definitive feasibility study with a principal recommendation that the project is ready to progress to FID.  While pricing terms of the Trafigura deal were kept confidential, Standard Lithium said they are “structured to support the anticipated financing for the project.” The JV is seeking to finalize customer offtake agreements for roughly 80% of the 22,500 tonnes of annual nameplate lithium carbonate capacity for the initial phase of the project. This agreement represents over 40% of the targeted offtake commitments. Formed in 2024, Smackover Lithium is developing multiple DLE projects in Southwest Arkansas and East Texas. Standard Lithium is operator of the projecs with 55% interest. Equinor holds the remaining 45% interest.

Read More »

Equinor makes oil and gas discoveries in the North Sea

Equinor Energy AS discovered oil in the Troll area and gas and condensate in the Sleipner area of the North Sea. Byrding C discovery well 35/11-32 S in production license (PL) 090 HS was made 5 km northwest of Fram field in Troll. The well was drilled by the COSL Innovator rig in 373 m of water to 3,517 m TVD subsea. It was terminated in the Heather formation from the Middle Jurassic. The primary exploration target was to prove petroleum in reservoir rocks from the Late Jurassic deep marine equivalent to the Sognefjord formation. The secondary target was to prove petroleum and investigate the presence of potential reservoir rocks in two prospective intervals from the Middle Jurassic in deep marine equivalents to the Fensfjord formation. The well encountered a 22-m oil column in sandstone layers in the Sognefjord formation with a total thickness of 82 m, of which 70 m was sandstone with moderate to good reservoir properties. The oil-water contact was encountered. The secondary exploration target in the Fensfjord formation did not prove reservoir rocks or hydrocarbons. The well was not formation-tested, but data and samples were collected. The well has been permanently plugged. Preliminary estimates indicate the size of the discovery is 4.4–8.2 MMboe. Oil discovered in Byrding C will be produced using existing or future infrastructure in the area. The Frida Kahlo discovery was drilled from the Sleipner B platform in production license PL 046 northwest of Sleipner Vest and is estimated to contain 5–9 MMboe of gas and condensate. The well will be brought on stream as early as April. The four most recent exploration wells in the Sleipner area, drilled over a 3-month period, include Lofn, Langemann, Sissel, and Frida Kahlo. All have all proven gas and condensate in the Hugin formation, with combined estimated

Read More »

IEA launches record strategic oil release as Middle East war disrupts supply

The International Energy Agency (IEA) on Mar. 11 approved the largest emergency oil stock release in its history, making 400 million bbl available from member-country reserves in response to market disruptions tied to the war in the Middle East. The coordinated action, agreed unanimously by the IEA’s 32 member countries, is intended to ease supply pressure and temper price volatility as crude markets react to disrupted flows through the Strait of Hormuz. “The conflict in the Middle East is having significant impacts on global oil and gas markets, with major implications for energy security, energy affordability and the global economy for oil,” IEA executive director Fatih Birol said. The release more than doubles the previous IEA record set in 2022, when member countries collectively made 182.7 million bbl available following Russia’s invasion of Ukraine. Under the IEA system, member countries are required to maintain emergency oil stocks equal to at least 90 days of net imports, giving the agency a mechanism to respond when severe disruptions threaten global supply. The move comes after crude prices surged amid concerns that the US-Iran war could lead to prolonged disruption of exports from the Gulf. Despite the planned stock release, traders remain uncertain about whether reserve barrels alone will be enough to offset losses if the disruption persists. IEA said the emergency barrels will be supplied to the market from government-controlled and obligated industry stocks held across member countries. The action marks the sixth coordinated stock release in the agency’s history and underscores the seriousness of the current supply shock. Earlier the day, Japanese Prime Minister Sanae Takaichi said that Japan might start using its strategic oil reserves as early as next week, citing Japan’s unusually high dependence on Middle Eastern crude oil.

Read More »

Infographic: Strait of Hormuz energy trade 2025

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: var(–color-primary-main); } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style { font-size: 0.75rem; opacity: .6; } #onetrust-pc-sdk [id*=btn-handler], #onetrust-pc-sdk [class*=btn-handler] { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-policy a, #onetrust-pc-sdk a, #ot-pc-content a { color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-pc-sdk .ot-active-menu { border-color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-accept-btn-handler, #onetrust-banner-sdk #onetrust-reject-all-handler, #onetrust-consent-sdk #onetrust-pc-btn-handler.cookie-setting-link { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-consent-sdk .onetrust-pc-btn-handler { color: #c19a06 !important; border-color: #c19a06 !important; } Coordinated attacks Feb. 28 by the US and Israel on Iran and the since-escalated conflict have nearly halted shipping traffic through the Strait of Hormuz, which typically carries about 20% of the world’s crude oil and natural gas. OGJ Statistics Editor Laura Bell-Hammer compiled data to showcase 2025 energy trade through the critical transit chokepoint.   <!–> –> <!–> ]–> <!–> ]–>

Read More »

BOEM: US OCS holds 65.8 billion bbl of technically recoverable reserves

The US Outer Continental Shelf (OCS) holds mean undiscovered technically recoverable resources (UTRR) of 65.8 billion bbl of oil and 218.43 tcf of natural gas, the US Bureau of Ocean Energy Management (BOEM) said Mar. 9. Based on current production trends, these undiscovered resources represent the potential for 100 or more years of energy production from the US Outer Continental Shelf (OCS), BOEM said. A large portion of undiscovered OSC resources is located offshore the Gulf of Mexico and Alaska, according to the report. The offshore Gulf holds 26.9 million bbl of oil and 45.59 tcf of gas, while offshore Alaska holds an estimated mean 24.1 million bbl of oil and 122.29 tcf of gas. Offshore Pacific holds a mean UTRR of 10.3 million barrels of oil and 16.2 trillion cubic feet of gas, the report said. Offshore Atlantic holds a mean UTRR of 10.3 billion barrels of oil and 16.2 trillion cubic feet of gas. The assessment also evaluates the impact of prices on hydrocarbon recovery. Alaska is particularly price-sensitive, with mean undiscovered economically recoverable resources (UERR) negligible until prices average $100/bbl and $17.79/Mcf. At those levels, the mean UERR stands at 6.25 billion bbl and 13.25 tcf. At $160/bbl and $28.47/Mcf, recoverable resources jump to 14.67 billion bbl and 58.78 tcf. In the Gulf of Mexico, the mean UERR is 17.51 billion bbl of oil and 13.71 tcf at average prices of $60/bbl and $3.20/Mcf, increasing to 20.51 billion bbl and 17.49 tcf at average prices of $100/bbl and $5.34/Mcf, respectively. BOEM conducts a national resource assessment every 4 years to understand the “distribution of undiscovered oil and gas resources on the OCS” and identify opportunities for additional oil and gas exploration and development. “The Outer Continental Shelf holds tremendous resource potential,” said BOEM Acting Director Matt Giacona. “This

Read More »

Data mining? Old servers could become new source of rare earths

For decades, he said, “the retirement of data center equipment was treated almost entirely as a compliance and disposal issue. Enterprises focused on secure decommissioning, certified recycling, and documented destruction of sensitive hardware. Once equipment left production environments, its economic life was assumed to be largely finished.” That assumption, he pointed out, “is beginning to change, because the hardware inside modern data centres contains a wide range of strategically important materials. Servers, storage systems, networking equipment, and power components contain copper, aluminum, silver, gold, and increasingly small but significant quantities of rare earth elements and other critical minerals.” These materials play a vital role in the manufacturing of semiconductors, energy systems, defense electronics, and advanced computing infrastructure, he explained, noting, “as global demand for digital infrastructure continues to expand, the volume of retired hardware entering disposal channels is rising quickly.” Electronic waste has already become one of the fastest growing waste streams in the world. “Global volumes now exceed 60 million tonnes annually and are projected to move toward eighty million tonnes by the end of the decade if current trends continue,” he said. “Data center infrastructure represents only a portion of that total, but it is a particularly important portion because it is concentrated, professionally managed, and replaced in structured cycles.” For a metals producer, he said, data center infrastructure represents a highly attractive feedstock, because unlike consumer electronics, enterprise hardware is replaced in large batches and flows through professional asset management channels. That predictability, said Gogia, “allows recyclers to design specialized processes that target specific components and materials. Over time, this creates the foundation for an industrial scale circular supply chain in which retired electronics feed back into the production of new materials.”

Read More »

Meta is developing more AI chips for itself

With demand for AI chips rising and supplies tightening, Meta is taking its AI computing needs into its own hands and developing more of its own chips: It will produce four new generations of chips over the next two years. Cloud computing giants including Meta, AWS, and Google have been keen to develop their own chips to improve the performance of their own data centers. Meta started its own chip program in 2023, when it implemented the Meta Training and Inference Accelerator (MTIA), a family of custom-built silicon chips to power its AI workloads efficiently. The MTIA 300, which Meta will use for ranking and recommendations training, is already  in production, Meta said. It will use the other planned chips, the MTIA 400, 450, and 500, mainly for generative AI inference production, it said.

Read More »

Arista targets AI data centers with new liquid cooled pluggable optic module

To prove their point, the authors imagined a 400 MW AI datacenter with 1024 GPU racks of 128 GPUs each for a total of 128,000 GPUs. “Assume 12.8T scale-up and 1.6T scale-out bandwidth per GPU. With OSFP switch racks that have a density of 1.6 Pbps per rack, this would require more than 1,400 switch racks for scale-up and scale-out fabrics. With XPO, this would require 75% fewer racks, saving over 1,050 racks or 44 % of the floor space,” Bechtolsheim and Vusirikala stated in the blog.  “Eliminating 75% of switch racks translates to massive reductions in construction and infrastructure costs, including power distribution, plumbing and installation costs, while accelerating deployment timelines,” Bechtolsheim and Vusirikala stated. Arista said the water-cooling capability of XPO is also an important feature. “All large AI data centers will be liquid cooled and the switches that go into these data centers also need to be liquid cooled,” Bechtolsheim and Vusirikala stated.  “While one can add liquid cooled cold plates on flat-top OSFP modules, this does not substantially improve thermal performance.” XPO solves this problem by integrating a liquid cold plate inside the module, with two 32-channel paddle cards sharing the common cold plate which can cool both low power as well as high-power optics such as 8x1600G-ZR/ZR+ with up to 400W of power, Bechtolsheim and Vusirikala stated. XPO modules are much simpler than OSPF modules which improves reliability as well. “Each 32-channel paddle card has only one microcontroller and one set of voltage converters, a 75% reduction in common components versus 4 OSFPs,” Bechtolsheim and Vusirikala wrote. 

Read More »

Cisco grows high-end optical support for AI clusters

Cisco has also upgraded its Network Conversion System (NCS) with a 1RU, 800GE line card offering 12.8T capacity, with 32 OSFP-based ports for 100GE, 400GE, and 800GE clients and 800ZR/ZR+ WDM trunks. The NCS 1014  doubles the density of previous-generation NCS versions and now includes MACsec encryption (IEEE 802.1AE) to secure point-to-point links with hardware-based encryption, data integrity, and authentication for Ethernet traffic, Ghioni stated. It supports enhanced capacity and performance with C&L-band support and NCS 1014 systems with the 2.4T WDM line card based on the Coherent Interconnect Module 8 and now supports 800 GE clients, which can be mapped directly to a wavelength or inverse multiplexed across two wavelengths to maximize reach, Ghioni wrote.  In the pluggable optic arena, Cisco is now offering a Quad Small Form Factor Pluggable Double Density (QSFP-DD) Pluggable Protection Switch Module that can monitor the optical link and switch traffic if it detects a fault in less than 50 milliseconds. The module occupies a quarter of the rack space compared to traditional protection devices—offering 90% rack space saving over available options, Ghioni wrote.  It is aimed at Metro and DCI network customers where sub-50 ms failure recovery is essential and data centers needing fiber protection without bulky hardware, Ghioni stated.  Cisco also added its Acacia developed Bright QSFP28 100ZR 0 dBm coherent optical pluggable in a standard QSFP28 form factor.  It is aimed at edge, access, enterprise, and campus network deployment. Cisco has been actively growing its optical portfolio  recently adding the Cisco Silicon One G300, which powers 102.4T N9000 and Cisco 8000 systems, as well as advanced 1.6T OSFP optics and 800G Linear Pluggable Optics. 

Read More »

Datalec targets rapid infrastructure deployment with new modular data centers

“We are engineering the data center with a new lens bringing pre-engineered system designs that are flexible and adaptable that enables a tailored solution for clients,” said John Lever, director of modular solutions at Datalec. The systems are flexible enough that these solutions cater for all types of data center, from standard server technology to AI and high-density compute. Datalec also provides “bolt-on” solutions, including a ‘digital wrapper’ including digital twinning and lifecycle and global support, Lever says. Another way Datalec says it differentiates from competing modular designs is a larger share of work is done offsite in a controlled manufacturing environment, which cuts onsite construction time, improves safety and limits disruption to live facilities, Lever says. The company competes with other modular data center vendors including Schneider Electric, Vertiv, Flex many others. DPI’s says its services are aimed at colocation providers, hyperscale and AI infrastructure teams, and large enterprises that need to add capacity quickly, safely and cost effectively across multiple regions.

Read More »

Study finds significant savings from direct current power for AI workloads

The result is a 50% to 80% reduction in copper usage, due to fewer conductors and less parallel cabling, and an 8% to 12% reduction in annual energy-related OpEx through lower conversion and distribution losses. By reducing conductor count, cabling, and redundant power components, 800VDC enables meaningful savings at both build-out and operational stages. AI-first facilities can see a $4 million to $8 million in CapEx savings per 10 MW build by reducing upstream AC. For a one-gigawatt data center, you’re saving a couple million pounds of copper wire, he said. Burke says an all-DC data center is best done with a whole new facility rather than retrofitting old facilities. “[DC] is going to be in a lot of greenfield data centers that are going to be built, and data centers that are going to go to higher compute power are also going to DC,” he said. He did recommend all-DC retrofits for existing data centers that are going to employ high power computing with GPUs. Enteligent’s unnamed and as yet unreleased product is a converter that takes 800 volts and partitions it to 50 volts for the computing servers. The company will provide a new power supply, power shelf that converts 800 volts DC to 50 volts DC much more efficiently than any current power supplies. Burke said the company is doing NDA level testing and pilot programs now with its product, but it will be making a formal announcement within the next few weeks. There are a number of players in the DC arena focusing on different parts of the power supply market including Vertiv, Rutherford, Siemens, Eaton and many more.

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »