Stay Ahead, Stay ONMINE

From MIPS to exaflops in mere decades: Compute power is exploding, and it will transform AI

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More At the recent Nvidia GTC conference, the company unveiled what it described as the first single-rack system of servers capable of one exaflop — one billion billion, or a quintillion, floating-point operations (FLOPS) per second. This […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


At the recent Nvidia GTC conference, the company unveiled what it described as the first single-rack system of servers capable of one exaflop — one billion billion, or a quintillion, floating-point operations (FLOPS) per second. This breakthrough is based on the latest GB200 NVL72 system, which incorporates Nvidia’s latest Blackwell graphics processing units (GPUs). A standard computer rack is about 6 feet tall, a little more than 3 feet deep and less than 2 feet wide.

Shrinking an exaflop: From Frontier to Blackwell

A couple of things about the announcement struck me. First, the world’s first exaflop-capable computer was installed only a few years ago, in 2022, at Oak Ridge National Laboratory. For comparison, the “Frontier” supercomputer built by HPE and powered by AMD GPUs and CPUs, originally consisted of 74 racks of servers. The new Nvidia system has achieved roughly 73X greater performance density in just three years, equivalent to a tripling of performance every year. This advancement reflects remarkable progress in computing density, energy efficiency and architectural design.

Secondly, it needs to be said that while both systems hit the exascale milestone, they are built for different challenges, one optimized for speed, the other for precision. Nvidia’s exaflop specification is based on lower-precision math — specifically 4-bit and 8-bit floating-point operations — considered optimal for AI workloads including tasks like training and running large language models (LLMs). These calculations prioritize speed over precision. By contrast, the exaflop rating for Frontier was achieved using 64-bit double-precision math, the gold standard for scientific simulations where accuracy is critical.

We’ve come a long way (very quickly)

This level of progress seems almost unbelievable, especially as I recall the state-of-the-art when I began my career in the computing industry. My first professional job was as a programmer on the DEC KL 1090. This machine, part of DEC’s PDP-10 series of timeshare mainframes, offered 1.8 million instructions per second (MIPS). Aside from its CPU performance, the machine connected to cathode ray tube (CRT) displays via hardwired cables. There were no graphics capabilities, just light text on a dark background. And of course, no Internet. Remote users connected over phone lines using modems running at speeds up to 1,200 bits per second.

DEC System 10; Source: By Joe Mabel, CC BY-SA 3.0.

500 billion times more compute

While comparing MIPS to FLOPS gives a general sense of progress, it is important to remember that these metrics measure different computing workloads. MIPS reflects integer processing speed, which is useful for general-purpose computing, particularly in business applications. FLOPS measures floating-point performance that is crucial for scientific workloads and the heavy number-crunching behind modern AI, such as the matrix math and linear algebra used to train and run machine learning (ML) models.

While not a direct comparison, the sheer scale of the difference between MIPS then and FLOPS now provides a powerful illustration of the rapid growth in computing performance. Using these as a rough heuristic to measure work performed, the new Nvidia system is approximately 500 billion times more powerful than the DEC machine. That kind of leap exemplifies the exponential growth of computing power over a single professional career and raises the question: If this much progress is possible in 40 years, what might the next 5 bring?

Nvidia, for its part, has offered some clues. At GTC, the company shared a roadmap predicting that its next-generation full-rack system based on the “Vera Rubin” Ultra architecture will deliver 14X the performance of the Blackwell Ultra rack shipping this year, reaching somewhere between 14 and 15 exaflops in AI-optimized work in the next year or two.

Just as notable is the efficiency. Achieving this level of performance in a single rack means less physical space per unit of work, fewer materials and potentially lower energy use per operation, although the absolute power demands of these systems remain immense.

Does AI really need all that compute power?

While such performance gains are indeed impressive, the AI industry is now grappling with a fundamental question: How much computing power is truly necessary and at what cost? The race to build massive new AI data centers is being driven by the growing demands of exascale computing and ever-more capable AI models.

The most ambitious effort is the $500 billion Project Stargate, which envisions 20 data centers across the U.S., each spanning half a million square feet. A wave of other hyperscale projects is either underway or in planning stages around the world, as companies and countries scramble to ensure they have the infrastructure to support the AI workloads of tomorrow.

Some analysts now worry that we may be overbuilding AI data center capacity. Concern intensified after the release of R1, a reasoning model from China’s DeepSeek that requires significantly less compute than many of its peers. Microsoft later canceled leases with multiple data center providers, sparking speculation that it might be recalibrating its expectations for future AI infrastructure demand.

However, The Register suggested that this pullback may have more to do with some of the planned AI data centers not having sufficiently robust ability to support the power and cooling needs of next-gen AI systems. Already, AI models are pushing the limits of what present infrastructure can support. MIT Technology Review reported that this may be the reason many data centers in China are struggling and failing, having been built to specifications that are not optimal for the present need, let alone those of the next few years.

AI inference demands more FLOPs

Reasoning models perform most of their work at runtime through a process known as inference. These models power some of the most advanced and resource-intensive applications today, including deep research assistants and the emerging wave of agentic AI systems.

While DeepSeek-R1 initially spooked the industry into thinking that future AI might require less computing power, Nvidia CEO Jensen Huang pushed back hard. Speaking to CNBC, he countered this perception: “It was the exact opposite conclusion that everybody had.” He added that reasoning AI consumes 100X more computing than non-reasoning AI.

As AI continues to evolve from reasoning models to autonomous agents and beyond, demand for computing is likely to surge once again. The next breakthroughs may come not just in language or vision, but in AI agent coordination, fusion simulations or even large-scale digital twins, each made possible by the kind of computing ability leap we have just witnessed.

Seemingly right on cue, OpenAI just announced $40 billion in new funding, the largest private tech funding round on record. The company said in a blog post that the funding “enables us to push the frontiers of AI research even further, scale our compute infrastructure and deliver increasingly powerful tools for the 500 million people who use ChatGPT every week.”

Why is so much capital flowing into AI? The reasons range from competitiveness to national security. Although one particular factor stands out, as exemplified by a McKinsey headline: “AI could increase corporate profits by $4.4 trillion a year.”

What comes next? It’s anybody’s guess

At their core, information systems are about abstracting complexity, whether through an emergency vehicle routing system I once wrote in Fortran, a student achievement reporting tool built in COBOL, or modern AI systems accelerating drug discovery. The goal has always been the same: To make greater sense of the world.

Now, with powerful AI beginning to appear, we are crossing a threshold. For the first time, we may have the computing power and the intelligence to tackle problems that were once beyond human reach.

New York Times columnist Kevin Roose recently captured this moment well: “Every week, I meet engineers and entrepreneurs working on AI who tell me that change — big change, world-shaking change, the kind of transformation we’ve never seen before — is just around the corner.” And that does not even count the breakthroughs that arrive each week.

Just in the past few days, we’ve seen OpenAI’s GPT-4o generate nearly perfect images from text, Google release what may be the most advanced reasoning model yet in Gemini 2.5 Pro and Runway unveil a video model with shot-to-shot character and scene consistency, something VentureBeat notes has eluded most AI video generators until now.

What comes next is truly a guess. We do not know whether powerful AI will be a breakthrough or breakdown, whether it will help solve fusion energy or unleash new biological risks. But with ever more FLOPS coming online over the next five years, one thing seems certain: Innovation will come fast — and with force. It is clear, too, that as FLOPS scale, so must our conversations about responsibility, regulation and restraint.

Gary Grossman is EVP of technology practice at Edelman and global lead of the Edelman AI Center of Excellence.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Cato Networks augments CASB with genAI security

CASBs sit between an end user and a cloud service to enforce security policies, protect data, and ensure compliance. CASBs provide enterprise network and security teams with information on how end users are accessing and using cloud resources such as data, applications, and services. They provide visibility into cloud usage,

Read More »

8 unusual Linux commands

3. The column command The column command will display text in columns. Here are two examples of how to use it: $ cat staff | columnJohn Doe Lisa Stone Joanne Zahn Eric Docker Ben MatsonMary Berry Elaine Henry David Bloom Sam Adams Sally Rose$ cat staff | column -tJohn DoeMary

Read More »

AI agents vs. agentic AI: What do enterprises want?

The cloud-provider technical people I know don’t like this approach; they see it as likely to raise barriers to the use of their online generative AI services. Enterprises see their AI agent vision as facilitating cloud AI services instead. If there’s one massive AI entity doing everything, then data sovereignty

Read More »

Algae to create 100 new jobs for oil workers in Grangemouth

Plans have been submitted to increase algae production at Grangemouth in a move that could create 100 new jobs. The firm behind the project, MiAlgae, has said that the expansion will enable it to continue creating “retraining opportunities for workers transitioning from the oil and gas industry”. The business secured £13.8 million in a Series A funding round last year to finance the scale-up. Douglas Martin, founder and managing director of MiAlgae, said: “Grangemouth has incredible potential for us as we look to the next stage of our growth. “This location offers an ideal position to support our scaling efforts and meet the increasing demand for our ‘biotech for good’ solutions, with the creation of green jobs across engineering, production, and research and development. “We are confident that this new facility will help build a greener future and bring high-quality, sustainable jobs to the local community.” © Supplied by MiAlgaeDouglas Martin, founder and managing director of MiAlgae. Grangemouth worker woes Workers in Grangemouth are currently facing career uncertainty as the owners of Scotland’s only oil refinery look to shut up shop. Petroineos launched the first wave of redundancies at the Grangemouth oil refinery early this year as it aims to close that part of the plant. The firm is set to cut 400 jobs at the site in the coming months in a move that unions labelled a “national disgrace”. © Andrew Milligan/PA WireMembers of Unite union take part in a demonstration to protest at Petroineos plans to close Grangemouth oil refinery, during the Scottish Labour Party conference at the Scottish Exhibition Centre (SEC) in Glasgow. Image: Andrew Milligan/PA Wire However, a £1.5 million feasibility study published in March claimed clean energy projects at the Grangemouth refinery could create around 800 jobs over the next 15 years. The government-backed initiative,

Read More »

USA EIA Cuts West Texas Intermediate Oil Price Forecasts

In its latest short term energy outlook (STEO), which was released on April 10, the U.S. Energy Information Administration (EIA) cut its West Texas Intermediate (WTI) oil price forecast for 2025 and 2026. The EIA now sees the WTI spot price averaging $63.88 per barrel in 2025 and $57.48 per barrel in 2026, according to the STEO. In its previous STEO, which was released last month, the EIA projected that the WTI spot price would come in at $70.68 per barrel this year and $64.97 per barrel next year. The EIA forecast in its April STEO that the WTI spot price will average $62.33 per barrel in the second quarter of 2025, $61.67 per barrel in the third quarter, $60 per barrel in the fourth quarter, $59 per barrel in the first quarter of next year, $58 per barrel in the second quarter, $57 per barrel in the third quarter, and $56 per barrel in the fourth quarter. In its previous STEO, the EIA projected that the WTI spot price would average $70.50 per barrel in the second quarter of this year, $71.50 per barrel in the third quarter, $69.52 per barrel in the fourth quarter, $67.50 per barrel in the first quarter of 2026, $65.50 per barrel in the second quarter, $64.50 per barrel in the third quarter, and $62.50 per barrel in the fourth quarter of next year. In its latest STEO, the EIA highlighted that the WTI spot price averaged $71.85 per barrel in the first quarter of this year and $76.60 per barrel overall in 2024. In its previous STEO, the EIA projected that the WTI spot price would average $71.25 per barrel in the first quarter of 2025. That STEO also highlighted that the WTI spot price averaged $76.60 per barrel last year. A research

Read More »

‘The end of an era’: Shell Aberdeen HQ demolition short film

Shell’s iconic Tullos HQ has dominated the Aberdeen skyline for decades, and as demolition wrapped up at the site, a short film was produced to document the process. Love it or hate it, the supermajor’s Aberdeen base stood out with its shining windows and unique architecture since the 1970s. A few years ago, Shell opted to move into the heart of Aberdeen, setting up shop on Union Street, alongside the likes of Neo Energy in the Silver Fin building. To document the process of the building’s teardown, Aberdeen-based film-makers Signal were brought on board. The firm’s managing director took to LinkedIn on Wednesday to share details of the process and share the final cut of the documentary. © Supplied by Kenny Elrick/DC ThomA picture taken during the demolition of Shell’s Tullos HQ. Image: Kenny Elrick/DC Thomson. Jamie Baikie wrote: “Solar-powered time lapse cameras captured the gradual dismantling over a year, supported by ground activities and aerial footage from drones, along with personal interviews with the project team.” Previously, Energy Voice shared fond memories that former workers have of the complex. One former employee told of when “Maggie Thatcher was on the phone” to discuss a World War 2 mine that had rolled up against the Brent export pipeline. Others told stories of workplace highjinks and “awesome” lunches. The Signal managing director continued: “Tullos has been a part of my life for decades, having started working with Shell in 1998 and there’s been an emotional attachment as it holds many fond memories of the people I had the pleasure to work with and the opportunities it gave me. ” © Supplied by Aberdeen Journals LtPremier Margaret Thatcher with her husband at the opening of the Shell complex at Tullos in September 1979. Shell began the demolition of its former Tullos base

Read More »

Wake loss could cost Equinor and Orsted up to £363m from North Sea wind farms

Equinor and Orsted have warned that the effects of wake loss on their projects could cost them up to £363 million in lost revenues. Both developers have submitted financial impact assessments warning that the development of the Outer Dowsing offshore wind farm could have knock-on effects on several of their own developments, including the major Hornsea projects. TotalEnergies, Corio Generation and Gulf Energy Development are currently looking to create Outer Dowsing. This 1.5GW project could potentially locate up to 100 turbines, each up to 403m high, 33 miles (54km) off the Lincolnshire coast. With a consent decision expected this year, construction on the wind farm could begin in 2027 with first power in 2030. However, once completed, Outer Dowsing could lead to wake loss, where its presence affects air currents downwind, impacting other wind farms around it. © Supplied by Outer Dowsing OffshoThe Outer Dowsing offshore wind farm and surrounding projects. What is wake loss? Wake loss describes the effect that wind turbines have on air currents as they pass around them. The turbines absorb energy from the wind, slowing the flow of air and making it more turbulent. For wind farms downwind, wake loss reduces the quality of the air currents and brings down their turbines’ energy yield. According to Wood Thilsted’s report, wake loss can affect wind farms at distances of 18.5-37 miles (30-60km), with impacts even up to 62 miles (100km). While wind dynamics are well understood, the issue of wake loss hasn’t been particularly considered by regulatory bodies and developers when creating wind farms. In the early days of the UK’s offshore wind rollout, the relative scarcity of wind farms meant the effects they have on each other were minimal. But as the UK looks to build more offshore wind farms, the North Sea is becoming

Read More »

Goldman Sachs Warns of Sub-$40 Oil Under One Scenario

In a report sent to Rigzone by the Goldman Sachs team on Monday, Goldman Sachs analysts revealed that they see the Brent oil price falling to under $40 per barrel under one scenario. “In a more extreme and less likely scenario with both a global GDP slowdown and a full unwind of OPEC cuts, which would discipline non-OPEC supply, we estimate that Brent would fall under $40 in late 2025,” the analysts warned in the report. This is one of four “downside price scenarios”, the Goldman Sachs analysts highlighted in the report. The analysts outlined that, under their base case, they expect Brent to average $63 per barrel in the remainder of 2025 and $58 per barrel in 2026. “Our base case assumes that one, the U.S. avoids recession and two, OPEC+ supply rises only modestly,” the Goldman Sachs analysts said in the report. The analysts went on to warn in the report that risks to their oil price forecast are mostly to the downside. Outlining another downside price scenario in the report, the analysts said, “assuming a typical U.S. recession shock and our OPEC baseline, we estimate that Brent would average $61/53 in Bal25/Cal26, respectively”. “In a global GDP slowdown scenario and keeping our OPEC baseline unchanged, we estimate that Brent would average $58/47 in Bal25/Cal26. We estimate a similar price path assuming our GDP baseline and a full unwind of the 2.2 million barrels per day of OPEC8+ cuts,” they added in the report, outlining another downside scenario. “We also see some downside risks to oil prices from a potentially larger rise in OECD commercial stocks even under our supply-demand baseline,” the analysts went on to state in the report. “Specifically, OECD stocks may rise more if floating storage – which tends to build on attractive storage incentives

Read More »

48 employees let go as Aberdeen’s Belmar Engineering enters liquidation

Aberdeen-based Belmar Engineering let go of 48 workers on Tuesday as it entered liquidation. Insolvency expert Michael Reid, who is also Aberdeen-based and works for MHA, has been appointed liquidator. A letter has been sent to all known creditors. There were around 80 in total, made up of trade suppliers, utility providers, finance companies and HMRC. Reid said: “Prior to liquidation, the board tried to sell the business, in whole or in part, but were unsuccessful. “They sought to review operating processes and pricing structures, however, without a base level of work, the factory was not viable.” © Supplied by Google MapsBelmar Engineering’s facility on Aberdeen’s Abbotswell Road. The liquidator added that the decision to close the doors of Belmar Engineering “was taken as the final option”. The equipment at Belmar Engineering’s Abbotswell Road facility is to be determined by a valuer, and an auction is likely to be held. Throughout its 50 years in business, the north-east firm has manufactured subsea equipment and landing strings, including lubricator valves, accumulator mandrels, retainer valves and subsea test trees. It went through a rebrand last year with CP&Co creating new logos and a strap line: “Engineering for excellence since 1975”. “I will ensure that all employees are fully advised about the process for claiming their employment entitlements, including unpaid pay, pension deductions, accrued holiday pay, notice pay and redundancy,” Reid added. © Supplied by MHAMHA’s Michael Reid has been appointed liquidator of Belmar Engineering. “This is the second local engineering company that has appointed me liquidator in the last month, and perhaps reflects the challenges facing this type of business at the moment.” MHA wrote that pressures from international competition, a dwindling order book and losses in recent years have resulted in the current state of the business. Reid was also appointed

Read More »

Intelligence at the edge opens up more risks: how unified SASE can solve it

In an increasingly mobile and modern workforce, smart technologies such as AI-driven edge solutions and the Internet of Things (IoT) can help enterprises improve productivity and efficiency—whether to address operational roadblocks or respond faster to market demands. However, new solutions also come with new challenges, mainly in cybersecurity. The decentralized nature of edge computing—where data is processed, transmitted, and secured closer to the source rather than in a data center—has presented new risks for businesses and their everyday operations. This shift to the edge increases the number of exposed endpoints and creates new vulnerabilities as the attack surface expands. Enterprises will need to ensure their security is watertight in today’s threat landscape if they want to reap the full benefits of smart technologies at the edge. Bypassing the limitations of traditional network security  For the longest time, enterprises have relied on traditional network security approaches to protect their edge solutions. However, these methods are becoming increasingly insufficient as they typically rely on static rules and assumptions, making them inflexible and predictable for malicious actors to circumvent.  While effective in centralized infrastructures like data centers, traditional network security models fall short when applied to the distributed nature of edge computing. Instead, organizations need to adopt more adaptive, decentralized, and intelligent security frameworks built with edge deployments in mind.  Traditional network security typically focuses on keeping out external threats. But today’s threat landscape has evolved significantly, with threat actors leveraging AI to launch advanced attacks such as genAI-driven phishing, sophisticated social engineering attacks, and malicious GPTs. Combined with the lack of visibility with traditional network security, a cybersecurity breach could remain undetected until it’s too late, resulting in consequences extending far beyond IT infrastructures.  Next generation of enterprise security with SASE As organizations look into implementing new technologies to spearhead their business, they

Read More »

Keysight tools tackle data center deployment efficiency

Test and performance measurement vendor Keysight Technologies has developed Keysight Artificial Intelligence (KAI) to identify performance inhibitors affecting large GPU deployments. It emulates workload profiles, rather than using actual resources, to pinpoint performance bottlenecks. Scaling AI data centers requires testing throughout the design and build process – every chip, cable, interconnect, switch, server, and GPU needs to be validated, Keysight says. From the physical layer through the application layer, KAI is designed to identify weak links that degrade the performance of AI data centers, and it validates and optimizes system-level performance for optimal scaling and throughput. AI providers, semiconductor fabricators, and network equipment manufacturers can use KAI to accelerate design, development, deployment, and operations by pinpointing performance issues before deploying in production.

Read More »

U.S. Advances AI Data Center Push with RFI for Infrastructure on DOE Lands

ORNL is also the home of the Center for Artificial Intelligence Security Research (CAISER), which Edmon Begoli, CAISER founding director, described as being in place to build the security necessary by defining a new field of AI research targeted at fighting future AI security risks. Also, at the end of 2024, Google partner Kairos Power started construction of their Hermes demonstration SMR in Oak Ridge. Hermes is a high-temperature gas-cooled reactor (HTGR) that uses triso-fueled pebbles and a molten fluoride salt coolant (specifically Flibe, a mix of lithium fluoride and beryllium fluoride). This demonstration reactor is expected to be online by 2027, with a production level system becoming available in the 2030 timeframe. Also located in a remote area of Oak Ridge is the Tennessee Valley Clinch River project, where the TVA announced a signed agreement with GE-Hitachi to plan and license a BWRX-300 small modular reactor (SMR). On Integrating AI and Energy Production The foregoing are just examples of ongoing projects at the sites named by the DOE’s RFI. Presuming that additional industry power, utility, and data center providers get on board with these locations, any of the 16 could be the future home of AI data centers and on-site power generation. The RFI marks a pivotal step in the U.S. government’s strategy to solidify its global dominance in AI development and energy innovation. By leveraging the vast resources and infrastructure of its national labs and research sites, the DOE is positioning the country to meet the enormous power and security demands of next-generation AI technologies. The selected locations, already home to critical energy research and cutting-edge supercomputing, present a compelling opportunity for industry stakeholders to collaborate on building integrated, sustainable AI data centers with dedicated energy production capabilities. With projects like Oak Ridge’s pioneering SMRs and advanced AI security

Read More »

Generac Sharpens Focus on Data Center Power with Scalable Diesel and Natural Gas Generators

In a digital economy defined by constant uptime and explosive compute demand, power reliability is more than a design criterion—it’s a strategic imperative. In response to such demand, Generac Power Systems, a company long associated with residential backup and industrial emergency power, is making an assertive move into the heart of the digital infrastructure sector with a new portfolio of high-capacity generators engineered for the data center market. Unveiled this week, Generac’s new lineup includes five generators ranging from 2.25 MW to 3.25 MW. These units are available in both diesel and natural gas configurations, and form part of a broader suite of multi-asset energy systems tailored to hyperscale, colocation, enterprise, and edge environments. The product introductions expand Generac’s commercial and industrial capabilities, building on decades of experience with mission-critical power in hospitals, telecom, and manufacturing, now optimized for the scale and complexity of modern data centers. “Coupled with our expertise in designing generators specific to a wide variety of industries and uses, this new line of generators is designed to meet the most rigorous standards for performance, packaging, and after-treatment specific to the data center market,” said Ricardo Navarro, SVP & GM, Global Telecom and Data Centers, Generac. Engineering for the Demands of Digital Infrastructure Each of the five new generators is designed for seamless integration into complex energy ecosystems. Generac is emphasizing modularity, emissions compliance, and high-ambient operability as central to the offering, reflecting a deep understanding of the real-world challenges facing data center operators today. The systems are built around the Baudouin M55 engine platform, which is engineered for fast transient response and high operating temperatures—key for data center loads that swing sharply under AI and cloud workloads. The M55’s high-pressure common rail fuel system supports low NOx emissions and Tier 4 readiness, aligning with the most

Read More »

CoolIT and Accelsius Push Data Center Liquid Cooling Limits Amid Soaring Rack Densities

The CHx1500’s construction reflects CoolIT’s 24 years of DLC experience, using stainless-steel piping and high-grade wetted materials to meet the rigors of enterprise and hyperscale data centers. It’s also designed to scale: not just for today’s most power-hungry processors, but for future platforms expected to surpass today’s limits. Now available for global orders, CoolIT is offering full lifecycle support in over 75 countries, including system design, installation, CDU-to-server certification, and maintenance services—critical ingredients as liquid cooling shifts from high-performance niche to a requirement for AI infrastructure at scale. Capex Follows Thermals: Dell’Oro Forecast Signals Surge In Cooling and Rack Power Infrastructure Between Accelsius and CoolIT, the message is clear: direct liquid cooling is stepping into its maturity phase, with products engineered not just for performance, but for mass deployment. Still, technology alone doesn’t determine the pace of adoption. The surge in thermal innovation from Accelsius and CoolIT isn’t happening in a vacuum. As the capital demands of AI infrastructure rise, the industry is turning a sharper eye toward how data center operators account for, prioritize, and report their AI-driven investments. To wit: According to new market data from Dell’Oro Group, the transition toward high-power, high-density AI racks is now translating into long-term investment shifts across the data center physical layer. Dell’Oro has raised its forecast for the Data Center Physical Infrastructure (DCPI) market, predicting a 14% CAGR through 2029, with total revenue reaching $61 billion. That revision stems from stronger-than-expected 2024 results, particularly in the adoption of accelerated computing by both Tier 1 and Tier 2 cloud service providers. The research firm cited three catalysts for the upward adjustment: Accelerated server shipments outpaced expectations. Demand for high-power infrastructure is spreading to smaller hyperscalers and regional clouds. Governments and Tier 1 telecoms are joining the buildout effort, reinforcing AI as a

Read More »

Podcast: Nomads at the Frontier – AI, Infrastructure, and Data Center Workforce Evolution at DCD Connect New York

The 25th anniversary of the latest Data Center Dynamics event in New York City last month (DCD Connect NY 2025) brought record-breaking attendance, underscoring the accelerating pace of change in the digital infrastructure sector. At the heart of the discussions were evolving AI workloads, power and cooling challenges, and the crucial role of workforce development. Welcoming Data Center Frontier at their show booth were Phill Lawson-Shanks of Aligned Data Centers and Phillip Koblence of NYI, who are respectively managing director and co-founder of the Nomad Futurist Foundation. Our conversation spanned the pressing issues shaping the industry, from the feasibility of AI factories to the importance of community-driven talent pipelines. AI Factories: Power, Cooling, and the Road Ahead One of the hottest topics in the industry is how to support the staggering energy demands of AI workloads. Reflecting on NVIDIA’s latest announcements at GTC, including the potential of a 600-kilowatt rack, Lawson-Shanks described the challenges of accommodating such density. While 120-130 kW racks are manageable today, scaling beyond 300 kW will require rethinking power distribution methods—perhaps moving power sleds outside of cabinets or shifting to medium-voltage delivery. Cooling is another major concern. Beyond direct-to-chip liquid cooling, air cooling still plays a role, particularly for DIMMs, NICs, and interconnects. However, advances in photonics, such as shared laser fiber interconnects, could reduce switch power consumption, marking a potential turning point in energy efficiency. “From our perspective, AI factories are highly conceivable,” said Lawson-Shanks. “But we’re going to see hybridization for a while—clients will want to run cloud infrastructure alongside inference workloads. The market needs flexibility.” Connectivity and the Role of Tier-1 Cities Koblence emphasized the continuing relevance of major connectivity hubs like New York City in an AI-driven world. While some speculate that dense urban markets may struggle to accommodate hyperscale AI workloads,

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »