Stay Ahead, Stay ONMINE

Meta’s answer to DeepSeek is here: Llama 4 launches with long context Scout and Maverick models, and 2T parameter Behemoth on the way!

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More The entire AI landscape shifted back in January 2025 after a then little-known Chinese AI startup DeepSeek (a subsidiary of the Hong Kong-based quantitative analysis firm High-Flyer Capital Management) launched its powerful open source language reasoning […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


The entire AI landscape shifted back in January 2025 after a then little-known Chinese AI startup DeepSeek (a subsidiary of the Hong Kong-based quantitative analysis firm High-Flyer Capital Management) launched its powerful open source language reasoning model DeepSeek R1 publicly to the world, besting U.S. giants such as Meta.

As DeepSeek usage spread rapidly among researchers and enterprises, Meta was reportedly sent into panic mode upon learning that this new R1 model had been trained for a fraction of the cost of many other leading models yet outclassed them, reportedly for as little as several million dollars — what it pays some of its own AI team leaders.

Meta’s whole generative AI strategy had until that point been predicated on releasing best-in-class open source models under its brand name “Llama” for researchers and companies to build upon freely (at least, if they had fewer than 700 million monthly users, at which point they are supposed to contact Meta for special paid licensing terms). Yet DeepSeek R1’s astonishingly good performance on a far smaller budget had allegedly shaken the company leadership and forced some kind of reckoning, with the last version of Llama, 3.3, having been released just a month prior in December 2024 yet already looking outdated.

Now we know the fruits of that effort: today, Meta founder and CEO Mark Zuckerberg took to his Instagram account to announced a new Llama 4 series of models, with two of them — the 400-billion parameter Llama 4 Maverick and 109-billion parameter Llama 4 Scout — available today for developers to download and begin using or fine-tuning now on llama.com and AI code sharing community Hugging Face.

A massive 2-trillion parameter Llama 4 Behemoth is also being previewed today, though Meta’s blog post on the releases said it was still being trained, and gave no indication of when it might be released. (Recall parameters refer to the settings that govern the model’s behavior and that generally more mean a more powerful and complex all around model.)

One headline feature of these models is that they are all multimodal — trained on, and therefore, capable of receiving and generating text, video, and imagery (hough audio was not mentioned).

Another is that they have incredibly long context windows — 1 million tokens for Llama 4 Maverick and 10 million for Llama 4 Scout — which is equivalent to about 1,500 and 15,000 pages of text, respectively, all of which the model can handle in a single input/output interaction. That means a user could theoretically upload or paste up to 7,500 pages-worth-of text and receive that much in return from Llama 4 Scout, which would be handy for information-dense fields such as medicine, science, engineering, mathematics, literature etc.

Here’s what else we’ve learned about this release so far:

All-in on mixture-of-experts

All three models use the “mixture-of-experts (MoE)” architecture approach popularized in earlier model releases from OpenAI and Mistral, which essentially combines multiple smaller models specialized (“experts”) in different tasks, subjects and media formats into a unified whole, larger model. Each Llama 4 release is said to be therefore a mixture of 128 different experts, and more efficient to run because only the expert needed for a particular task, plus a “shared” expert, handles each token, instead of the entire model having to run for each one.

As the Llama 4 blog post notes:

As a result, while all parameters are stored in memory, only a subset of the total parameters are activated while serving these models. This improves inference efficiency by lowering model serving costs and latency—Llama 4 Maverick can be run on a single [Nvidia] H100 DGX host for easy deployment, or with distributed inference for maximum efficiency.

Both Scout and Maverick are available to the public for self-hosting, while no hosted API or pricing tiers have been announced for official Meta infrastructure. Instead, Meta focuses on distribution through open download and integration with Meta AI in WhatsApp, Messenger, Instagram, and web.

Meta estimates the inference cost for Llama 4 Maverick at $0.19 to $0.49 per 1 million tokens (using a 3:1 blend of input and output). This makes it substantially cheaper than proprietary models like GPT-4o, which is estimated to cost $4.38 per million tokens, based on community benchmarks.

All three Llama 4 models—especially Maverick and Behemoth—are explicitly designed for reasoning, coding, and step-by-step problem solving — though they don’t appear to exhibit the chains-of-thought of dedicated reasoning models such as the OpenAI “o” series, nor DeepSeek R1.

Instead, they seem designed to compete more directly with “classical,” non-reasoning LLMs and multimodal models such as OpenAI’s GPT-4o and DeepSeek’s V3 — with the exception of Llama 4 Behemoth, which does appear to threaten DeepSeek R1 (more on this below!)

In addition, for Llama 4, Meta built custom post-training pipelines focused on enhancing reasoning, such as:

  • Removing over 50% of “easy” prompts during supervised fine-tuning.
  • Adopting a continuous reinforcement learning loop with progressively harder prompts.
  • Using pass@k evaluation and curriculum sampling to strengthen performance in math, logic, and coding.
  • Implementing MetaP, a new technique that lets engineers tune hyperparameters (like per-layer learning rates) on models and apply them to other model sizes and types of tokens while preserving the intended model behavior.

MetaP is of particular interest as it could be used going forward to set hyperparameters on on model and then get many other types of models out of it, increasing training efficiency.

As my VentureBeat colleague and LLM expert Ben Dickson opined ont the new MetaP technique: “This can save a lot of time and money. It means that they run experiments on the smaller models instead of doing them on the large-scale ones.”

This is especially critical when training models as large as Behemoth, which uses 32K GPUs and FP8 precision, achieving 390 TFLOPs/GPU over more than 30 trillion tokens—more than double the Llama 3 training data.

In other words: the researchers can tell the model broadly how they want it to act, and apply this to larger and smaller version of the model, and across different forms of media.

A powerful – but not yet the most powerful — model family

In his announcement video on Instagram (a Meta subsidiary, naturally), Meta CEO Mark Zuckerberg said that the company’s “goal is to build the world’s leading AI, open source it, and make it universally accessible so that everyone in the world benefits…I’ve said for a while that I think open source AI is going to become the leading models, and with Llama 4, that is starting to happen.”

It’s a clearly carefully worded statement, as is Meta’s blog post calling Llama 4 Scout, “the best multimodal model in the world in its class and is more powerful than all previous generation Llama models,” (emphasis added by me).

In other words, these are very powerful models, near the top of the heap compared to others in their parameter-size class, but not necessarily setting new performance records. Nonetheless, Meta was keen to trumpet the models its new Llama 4 family beats, among them:

Llama 4 Behemoth

  • Outperforms GPT-4.5, Gemini 2.0 Pro, and Claude Sonnet 3.7 on:
    • MATH-500 (95.0)
    • GPQA Diamond (73.7)
    • MMLU Pro (82.2)

Llama 4 Maverick

  • Beats GPT-4o and Gemini 2.0 Flash on most multimodal reasoning benchmarks:
    • ChartQA, DocVQA, MathVista, MMMU
  • Competitive with DeepSeek v3.1 (45.8B params) while using less than half the active parameters (17B)
  • Benchmark scores:
    • ChartQA: 90.0 (vs. GPT-4o’s 85.7)
    • DocVQA: 94.4 (vs. 92.8)
    • MMLU Pro: 80.5
  • Cost-effective: $0.19–$0.49 per 1M tokens

Llama 4 Scout

  • Matches or outperforms models like Mistral 3.1, Gemini 2.0 Flash-Lite, and Gemma 3 on:
    • DocVQA: 94.4
    • MMLU Pro: 74.3
    • MathVista: 70.7
  • Unmatched 10M token context length—ideal for long documents, codebases, or multi-turn analysis
  • Designed for efficient deployment on a single H100 GPU

But after all that, how does Llama 4 stack up to DeepSeek?

But of course, there are a whole other class of reasoning-heavy models such as DeepSeek R1, OpenAI’s “o” series (like GPT-4o), Gemini 2.0, and Claude Sonnet.

Using the highest-parameter model benchmarked—Llama 4 Behemoth—and comparing it to the intial DeepSeek R1 release chart for R1-32B and OpenAI o1 models, here’s how Llama 4 Behemoth stacks up:

BenchmarkLlama 4 BehemothDeepSeek R1OpenAI o1-1217
MATH-50095.097.396.4
GPQA Diamond73.771.575.7
MMLU82.290.891.8

What can we conclude?

  • MATH-500: Llama 4 Behemoth is slightly behind DeepSeek R1 and OpenAI o1.
  • GPQA Diamond: Behemoth is ahead of DeepSeek R1, but behind OpenAI o1.
  • MMLU: Behemoth trails both, but still outperforms Gemini 2.0 Pro and GPT-4.5.

Takeaway: While DeepSeek R1 and OpenAI o1 edge out Behemoth on a couple metrics, Llama 4 Behemoth remains highly competitive and performs at or near the top of the reasoning leaderboard in its class.

Safety and less political ‘bias’

Meta also emphasized model alignment and safety by introducing tools like Llama Guard, Prompt Guard, and CyberSecEval to help developers detect unsafe input/output or adversarial prompts, and implementing Generative Offensive Agent Testing (GOAT) for automated red-teaming.

The company also claims Llama 4 shows substantial improvement on “political bias” and says “specifically, [leading LLMs] historically have leaned left when it comes to debated political and social topics,” that that Llama 4 does better at courting the right wing…in keeping with Zuckerberg’s embrace of Republican U.S. president Donald J. Trump and his party following the 2024 election.

Where Llama 4 stands so far

Meta’s Llama 4 models bring together efficiency, openness, and high-end performance across multimodal and reasoning tasks.

With Scout and Maverick now publicly available and Behemoth previewed as a state-of-the-art teacher model, the Llama ecosystem is positioned to offer a competitive open alternative to top-tier proprietary models from OpenAI, Anthropic, DeepSeek, and Google.

Whether you’re building enterprise-scale assistants, AI research pipelines, or long-context analytical tools, Llama 4 offers flexible, high-performance options with a clear orientation toward reasoning-first design.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Broadcom favors stock buyback over investing in innovation

“They spend a premium on acquiring, but they then [implement] ruthless cost cutting and raising prices to pay that debt down quickly,” Bickley said. That’s been apparent since the company’s acquisition of virtualization software vendor VMware, where licensing prices have leapt upwards — although many customers have little choice but

Read More »

HPE Aruba grows network mgmt. deployment options

Specifically, HPE Aruba Networking Central On-Premises for Government’s new deployment option includes FIPS 140-2 certified server hardware to meet related government security requirements, according to the company. “Organizations are increasingly prioritizing data sovereignty, requiring regional and local presence for mission-critical IT solutions. With these innovations, HPE now uniquely addresses the most pressing

Read More »

Riverbed bolsters AIOps platform with predictive and agentic AI

Shamus McGillicuddy, research director for the network management practice at Enterprise Management Associates (EMA), explains Riverbed’s Data Store approach as an enabling technology to Riverbed’s overall observability platform that references multiple siloed databases and uses AI to intelligently gather only the data that is relevant to a given operation. “For

Read More »

Aramco Places 20-Year Order from Rio Grande LNG Expansion Project

NextDecade Corp. has secured a 20-year, 1.2 million metric tons per annum (MMtpa) offtake from Saudi Arabian Oil Co. (Aramco) for the planned fourth liquefaction train of the under-construction Rio Grande LNG project in Brownsville, Texas. Houston, Texas-based NextDecade is seeking purchase commitments to be able to make an FID (final investment decision) on trains 4 and 5. NextDecade has so far approved 3 trains, which comprise phase 1, out of 8 planned for the project. “The Rio Grande LNG Facility continues to attract outstanding LNG customers, which we believe is a testament to the quality of our project”, NextDecade chair and chief executive Matt Schatzman said in a company statement announcing the Aramco deal. Under the agreement with Aramco, a subsidiary of the state-owned oil giant will receive the LNG free on board at a price indexed to the Henry Hub benchmark. Last month NextDecade said it has contracted Houston-based Baker Hughes Co. for the latter to provide its gas turbine and refrigerant compressor technology for trains 4-8. Baker Hughes has already been contracted to supply refrigerant compressors and gas turbines for the 3-train phase 1. Phase 1, under construction by Bechtel Energy Inc., will have a total nameplate capacity of about 18 MMtpa. “NextDecade is making excellent progress on commercializing Rio Grande LNG Trains 4 and 5”, said a press release by NextDecade and Baker Hughes March 11. “The Company expects to make positive final investment decisions and commence construction on Trains 4 and 5 and related infrastructure at the Rio Grande LNG Facility, subject to, among other things, maintaining requisite governmental approvals, finalizing and entering into EPC [engineering, procurement and construction] contracts, entering into appropriate commercial arrangements, and obtaining adequate financing to construct each train and related infrastructure”, the release said. “NextDecade is developing and beginning the permitting process for

Read More »

Building a ‘digital twin’ of your asset can power your future

As the world shifts towards renewable resources and sustainable practices, energy companies must adapt their operations to remain compliant and competitive. Aize embraces this transformation by offering solutions that improve safety, efficiency and decision-making within the industry.   About partnership content Some Energy Voice online content is funded by outside parties. The revenue from this helps to sustain our independent news gathering. You will always know if you are reading paid-for material as it will be clearly labelled as “Partnership” on the site and on social media channels, This can take two different forms. “Presented by”This means the content has been paid for and produced by the named advertiser. “In partnership with”This means the content has been paid for and approved by the named advertiser but written and edited by our own commercial content team. The energy sector generates vast amounts of data, from sensor readings and equipment performance metrics to operational and financial information. However, most companies struggle to harness this wealth of data effectively. As a cutting-edge technology company, Aize empowers companies with the tools they need to achieve their operational goals. It works to break down data silos and integrate information into a unified platform, enabling companies to analyse performance in real time and make informed decisions. How Aize makes your heavy assets feel lighter Aize accomplishes this by using digital twin technology, creating digital representations of large assets like offshore wind farms or oil refineries. Tom Brian, group product manager at Aize, explained: “In one facility like an oil and gas platform, you can have data coming from five, ten different places. I’ve worked in industry trying to operate in ten different systems. It’s impossible to see the big picture. “What we do is pull all that together and let users access and visualise data in

Read More »

Perenco completes first UK CO2 injection

Perenco has completed the UK’s first carbon dioxide injection at its Poseidon carbon capture storage (CCS) project. The firm kicked off the project’s CO2 injection test in February and has now reported its successful conclusion. Perenco, along with Carbon Catalyst and Harbour Energy conducted the CO2 injection test from the ERDA rig, newly equipped with the injection test facilities. Petrodec’s ERDA, the first rig in the UK to have achieved an approved safety case for CO2 injection support, has now sailed away from the Leman 27H platform. There were a total of 15 injection cycles performed into the depleted Leman gas field, which mobilised 11 CO2 offshore batch refills. The firm said that the test was completed without injection issues and that it collected valuable data throughout. Perenco argued that the test proved that carbon “can be a reality for depleted fields of the Southern North Sea” and that it is possible to widely reuse North Sea production infrastructure to “unlock cost-effective projects”. Armel Simondin, chief executive of Perenco, commented: “This test has met our expectations, both in terms of technical execution and the quality of data gathered. “These insights are instrumental as we move to the next phase of developing the Poseidon Project. “The successful test highlights the role our industry can play in the UK’s decarbonisation efforts. “The Perenco CCS team and our joint venture partners are now fully focused on interpreting the results and converting new insights into an evidence-based development plan for Project Poseidon.” The Poseidon project will use the Leman field to store CO2, which will be delivered from the Bacton Gas Terminal on the East Anglian coast. The project is expected to have a capacity of 935m tonnes of CO2, going from an injection rate of around 10 million tonnes per annum (mtpa) when

Read More »

Former Grampian Police detective Alan Smith on life in industrial investigations

For 30 years Alan Smith, now head of investigations at COMET, served in the police before turning his attention to industrial investigations. Smith sat down with Energy Voice Aberdeen features lead, Ryan Duff, to discuss his career in the police, his shift to the private sector and how his approach to investigations had to change when he left law enforcement. “I knew instinctively that when my police career came to an end that I was going to move into the world of industry investigations,” said Smith. The self-proclaimed “career investigator” joined what was then called Grampian Police where he became a detective. Investigating incidents in the North Sea fell under his remit while in the police, something that set him up for a life in industrial investigations. One incident that Smith investigated while with the police was the capsize of the Bourbon Dolphin off the coast of Shetland in 2007. The vessel was on the Chevron field when it run into trouble that resulted in the deaths of eight people, including captain Oddne Arve Remoy and his 14-year-old son David Remoy. The fact that a 14-year-old was onboard the vessel shocked local police at the time, however, it was a Norwegian custom for the captain to bring their son onto their vessel. This lesson has stayed with Smith as he investigates incidents in other parts of the world that have different cultures to that of the UK. Moving from law enforcement to investigating industrial accidents required a change in mindset for Smith, he explained. “A criminal investigation is absolutely all about blame, it’s all about guilt, it’s all about getting the person responsible and bringing them to justice,” Smith explained. “Whereas in industry, it’s the polar opposite… you want to understand what and why the incident happened but you’re not

Read More »

Community and connections are a necessity

Having some time to ourselves can feel like a gift; cherishing those moments for headspace. This time can prove to be energising yet calming, and can help us feel grounded and able to catch our breath in a fast-paced life. Yet, as human beings, social interaction is important for our general wellbeing, even when we may favour solitude. This need became even more apparent during the Covid pandemic, where there was no choice about whether we could spend time with people or not; we had to isolate for long periods of time. © Supplied by Usana MindsetDr Gillian Fowler. The knock-on impact on our psychological wellbeing was huge, and is still felt to this day by many, both in the workplace and at home. ‘The Big Mental Health Report 2024’ has highlighted that one in four people in England will experience mental health issues at some point each year, with these statistics also being echoed in the Scottish Government’s findings. The report also revealed the number of those experiencing loneliness has increased, with nearly 8% of UK adults saying they felt lonely ‘always or often’, with people aged over 65 also sharing these sentiments, as well as experiencing significant anxiety or low moods, with 10% of this population stating this was frequent or all the time. It is also stated that the life expectancy of people with severe mental health illness is about 15-20 years shorter than those without. The shift to open dialogue, mental health policies and campaigns, and more routes to support has helped lessen the stigma of having a mental health illness, but what can we do ourselves to proactively safeguard our health and wellbeing now and going forward? This year’s Mental Health Awareness Week, which runs from 12 to 18 May, has a theme focusing on

Read More »

Simple sustainable energy development which is British

A team of scientists at the UK’s number one ranked university have developed a way of making liquid fuels using the power of the sun to pull carbon dioxide from the air and make synthesis gas. This is basically a mixture of hydrogen and carbon monoxide (CO) in various ratios, and the process is the latest development in so-called direct air capture (DAC). The reactor has been developed by Cambridge University researchers who claim the output of their solar panel-powered reactor could be used as a feedstock to make fuels and the many chemicals and pharmaceutical products we rely on. The researchers say their system is compact and simple enough to be deployed in off-grid locations. Unlike most carbon capture technologies, the DAC reactor developed by the Cambridge team does not require fossil-fuel-based power, or the transport and storage of CO2. It literally can convert atmospheric CO2 into hydrocarbon gas using sunlight and is inspired by one of nature’s greatest processes – photosynthesis. In the UK, carbon capture and storage (CCS) has long been touted as a way of securely dumping below ground the CO2 produced in industrial processes as a partial mitigation of Earth’s accelerating climate crisis. But CCS per se is energy-intensive and there are concerns about the long-term safety of storing pressurised CO2 deep underground long-term. Though there is the CCUS (utilisation) option to consider too. According to the International Energy Agency, around 45 commercial facilities are already in operation applying CCUS to industrial processes, fuel transformation and power generation. CCUS deployment has trailed behind expectations in the past, but momentum has grown substantially in recent years, with over 700 projects in various stages of development across the CCUS value chain. © Supplied by Reisner Lab, CambridProfessor Erwin Reisner, Cambridge University. “Aside from the expense and the

Read More »

China’s rare earth export controls threaten enterprise IT hardware supply chains

“AI-first infrastructure rollouts — particularly those involving GPUs, edge accelerators, and high-efficiency cooling — are directly in the crosshairs,” Gogia noted. “So are quantum computing R&D efforts and high-reliability storage systems where thermal and magnetic materials matter.” China, responsible for 70% of global rare earth mining output and 87% of refined supply, poses a serious threat to enterprise IT hardware supply chains with these restrictions — especially for companies with AI-optimized server lines. AI chip production under threat The impact on semiconductor manufacturing comes at a critical time when enterprise demand for AI chips is soaring. Companies including Nvidia, AMD, Intel, and TSMC rely on rare earth elements during the manufacturing of advanced chips. “We see the greatest exposure in private data center expansion projects, AI inferencing at the edge, and next-gen device manufacturing, including specialized industrial IoT and robotics,” noted Gogia. Major cloud providers have been aggressively expanding their AI compute capacity, with substantial hardware refreshes planned for late 2025. These plans may now face delays or cost increases as chip manufacturers grapple with supply constraints. Pricing pressures to be felt in 3-6 months The immediate impact is expected to be limited as manufacturers work through existing inventory, but pricing pressure could emerge within 3-6 months, experts feel.

Read More »

DARPA backs multiple quantum paths in benchmarking initiative

Nord Quantique plans to use the money to expand its team, says Julien Camirand Lemyre, the company’s president, CTO and co-founder. That’s an opportunity to accelerate the development of the technology, he says. “By extension, what this will mean for enterprise users is that quantum solutions to real-world business problems will be available sooner, due to that acceleration,” he says. “And so enterprise customers need to also accelerate how they are thinking about adoption because the advantages quantum will provide will be tangible.” Lemyre predicts that useful quantum computers will be available for enterprises before the end of the decade. “In fact, there has been tremendous progress across the entire quantum sector in recent years,” he says. “This means industry needs to begin thinking seriously about how they will integrate quantum computing into their operations over the medium term.” “We’re seeing, with the deployment of programs like the QBI in the US and investments of billions of dollars from  public and private investors globally, an increasing maturity of quantum technologies,” said Paul Terry, CEO at Photonic, which is betting on optically-linked silicon spin qubits.  “Our architecture has been designed from day one to build modular, scalable, fault-tolerant quantum systems able to be deployed in data centers,” he said. He’s not the only one to mention fault-tolerance. DARPA stressed fault-tolerance in its announcement, and its selections point to the importance of error correction for the future of quantum computing. The biggest problem with today’s quantum computers is that the number of errors increases faster than the number of qubits, making them impossible to scale up. Quantum companies are working on a variety of approaches to reduce the error rates low enough that quantum computers can get big enough to actually to real work.

Read More »

Zayo’s Fiber Bet: Scaling Long-Haul and Metro Networks for AI Data Centers

Zayo Group Holdings Inc. has emerged as one of the most aggressive fiber infrastructure players in North America, particularly in the context of AI-driven growth. With a $4 billion investment in AI-related long-haul fiber expansion, Zayo is positioning itself as a critical enabler of the AI and cloud computing boom. The company is aggressively expanding its long-haul fiber network, adding over 5,000 route miles to accommodate the anticipated 2-6X increase in AI-driven data center capacity by 2030. This initiative comes as AI workloads continue to push the limits of existing network infrastructure, particularly in long-haul connectivity. New Fiber Routes The new routes include critical connections between key AI data center hubs, such as Chicago-Columbus, Las Vegas-Reno, Atlanta-Ashburn, and Columbus-Indianapolis, among others. Additionally, Zayo is overbuilding seven existing routes to further enhance network performance, resiliency, and low-latency connectivity. This new development is a follow-on to 15 new long haul routes representing over 5300 route miles of new and expanded capacity deployed over the last five years. These route locations were selected based on expected data center growth, power availability, existing capacity constraints, and specific regional considerations. The AI Data Center Sector: A Significant Driver of Fiber Infrastructure The exponential growth of AI-driven data center demand means that the U.S. faces a potential bandwidth shortage. Zayo’s investments look to ensure that long-haul fiber capacity keeps pace with this growth, allowing AI data centers to efficiently transmit data between key markets. This is especially important as data center development locations are being driven more by land and power availability rather than proximity to market. Emerging AI data center markets get the high speed fiber they need, especially as they are moving away from expensive power regions (e.g., California, Virginia) to lower-cost locations (e.g., Ohio, Nevada, Midwest). Without the high-speed networking capabilities offered by

Read More »

Crusoe Adds 4.5 GW Natural Gas to Fuel AI, Expands Abilene Data Center to 1.2 GW

Crusoe and the Lancium Clean Campus: A New Model for Power-Optimized Compute Crusoe Energy’s 300-megawatt deployment at the Lancium Clean Campus in Abilene is a significant marker of how data center strategies are evolving to integrate more deeply with energy markets. By leveraging demand flexibility, stranded power, and renewable energy, Crusoe is following a path similar to some of the most forward-thinking projects in the data center industry. But it’s also pushing the model further—fusing AI and high-performance computing (HPC) with the next generation of power-responsive infrastructure. Here’s how Crusoe’s strategy compares to some of the industry’s most notable power-driven data center deployments: Google’s Oklahoma Data Center: Proximity to Renewable Growth A close parallel to Crusoe’s energy-centric site selection strategy is Google’s Mayes County data center in Oklahoma. Google sited its facility there to take advantage of abundant wind energy, aligning with the local power grid’s renewable capacity. Similarly, Crusoe is tapping into Texas’s deregulated energy market, optimizing for low-cost renewable power and the ability to flexibly scale compute operations in response to grid conditions. Google has also been an industry leader in time-matching workloads to renewable energy availability, something that Crusoe is enabling in real time through grid-responsive compute orchestration. Sabey Data Centers in Quincy: Low-Cost Power as a Foundation Another instructive comparison is Sabey Data Centers’ Quincy, Washington, campus, which was built around one of the most cost-effective power sources in the U.S.—abundant hydroelectric energy. Sabey’s long-term strategy has been to co-locate power-intensive compute infrastructure near predictable, low-cost energy sources. Crusoe’s project applies a similar logic but adapts it for a variable grid environment. Instead of relying on a fixed low-cost power source like hydro, Crusoe dynamically adjusts to real-time energy availability, a strategy that could become a model for future power-aware, AI-driven workloads. Compass and Aligned: Modular, Energy-Adaptive

Read More »

Executive Roundtable: Data Center Site Selection and Market Evolution in a Constrained Environment

For the third installment of our Executive Roundtable for the First Quarter of 2025, we asked our panel of seasoned industry experts about how the dynamics of data center site selection have never been more complex—or more critical to long-term success. In an industry where speed to market is paramount, operators must now navigate an increasingly constrained landscape in the age of AI, ultra cloud and hyperscale expansion, marked by fierce competition for land, tightening power availability, and evolving local regulations.  Traditional core markets such as Northern Virginia, Dallas, and Phoenix remain essential, but supply constraints and permitting challenges are prompting developers to rethink their approach. As hyperscalers and colocation providers push the boundaries of site selection strategy, secondary and edge markets are emerging as viable alternatives, driven by favorable energy economics, infrastructure investment, and shifting customer demand.  At the same time, power procurement is now reshaping the equation. With grid limitations and interconnection delays creating uncertainty in major hubs, operators are exploring new solutions, from direct utility partnerships to on-site generation with renewables, natural gas, and burgeoning modular nuclear concepts. The question now is not just where to build but how to ensure long-term operational resilience. As data center demand accelerates, operators face mounting challenges in securing suitable land, reliable power, and regulatory approvals in both established and emerging markets.  And so we asked our distinguished executive panel for the First Quarter of 2025, with grid capacity constraints, zoning complexities, and heightened competition shaping development decisions, how are companies refining their site selection strategies in Q1 2025 to balance speed to market, scalability, and sustainability? And, which North American regions are showing the greatest potential as the next wave of data center expansion takes shape?

Read More »

Podcast: iMasons CEO Santiago Suinaga on the Future of Sustainable AI Data Centers

For this episode of the DCF Show podcast, host Matt Vincent, Editor in Chief of Data Center Frontier, is joined by Santiago Suinaga, CEO of Infrastructure Masons (iMasons), to explore the urgent challenges of scaling data center construction while maintaining sustainability commitments, among other pertinent industry topics. The AI Race and Responsible Construction “Balancing scale and sustainability is key because the AI race is real,” Suinaga emphasizes. “Forecasted capacities have skyrocketed to meet AI demand. Hyperscale end users and data center developers are deploying high volumes to secure capacity in an increasingly constrained global market.” This surge in demand pressures the industry to build faster than ever before. Yet, as Suinaga notes, speed and sustainability must go hand in hand. “The industry must embrace a build fast, build smart mentality. Leveraging digital twin technology, AI-driven design optimization, and circular economy principles is critical.” Sustainability, he argues, should be embedded at every stage of new builds, from integrating low-carbon materials to optimizing energy efficiency from the outset. “We can’t afford to compromise sustainability for speed. Instead, we must integrate renewable energy sources and partner with local governments, utilities, and energy providers to accelerate responsible construction.” A key example of this thinking is peak shaving—using redundant infrastructure and idle capacities to power the grid when data center demand is low. “99.99% of the time, this excess capacity can support local communities, while ensuring the data center retains prioritized energy supply when needed.” Addressing Embodied Carbon and Supply Chain Accountability Decarbonization is a cornerstone of iMasons’ efforts, particularly through the iMasons Climate Accord. Suinaga highlights the importance of tackling embodied carbon—the emissions embedded in data center construction materials and IT hardware. “We need standardized reporting metrics and supplier accountability to drive meaningful change,” he says. “Greater transparency across the supply chain can be

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »