Stay Ahead, Stay ONMINE

Mistral AI drops new open-source model that outperforms GPT-4o Mini with fraction of parameters

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More French artificial-intelligence startup Mistral AI unveiled a new open-source model today that the company says outperforms similar offerings from Google and OpenAI, setting the stage for increased competition in a market dominated by U.S. tech giants. […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


French artificial-intelligence startup Mistral AI unveiled a new open-source model today that the company says outperforms similar offerings from Google and OpenAI, setting the stage for increased competition in a market dominated by U.S. tech giants.

The model, called Mistral Small 3.1, processes both text and images with just 24 billion parameters—a fraction of the size of leading proprietary models—while matching or exceeding their performance, according to the company.

“This new model comes with improved text performance, multimodal understanding, and an expanded context window of up to 128k tokens,” Mistral said in a company blog post announcing the release. The firm claims the model processes information at speeds of 150 tokens per second, making it suitable for applications requiring rapid response times.

By releasing the model under the permissive Apache 2.0 license, Mistral is pursuing a markedly different strategy than its larger competitors, which have increasingly restricted access to their most powerful AI systems. The approach highlights a growing divide in the AI industry between closed, proprietary systems and open, accessible alternatives.

How a $6 billion European startup is taking on Silicon Valley’s AI giants

Founded in 2023 by former researchers from Google DeepMind and Meta, Mistral AI has rapidly established itself as Europe’s leading AI startup, with a valuation of approximately $6 billion after raising around $1.04 billion in capital. This valuation, while impressive for a European startup, remains a fraction of OpenAI’s reported $80 billion or the resources available to tech giants like Google and Microsoft.

Mistral has achieved notable traction, particularly in its home region. Its chat assistant Le Chat recently reached one million downloads in just two weeks following its mobile release, bolstered by vocal support from French President Emmanuel Macron, who urged citizens to “download Le Chat, which is made by Mistral, rather than ChatGPT by OpenAI — or something else” during a television interview.

The company strategically positions itself as “the world’s greenest and leading independent AI lab,” emphasizing European digital sovereignty as a key differentiator from American competitors.

Small but mighty: How Mistral’s 24 billion parameter model punches above its weight class

Mistral Small 3.1 stands out for its remarkable efficiency. With just 24 billion parameters—a fraction of models like GPT-4—the system delivers multimodal capabilities, multilingual support, and handles long-context windows of up to 128,000 tokens.

This efficiency represents a significant technical achievement. While the AI industry has generally pursued ever-larger models requiring massive computational resources, Mistral has focused on algorithmic improvements and training optimizations to extract maximum capability from smaller architectures.

The approach addresses one of the most pressing challenges in AI deployment: the enormous computational and energy costs associated with state-of-the-art systems. By creating models that run on relatively modest hardware—including a single RTX 4090 graphics card or a Mac with 32GB of RAM—Mistral makes advanced AI accessible for on-device applications where larger models prove impractical.

This emphasis on efficiency may ultimately prove more sustainable than the brute-force scaling pursued by larger competitors. As climate concerns and energy costs increasingly constrain AI deployment, Mistral’s lightweight approach could transition from alternative to industry standard.

Why Europe’s AI champion could benefit from growing geopolitical tensions

Mistral’s latest release emerges amid growing concerns about Europe’s ability to compete in the global AI race, traditionally dominated by American and Chinese companies.

“Not being American or Chinese may now be a help, not a hindrance,” The Economist reported in a recent analysis of Mistral’s position, suggesting that as geopolitical tensions rise, a European alternative may become increasingly attractive for certain markets and governments.

Arthur Mensch, Mistral’s CEO, has advocated forcefully for European digital sovereignty. At the Mobile World Congress in Barcelona this month, he urged European telecoms to “get into the hyperscaler game” by investing in data center infrastructure.

“We would welcome more domestic effort in making more data centers,” Mensch said, suggesting that “the AI revolution is also bringing opportunities to decentralize the cloud.”

The company’s European identity provides significant regulatory advantages. As the EU’s AI Act takes effect, Mistral enters the market with systems designed from inception to align with European values and regulatory expectations. This contrasts sharply with American and Chinese competitors who must retrofit their technologies and business practices to comply with an increasingly complex global regulatory landscape.

Beyond text: Mistral’s expanding portfolio of specialized AI models

Mistral Small 3.1 joins a rapidly expanding suite of AI products from the company. In February, Mistral released Saba, a model focused specifically on Arabic language and culture, demonstrating an understanding that AI development has concentrated excessively on Western languages and contexts.

Earlier this month, the company introduced Mistral OCR, an optical character recognition API that converts PDF documents into AI-ready Markdown files—addressing a critical need for enterprises seeking to make document repositories accessible to AI systems.

These specialized tools complement Mistral’s broader portfolio, which includes Mistral Large 2 (their flagship large language model), Pixtral (for multimodal applications), Codestral (for code generation), and “Les Ministraux,” a family of models optimized for edge devices.

This diversified portfolio reveals a sophisticated product strategy that balances innovation with market demands. Rather than pursuing a single monolithic model, Mistral creates purpose-built systems for specific contexts and requirements — an approach that may prove more adaptable to the rapidly evolving AI landscape.

From Microsoft to military: How strategic partnerships are fueling Mistral’s growth

Mistral’s rise has accelerated through strategic partnerships, including a deal with Microsoft that includes distribution of its AI models through Microsoft’s Azure platform and a $16.3 million investment.

The company has also secured partnerships with France’s army and job agency, German defense tech startup Helsing, IBM, Orange, and Stellantis, positioning itself as a key player in Europe’s AI ecosystem.

In January, Mistral signed a deal with press agency Agence France-Presse (AFP) to allow its chat assistant to query AFP’s entire text archive dating back to 1983, enriching its knowledge base with high-quality journalistic content.

These partnerships reveal a pragmatic approach to growth. Despite positioning itself as an alternative to American tech giants, Mistral recognizes the necessity of working within existing technological ecosystems while building the foundation for greater independence.

The open source advantage: Why Mistral is betting against big tech’s closed AI systems

Mistral’s continued commitment to open source represents its most distinctive strategic choice in an industry increasingly dominated by closed, proprietary systems.

While Mistral maintains some premier models for commercial purposes, its strategy of releasing powerful models like Mistral Small 3.1 under permissive licenses challenges conventional wisdom about intellectual property in AI development.

This approach has already produced tangible benefits. The company noted that “several excellent reasoning models” have been built on top of its previous Mistral Small 3, such as DeepHermes 24B by Nous Research—evidence that open collaboration can accelerate innovation beyond what any single organization might achieve independently.

The open-source strategy also serves as a force multiplier for a company with limited resources compared to its competitors. By enabling a global community of developers to build upon and extend its models, Mistral effectively expands its research and development capacity far beyond its direct headcount.

This approach represents a fundamentally different vision for AI’s future — one where foundational technologies function more like digital infrastructure than proprietary products. As large language models become increasingly commoditized, the true value may shift to specialized applications, industry-specific implementations, and service delivery rather than the base models themselves.

The strategy carries significant risks. If core AI capabilities become widely available commodities, Mistral will need to develop compelling differentiation in other areas. Yet it also protects the company from becoming trapped in an escalating arms race with vastly better-funded competitors — a contest few European startups could hope to win through conventional means.

By positioning itself at the center of an open ecosystem rather than attempting to control it entirely, Mistral may ultimately build something more resilient than what any single organization could create alone.

The $6 billion question: Can Mistral’s business model support its ambitious vision?

Mistral faces significant challenges despite its technical achievements and strategic vision. The company’s revenue reportedly remains in the “eight-digit range,” according to multiple sources—a fraction of what might be expected for its nearly $6 billion valuation.

Mensch has ruled out selling the company, stating at the World Economic Forum in Davos that Mistral is “not for sale” and that “of course, [an IPO is] the plan.” However, the path to sufficient revenue growth remains unclear in an industry where deep-pocketed competitors can afford to operate at a loss for extended periods.

The company’s open-source strategy, while innovative, introduces its own challenges. If base models become commoditized as Lample predicts, Mistral must develop additional revenue streams through specialized services, enterprise deployments, or unique applications that leverage but extend beyond their foundational technologies.

Mistral’s European identity, while providing regulatory advantages and appeal to sovereignty-conscious customers, also potentially limits its immediate growth potential compared to American and Chinese markets where AI adoption typically moves faster.

Nevertheless, Mistral Small 3.1 represents a compelling technical achievement and strategic statement. By demonstrating that advanced AI capabilities can be delivered in smaller, more efficient packages under open licenses, Mistral challenges fundamental assumptions about how AI development and commercialization should proceed.

For a technology industry increasingly concerned about concentration of power among a handful of American tech giants, Mistral’s European-led, open-source alternative offers a vision of a more distributed, accessible AI future—provided it can build a sustainable business model to support its ambitious technical agenda.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

IonQ, Alice & Bob roll out quantum breakthroughs

“And therefore, does bring us closer to escape velocity,” he added. “When this will happen, and we’ll be able to present an Alice & Bob’s logical qubit under threshold depends on a variety of factors, but we can openly say that this is the current big work happening in the

Read More »

Bad data in, bad data out: Protecting your investment in ADMS

Congratulations! Your utility has successfully implemented a cutting-edge ADMS application. Your GIS team has spent months working closely with the implementation team to clean and correct the data within the GIS application. The teams have validated voltage and phasing, eliminated loops, resolved open points, populated missing attribution with default values,

Read More »

But first, eat your vegetables: A guide to smarter load management

For more than a decade, the phrase “first, eat your energy efficiency vegetables” has been used to encourage consumers and utilities to prioritize reducing energy use before making changes to how energy is sourced. This principle remains true today, but with a new twist: electrification. As utilities navigate the evolving energy landscape, ensuring that efficiency and electrification work hand-in-hand is critical to reducing emissions, optimizing grid performance, and keeping customer bills in check. The Foundation Energy efficiency (EE) has long been the go-to strategy for utilities looking to lower energy consumption, reduce peak demand, and help customers save money. Electrification, on the other hand, can increase energy demand—especially if customers adopt electric vehicles (EVs) and heat pumps without implementing efficiency measures first. But when done right, electrification can be a powerful tool for decarbonization while still maintaining the benefits of energy efficiency. Why Smart Load Management Matters Electrification without efficiency is like skipping the vegetables and going straight for dessert—it may taste good in the short term but leads to problems down the line. Increased electricity consumption from EVs, heat pumps, and other electric technologies can put additional stress on the grid. The key is to integrate energy efficiency strategies with electrification efforts to balance load growth and avoid excessive peak demand. Targeting Customers Who Are Ready for the Next Step Not every customer is ready for a full-scale energy retrofit, but many are open to incremental improvements. The good news? Data shows that those who adopt smart thermostats or EVs are often inclined to pursue further energy efficiency improvements, and vice versa. Utilities can leverage this customer propensity by using targeted marketing and strategic program design to enhance energy efficiency and demand response (DR) participation. Here’s how: 1. Promote Heat Pumps to Homes with Recent Insulation Work A home

Read More »

APA Hits Oil Discovery on Alaskan North Slope

APA Corp. and its partners have declared a “significant” petroleum discovery in the Lagniappe area on Alaska’s North Slope. Drilled to about 10,500 feet, the Sockeye-2 well “encountered a high-quality reservoir with approximately 25 feet of net oil pay in one blocky, Paleocene-aged sand with an average porosity of 20 percent”, Houston, Texas-based APA said in an online statement. “As compared to recent regional field analogs in the Brookian play, the porosity and permeability are both better than expected, with the permeability to be confirmed through a planned flow test”, the oil and gas explorer and producer said. “Additional zones of potential pay were also encountered in the shallower Staines Tongue formation. “The Sockeye prospect is amplitude supported across 25,000 to 30,000 acres, and confirms the partners’ geologic and geophysical models, derisking numerous additional prospects in the area. “Wireline logging is complete and additional data collection is underway, including acquiring core and flow testing the well”. APA holds a 50 percent stake in the leasehold through subsidiary Apache Corp. Operator Lagniappe Alaska LLC of Armstrong Oil & Gas Inc. owns 25 percent. Santos Ltd. has the remaining 25 percent. Sockeye-2 is the second successful exploratory well drilled by the consortium on their 325,411-acre position on state lands. The first discovery, King Street-1, struck oil in two Brookian zones. “The Sockeye-2 well further demonstrates the potential of the play, presenting an exciting opportunity in an active area of the North Slope with significant existing infrastructure”, commented Bill Armstrong, chief executive of Armstrong Oil & Gas. The leasehold sits east of Prudhoe Bay, one of the biggest oilfields in the United States. In a separate statement, Santos said, “The exploration well cost is carried by APA as part of a 2023 farm-in agreement”. On a separate upstream development in Alaska, Santos also said phase

Read More »

New York state offers $5M for existing building energy innovations

The New York State Energy Research and Development Authority has allocated $5 million to support building solutions that can be demonstrated in existing buildings, including small- to medium-sized commercial, multifamily and single-family buildings, the department announced Friday. Through the Enabling Innovative Clean Energy Building Solutions program, the funding aims to accelerate the market-readiness of new and emerging clean energy and energy efficiency technologies that can help to manage future energy use, enhance energy efficiency and lower operating costs in buildings across the state, NYSERDA says. “By addressing critical areas like clean heating and cooling, energy efficiency, and building decarbonization, NYSERDA … will not only help the State meet its climate goals but also support long-term resilience and cost savings for tenants and building owners across New York State,” Richard Yancey, CEO of the Building Energy Exchange, said in a statement.  The program seeks applications for projects and programs that demonstrate the “latest innovations in integrated systems, materials and technologies that can greatly improve building or home performance, reduce costs, and/or improve resiliency,” NYSERDA said. The maximum funding for each selected proposal is $1 million, with eligible projects being required to have been previously awarded funding from a federal, state or local government entity — besides NYSERDA — foundation or non-profit to develop an innovative clean building technology product or solution, per the release. Eligible funding areas include solutions related to building envelopes, such as components or materials that improve performance or enable building envelope retrofit; clean heating and cooling; intelligent grid-interactive building controls; low-global warming potential refrigerant applications and leak detection; or thermal storage.  Clean heating and cooling solutions include new heat pump equipment, ground source systems and hybrid systems, with grid-interactive building controls including those that enable electric load and energy asset management. Thermal storage solutions that integrate heating,

Read More »

Microgrids called a low-burden way to ensure backup power

Microgrids don’t pose an immediate threat to diesel generators as the way most managers ensure their facilities have backup power, but Allan Schurr, the chief commercial officer at microgrid company Enchanted Rock, expects more operators to look at what he says is a cleaner, administratively easier approach that microgrids offer.  “Diesel generators are in short supply and sometimes it’s two years before you can get a diesel generator delivered,” Schurr told Facilities Dive. “By having a microgrid … you can [have something] more reliable and cleaner” and have it up and running within a year.” A microgrid is a self-contained electrical network that companies like Schurr’s install on their client’s property. Enchanted Rock builds its network using half-megawatt natural gas generators that it gangs together based on the amount of load the client wants to back up. For a hospital, it might be a network of 10 generators. For a grocery store, it might be two or three.  “Our standardization of equipment is part of our differentiator,” said Schurr, referring to his company’s use of a single-size generator. “That allows us to dial in exactly the amount of backup power that’s needed for any facility size.” Power from the microgrid is cleaner, Schurr says. In the case of Enchanted Rock, it’s cleaner because it’s generated from natural gas. But microgrids can also use solar, wind and other non-fossil-fuel burning sources.  The cleaner power might be important to organizations that make sustainability a priority but it can also help lower energy costs by enabling the microgrid to run power outside of emergency situations and sell it as a supplemental power source to the local utility grid. That revenue goes to the microgrid and helps offset what the property owner pays for its backup power.  “So, [our clients] can get the air

Read More »

Industry groups, utilities encourage more ‘efficient’ Part 53 changes

Dive Brief: The U.S. Nuclear Regulatory Commission’s proposed Part 53 rule must be changed to comport with recent directives from Congress and the Executive Branch to “achieve the most effective and efficient regulatory framework possible,” the Nuclear Energy Institute said in February. The industry group’s recommendation was one of dozens submitted by utilities, trade associations, nuclear technology developers and other industry stakeholders since the NRC released the proposed rule last fall. Many recommendations, though not all, urged the NRC to amend the proposed rule to make it less prescriptive or reduce perceived burdens on future licensees. The current version of the “risk-informed, technology-inclusive” Part 53 framework, which would provide an alternative to the existing Part 50 and Part 52 licensing pathways, incorporates changes requested by lawmakers in 2023 and ordered by NRC commissioners in 2024 to improve the new framework’s efficiency. Dive Insight: The revised Part 53 framework eliminates quantitative health objectives, a set of safety metrics that some in the industry feared would restrict future license applicants’ flexibility. In a March 2024 staff memo, the NRC suggested applicants propose “a comprehensive plant risk metric (or set of metrics)” instead. The revised Part 53 also removes a second possible licensing pathway, known as Framework B. Framework B more closely resembled the existing licensing pathways, but with technology-neutral modifications that reduce burdens on non-light water reactor designs.  Though the NRC voted last year to explore whether aspects of Framework B could inform existing licensing frameworks or possibly find life as a new, standalone framework, NEI in its comments last month recommended against the standalone approach, arguing that it was distracting NRC staff from pursuing more significant improvements to Part 53. In its comments, NEI encouraged NRC staff to take a “systematic and aggressive search for potential changes, in requirements, policy and

Read More »

TIPRO, TXOGA Look at Texas Upstream Employment

According to the Texas Independent Producers and Royalty Owners Association’s (TIPRO) analysis, direct Texas upstream employment for January totaled 203,400. That’s what TIPRO said in a statement sent to Rigzone on Friday, which cited the latest Current Employment Statistics (CES) report from the U.S. Bureau of Labor Statistics (BLS). In the statement, TIPRO highlighted that the January figure was “an increase of 2,500 industry positions from December employment numbers, subject to revisions”. TIPRO also noted in the statement that this represented an increase of 1,600 jobs in Oil and Gas Extraction and 900 jobs in the Services sector. In the statement, TIPRO said its new workforce data “indicated strong job postings for the Texas oil and natural gas industry”. It added that there were 10,724 active unique jobs postings for the Texas oil and natural gas industry last month, “including 5,140 new postings”. “In comparison, the state of California had 3,017 unique job postings in January, followed by New York (2,437), Florida (1,936) and Colorado (1,544),” TIPRO stated in the report. “TIPRO reported a total of 54,402 unique job postings nationwide last month within the oil and natural gas sector,” it added. TIPRO also highlighted in the statement that, among the 19 specific industry sectors it uses to define the Texas oil and natural gas industry, Gasoline Stations with Convenience Stores led in the ranking for unique job listings in January with 3,160 postings. This was followed by Support Activities for Oil and Gas Operations, with 2,321 postings, and Petroleum Refineries, with 801 postings, TIPRO outlined in the statement. The leading three cities by total unique oil and natural gas job postings were Houston (2,437), Midland (688) and Odessa (446), TIPRO noted in the statement. The top three companies ranked by unique job postings in January were Cefco (1,651), Love’s (693) and Energy Transfer (337), the organization added. “Of the top ten companies listed by unique job postings last month, four companies were in the

Read More »

Enterprises reevaluate virtualization strategies amid Broadcom uncertainty

This dilemma of whether to absorb the Broadcom price hikes or embark on the arduous and risky journey of untangling from the VMware ecosystem is triggering a broader C-level conversation around virtualization strategy. “For enterprises navigating this uncertainly, the challenge isn’t just finding a replacement for VMware. IT shops of all sizes see Broadcom’s actions as an opportunity to rethink their approach to virtualization, cloud strategy and IT modernization,” says Steve McDowell, chief analyst at NAND Research. Elliot says that server virtualization has been taken for granted for a long time, and the Broadcom-driven wake-up call is forcing organizations to reevaluate their virtualization strategies at the board level. “That kind of strategic conversation hasn’t happened for years. Customers are saying, ‘What can we do as this platform emerges from VMware. How do we think about this relative to our multi-cloud strategy and private cloud and the efficiencies we can gain? Let’s talk about risk reduction. Let’s talk about platform strategy.’ This is an opportunity to identify business value. It’s triggering this plethora of swim lanes.” Check the waters before diving in While there are multiple alternatives to the VMware platform, none of them are as good from a feature perspective, and there’s a risk associated with moving off a tried-and-true platform. In estimating the cost of a large-scale VMware migration, Gartner cautions: “VMware’s server virtualization platform has become the point of integration for its customers across server, storage and network infrastructure in the data center. Equally, it is a focus of IT operational duties including workload provisioning, backup and disaster recovery. Migrating from VMware’s server virtualization platform would require untangling many aspects of these investments.” It would take a midsize enterprise at least two years to untangle much of its dependency upon VMware, and it could take a large enterprise

Read More »

5 alternatives to VMware vSphere virtualization platform

Nutanix – which is actively courting disgruntled VMware customers – provides storage services that aggregate storage in a global pool that enables any VM to access and consume storage resources. Features include compression, deduplication, high-availability and snapshots. Enterprises running high-performance databases often require external storage arrays, and Nutanix has addressed that need by certifying storage with SAP HANA and Oracle RAC. (Read more: Cisco, Nutanix strengthen joint HCI package) 4. Scale Computing Platform Scale provides an all-in-one hardware and software package that includes all software licenses. Software features offered at no additional charge include high-availability clustering, built-in disaster recovery, replication and software-defined storage. Scale also offers a tool to automate migrations off vSphere, a centralized management feature for HCI clusters, and the ability to mix and match dissimilar hardware appliances in a cluster. In addition, all storage is pooled. Last summer, Scale Computing said in a quarterly earnings announcement that sales have taken off, thanks in part to Broadcom’s changes to VMware sales operations.  5. VergeIO VergeIO takes HCI to the next level with something it calls ultraconverged infrastructure (UCI). This means VergeIO can not only virtualize the normal stack of compute, networking and storage, it can also implement multi-tenancy, creating multiple virtual data centers (VDCs). Each VDC has its own compute, network, storage, management and VergeOS assigned to it. Enterprises can manage and use each VDC much like the virtual private clouds offered by the hyperscalers. VergeIO says this model creates greater workload density, which means lower costs, improved availability, and simplified IT.

Read More »

IBM laying foundation for mainframe as ultimate AI server

“It will truly change what customers are able to do with AI,” Stowell said. IBM’s mainframe processors The next generation of processors is expected to continue a long history of generation-to-generation improvements, IBM stated in a new white paper on AI and the mainframe. “They are projected to clock in at 5.5 GHz. and include ten 36 MB level 2 caches. They’ll feature built-in low-latency data processing for accelerated I/O as well as a completely redesigned cache and chip-interconnection infrastructure for more on-chip cache and compute capacity,” IBM wrote.  Today’s mainframes also have extensions and accelerators that integrate with the core systems. These specialized add-ons are designed to enable the adoption of technologies such as Java, cloud and AI by accelerating computing paradigms that are essential for high-volume, low-latency transaction processing, IBM wrote.  “The next crop of AI accelerators are expected to be significantly enhanced—with each accelerator designed to deliver 4 times more compute power, reaching 24 trillion operations per second (TOPS),” IBM wrote. “The I/O and cache improvements will enable even faster processing and analysis of large amounts of data and consolidation of workloads running across multiple servers, for savings in data center space and power costs. And the new accelerators will provide increased capacity to enable additional transaction clock time to perform enhanced in-transaction AI inferencing.” In addition, the next generation of the accelerator architecture is expected to be more efficient for AI tasks. “Unlike standard CPUs, the chip architecture will have a simpler layout, designed to send data directly from one compute engine, and use a range of lower- precision numeric formats. These enhancements are expected to make running AI models more energy efficient and far less memory intensive. As a result, mainframe users can leverage much more complex AI models and perform AI inferencing at a greater scale

Read More »

VergeIO enhances VergeFabric network virtualization offering

VergeIO is not, however, using an off-the-shelf version of KVM. Rather, it is using what Crump referred to as a heavily modified KVM hypervisor base, with significant proprietary enhancements while still maintaining connections to the open-source community. VergeIO’s deployment profile is currently 70% on premises and about 30% via bare-metal service providers, with a particularly strong following among cloud service providers that host applications for their customers. The software requires direct hardware access due to its low-level integration with physical resources. “Since November of 2023, the normal number one customer we’re attracting right now is guys that have had a heart attack when they got their VMware renewal license,” Crump said. “The more of the stack you own, the better our story becomes.” A 2024 report from Data Center Intelligence Group (DCIG) identified VergeOS as one of the top 5 alternatives to VMware. “VergeIO starts by installing VergeOS on bare metal servers,” the report stated. “It then brings the servers’ hardware resources under its management, catalogs these resources, and makes them available to VMs. By directly accessing and managing the server’s hardware resources, it optimizes them in ways other hypervisors often cannot.” Advanced networking features in VergeFabric VergeFabric is the networking component within the VergeOS ecosystem, providing software-defined networking capabilities as an integrated service rather than as a separate virtual machine or application.

Read More »

Podcast: On the Frontier of Modular Edge AI Data Centers with Flexnode’s Andrew Lindsey

The modular data center industry is undergoing a seismic shift in the age of AI, and few are as deeply embedded in this transformation as Andrew Lindsey, Co-Founder and CEO of Flexnode. In a recent episode of the Data Center Frontier Show podcast, Lindsey joined Editor-in-Chief Matt Vincent and Senior Editor David Chernicoff to discuss the evolution of modular data centers, the growing demand for high-density liquid-cooled solutions, and the industry factors driving this momentum. A Background Rooted in Innovation Lindsey’s career has been defined by the intersection of technology and the built environment. Prior to launching Flexnode, he worked at Alpha Corporation, a top 100 engineering and construction management firm founded by his father in 1979. His early career involved spearheading technology adoption within the firm, with a focus on high-security infrastructure for both government and private clients. Recognizing a massive opportunity in the data center space, Lindsey saw a need for an innovative approach to infrastructure deployment. “The construction industry is relatively uninnovative,” he explained, citing a McKinsey study that ranked construction as the second least-digitized industry—just above fishing and wildlife, which remains deliberately undigitized. Given the billions of square feet of data center infrastructure required in a relatively short timeframe, Lindsey set out to streamline and modernize the process. Founded four years ago, Flexnode delivers modular data centers with a fully integrated approach, handling everything from site selection to design, engineering, manufacturing, deployment, operations, and even end-of-life decommissioning. Their core mission is to provide an “easy button” for high-density computing solutions, including cloud and dedicated GPU infrastructure, allowing faster and more efficient deployment of modular data centers. The Rising Momentum for Modular Data Centers As Vincent noted, Data Center Frontier has closely tracked the increasing traction of modular infrastructure. Lindsey has been at the forefront of this

Read More »

Last Energy to Deploy 30 Microreactors in Texas for Data Centers

As the demand for data center power surges in Texas, nuclear startup Last Energy has now announced plans to build 30 microreactors in the state’s Haskell County near the Dallas-Fort Worth Metroplex. The reactors will serve a growing customer base of data center operators in the region looking for reliable, carbon-free energy. The plan marks Last Energy’s largest project to date and a significant step in advancing modular nuclear power as a viable solution for high-density computing infrastructure. Meeting the Looming Power Demands of Texas Data Centers Texas is already home to over 340 data centers, with significant expansion underway. Google is increasing its data center footprint in Dallas, while OpenAI’s Stargate has announced plans for a new facility in Abilene, just an hour south of Last Energy’s planned site. The company notes the Dallas-Fort Worth metro area alone is projected to require an additional 43 gigawatts of power in the coming years, far surpassing current grid capacity. To help remediate, Last Energy has secured a 200+ acre site in Haskell County, approximately three and a half hours west of Dallas. The company has also filed for a grid connection with ERCOT, with plans to deliver power via a mix of private wire and grid transmission. Additionally, Last Energy has begun pre-application engagement with the U.S. Nuclear Regulatory Commission (NRC) for an Early Site Permit, a key step in securing regulatory approval. According to Last Energy CEO Bret Kugelmass, the company’s modular approach is designed to bring nuclear energy online faster than traditional projects. “Nuclear power is the most effective way to meet Texas’ growing energy demand, but it needs to be deployed faster and at scale,” Kugelmass said. “Our microreactors are designed to be plug-and-play, enabling data center operators to bypass the constraints of an overloaded grid.” Scaling Nuclear for

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »