Stay Ahead, Stay ONMINE

Hugging Face just launched a $299 robot that could disrupt the entire robotics industry

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Hugging Face, the $4.5 billion artificial intelligence platform that has become the GitHub of machine learning, announced Tuesday the launch of Reachy Mini, a $299 desktop robot designed to […]

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now


Hugging Face, the $4.5 billion artificial intelligence platform that has become the GitHub of machine learning, announced Tuesday the launch of Reachy Mini, a $299 desktop robot designed to bring AI-powered robotics to millions of developers worldwide. The 11-inch humanoid companion represents the company’s boldest move yet to democratize robotics development and challenge the industry’s traditional closed-source, high-cost model.

The announcement comes as Hugging Face crosses a significant milestone of 10 million AI builders using its platform, with CEO Clément Delangue revealing in an exclusive interview that “more and more of them are building in relation to robotics.” The compact robot, which can sit on any desk next to a laptop, addresses what Delangue calls a fundamental barrier in robotics development: accessibility.

“One of the challenges with robotics is that you know you can’t just build on your laptop. You need to have some sort of robotics partner to help in your building, and most people won’t be able to buy $70,000 robots,” Delangue explained, referring to traditional industrial robotics systems and even newer humanoid robots like Tesla’s Optimus, which is expected to cost $20,000-$30,000.

How a software company is betting big on physical AI robots

Reachy Mini emerges from Hugging Face’s April acquisition of French robotics startup Pollen Robotics, marking the company’s most significant hardware expansion since its founding. The robot represents the first consumer product to integrate natively with the Hugging Face Hub, allowing developers to access thousands of pre-built AI models and share robotics applications through the platform’s “Spaces” feature.

The timing appears deliberate as the AI industry grapples with the next frontier: physical AI. While large language models have dominated the past two years, industry leaders increasingly believe that artificial intelligence will need physical embodiment to achieve human-level capabilities. Goldman Sachs projects the humanoid robotics market could reach $38 billion by 2035, while the World Economic Forum identifies robotics as a critical frontier technology for industrial operations.

“We’re seeing more and more people moving to robotics, which is extremely exciting,” Delangue said. “The idea is to really become the desktop, open-source robot for AI builders.”

Inside the $299 robot that could democratize AI development

Reachy Mini packs sophisticated capabilities into its compact form factor. The robot features six degrees of freedom in its moving head, full body rotation, animated antennas, a wide-angle camera, multiple microphones, and a 5-watt speaker. The wireless version includes a Raspberry Pi 5 computer and battery, making it fully autonomous.

The robot ships as a DIY kit and can be programmed in Python, with JavaScript and Scratch support planned. Pre-installed demonstration applications include face and hand tracking, smart companion features, and dancing moves. Developers can create and share new applications through Hugging Face’s Spaces platform, potentially creating what Delangue envisions as “thousands, tens of thousands, millions of apps.”

This approach contrasts sharply with traditional robotics companies that typically release one product annually with limited customization options. “We want to have a model where we release tons of things,” Delangue explained. “Maybe we’ll release 100 prototypes a year. Out of this 100 prototypes, maybe we’ll assemble only 10 ourselves… and maybe fully assembled, fully packaged, fully integrated with all the software stack, maybe there’s going to be just a couple of them.”

Why open source hardware might be the future of robotics

The launch represents a fascinating test of whether open-source principles can translate successfully to hardware businesses. Hugging Face plans to release all hardware designs, software, and assembly instructions as open source, allowing anyone to build their own version. The company monetizes through convenience, selling pre-assembled units to developers who prefer to pay rather than build from scratch.

“You try to share as much as possible to really empower the community,” Delangue explained. “There are people who, even if they have all the recipes open source to build their own Reachy Mini, would prefer to pay 300 bucks, 500 bucks, and get it already ready, or easy to assemble at home.”

This freemium approach for hardware echoes successful software models but faces unique challenges. Manufacturing costs, supply chain complexity, and physical distribution create constraints that don’t exist in pure software businesses. However, Delangue argues this creates valuable feedback loops: “You learn from the open source community about what they want to build, how they want to build, and you can reintegrate it into what you sell.”

The privacy challenge facing AI robots in your home

The move into robotics raises new questions about data privacy and security that don’t exist with purely digital AI systems. Robots equipped with cameras, microphones, and the ability to take physical actions in homes and workplaces create unprecedented privacy considerations.

Delangue positions open source as the solution to these concerns. “One of my personal motivations to do open source robotics is that I think it’s going to fight concentration of power… the natural tendency of creating black box robots that users don’t really understand or really control,” he said. “The idea of ending up in a world where just a few companies are controlling millions of robots that are in people’s homes, being able to take action in real life, is quite scary.”

The open-source approach allows users to inspect code, understand data flows, and potentially run AI models locally rather than relying on cloud services. For enterprise customers, Hugging Face’s existing enterprise platform could provide private deployment options for robotics applications.

From prototype to production: Hugging Face’s manufacturing gamble

Hugging Face faces significant manufacturing and scaling challenges as it transitions from a software platform to a hardware company. The company plans to begin shipping Reachy Mini units as early as next month, starting with more DIY-oriented versions where customers complete final assembly.

“The first versions, the first orders shipping will be a bit DIY, in the sense that we’ll split the weight of assembling with the user,” Delangue explained. “We’ll do some of the assembling ourselves, and then the user will be doing some of the assembling themselves too.”

This approach aligns with the company’s goal of engaging the AI builder community in hands-on robotics development while managing manufacturing complexity. The strategy also reflects uncertainty about market demand for the new product category.

Taking on Tesla and Boston Dynamics with radical transparency

Reachy Mini enters a rapidly evolving robotics landscape. Tesla’s Optimus program, Figure’s humanoid robots, and Boston Dynamics‘ commercial offerings represent the high-end of the market, while companies like Unitree have introduced more affordable humanoid robots at around $16,000.

Hugging Face’s approach differs fundamentally from these competitors. Rather than creating a single, highly capable robot, the company is building an ecosystem of affordable, modular, open-source robotics components. Previous releases include the SO-101 robotic arm (starting at $100) and plans for the HopeJR humanoid robot (around $3,000).

The strategy reflects broader trends in AI development, where open-source models from companies like Meta and smaller players have challenged closed-source leaders like OpenAI. In January, Chinese startup DeepSeek shocked the industry by releasing a powerful AI model developed at significantly lower cost than competing systems, demonstrating the potential for open-source approaches to disrupt established players.

Building an ecosystem: The partnerships powering open robotics

Hugging Face’s robotics expansion benefits from strategic partnerships across the industry. The company collaborates with NVIDIA on robotics simulation and training through Isaac Lab, enabling developers to generate synthetic training data and test robot behaviors in virtual environments before deployment.

The recent release of SmolVLA, a 450-million parameter vision-language-action model, demonstrates the technical foundation underlying Reachy Mini. The model is designed to be efficient enough to run on consumer hardware, including MacBooks, making sophisticated AI capabilities accessible to individual developers rather than requiring expensive cloud infrastructure.

Physical Intelligence, a startup co-founded by UC Berkeley professor Sergey Levine, has made its Pi0 robot foundation model available through Hugging Face, creating opportunities for cross-pollination between different robotics approaches. “Making robotics more accessible increases the velocity with which technology advances,” Levine noted in previous statements about open-source robotics.

What a $299 robot means for the billion-dollar AI hardware race

The Reachy Mini launch signals Hugging Face’s ambition to become the dominant platform for AI development across all modalities, not just text and image generation. With robotics representing a potential $38 billion market by 2035, according to Goldman Sachs projections, early platform positioning could prove strategically valuable.

Delangue envisions a future where hardware becomes an integral part of AI development workflows. “We see hardware as part of the AI builder building blocks,” he explained. “Always with our approach of being open, being community driven, integrating everything with as many community members, as many other organizations as possible.”

The company’s financial position provides flexibility to experiment with hardware business models. As a profitable company with significant funding, Hugging Face can afford to prioritize market development over immediate revenue optimization. Delangue mentioned potential subscription models where Hugging Face platform access could include hardware components, similar to how some software companies bundle services.

How affordable robots could transform education and research

Beyond commercial applications, Reachy Mini could significantly impact robotics education and research. At $299, the robot costs less than many smartphones while providing full programmability and AI integration. Universities, coding bootcamps, and individual learners could use the platform to explore robotics concepts without requiring expensive laboratory equipment.

The open-source nature enables educational institutions to modify hardware and software to suit specific curricula. Students could progress from basic programming exercises to sophisticated AI applications using the same platform, potentially accelerating robotics education and workforce development.

Delangue revealed that community feedback has already influenced product development. A colleague’s five-year-old daughter wanted to carry the robot around the house, leading to the development of the wireless version. “She started to want to take the Reachy Mini and bring it everywhere. That’s when the wires started to be a problem,” he explained.

The disruption that could reshape the entire robotics industry

Hugging Face’s approach could fundamentally alter robotics industry dynamics. Traditional robotics companies invest heavily in proprietary technology, limiting innovation to internal teams. The open-source model could unlock distributed innovation across thousands of developers, potentially accelerating advancement while reducing costs.

The strategy mirrors successful disruptions in other technology sectors. Linux challenged proprietary operating systems, Android democratized mobile development, and TensorFlow accelerated machine learning adoption. If successful, Hugging Face’s robotics platform could follow a similar trajectory.

However, hardware presents unique challenges compared to software. Manufacturing quality control, supply chain management, and physical safety requirements create complexity that doesn’t exist in purely digital products. The company’s ability to manage these challenges while maintaining its open-source philosophy will determine the platform’s long-term success.

Whether Reachy Mini succeeds or fails, its launch marks a pivotal moment in robotics development. For the first time, a major AI platform is betting that the future of robotics belongs not in corporate research labs, but in the hands of millions of individual developers armed with affordable, open-source tools. In an industry long dominated by secrecy and six-figure price tags, that might just be the most revolutionary idea of all.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

IBM Power11 challenges x86 and GPU giants with security-first server strategy

The IBM Power Cyber Vault solution is designed to provide protection against cyberattacks such as data corruption and encryption with proactive immutable snapshots that are automatically captured, stored, and tested on a custom-defined schedule, IBM said. Power11 also uses NIST-approved built-in quantum-safe cryptography designed to help protect systems from harvest-now, decrypt-later attacks

Read More »

Strathcona Closes Montney Assets Sale, Becomes Pure-Play Heavy Oil Firm

Strathcona Resources Ltd. has closed its Montney asset sale in Western Canada, consisting of Groundbirch, Kakwa, and Grande Prairie assets, for a total value of approximately $2.09 billion (CAD 2.86 billion), including closing adjustments. Strathcona said it is now a pure-play heavy oil company producing approximately 120 million barrels per day (MMbpd) with a 50-year 2P reserve life index. The company has plans to grow to 195 MMbpd by 2031, according to a news release. After deducting all debt, Strathcona said it currently has approximately CAD 200 million in positive net cash and marketable securities, inclusive of approximately 4.6 million shares in Tourmaline Oil Corp. and 23.4 million shares in MEG Energy Corp. ARC Resources Ltd. acquired the condensate-rich Montney assets in Alberta’s Kakwa in an all-cash transaction valued at approximately CAD 1.6 billion. The assets are “underpinned by a substantial drilling inventory and include owned and operated infrastructure, reinforcing ARC’s position as Canada’s largest Montney and condensate producer,” the company said in a separate statement. For the remainder of 2025, the assets are expected to deliver average production between 35,000 and 40,000 barrels of oil equivalent per day (boepd), consisting of approximately 50 percent crude oil and liquids and 50 percent natural gas, ARC said. The acquisition increases ARC’s Kakwa production 24 percent to over 210,000 boepd and increases the Montney inventory duration at Kakwa from 12 years to over 15 years, according to the statement. The transaction includes 100 percent ownership of two natural gas processing facilities and condensate handling infrastructure. In addition, the assets include a 19 percent interest in a third-party natural gas processing facility with deep-cut natural gas liquids (NGL) recovery, ARC said. ARC said its development plans for the assets will target the Montney, which is approximately 100 percent working interest land. ARC will

Read More »

Delfin Midstream Moves Closer to FID for LNG Export Project in Louisiana

Delfin Midstream Inc. is moving closer to a positive final investment decision (FID) for its deepwater port project in Louisiana with the signing of two agreements. Delfin entered into an agreement with Siemens Energy Inc. to reserve manufacturing capacity for four SGT-750 gas turbine mechanical drive packages, the company said in a news release. The equipment will be used to drive the mixed-refrigerant compressors for Delfin’s liquefied natural gas (LNG) liquefaction system. Delfin said it also agreed to an Early Works program with Samsung Heavy Industries and Black & Veatch Inc. to further detail floating liquefied natural gas (FLNG) vessel design specifications as the basis for the lump-sum, turn-key engineering, procurement, construction, and integration (EPCI) contract and to prepare both contractors for the execution of the project. Financial terms of the contracts were not disclosed. The work will de-risk project execution and ensure both contractors are prepared for immediate project execution following a positive FID, expected in the fall of 2025, Delfin said. Delfin CEO Dudley Poston said, “This is an incredibly exciting time for the development of Delfin’s critical energy infrastructure project. Following the successful issuance of the deepwater port license by MARAD, all workstreams are on schedule and the project is currently on track for FID in the fall of 2025. By making this large investment to lock-in critical manufacturing capacity, we have secured our execution schedule with the anticipated delivery of our first FLNG vessel from Samsung Heavy Industries shipyard in 2029”. Siemens Energy executive board member Karim Amin said, “Siemens Energy is excited to support Delfin’s energy infrastructure project by providing the critical Gas Turbine Mechanical Drive packages the Company needs as it moves towards delivering the first offshore LNG project in the United States. The modular design, high power-to-weight ratio and ability to operate under

Read More »

Executives Reveal Where They See Henry Hub Price Landing in Future

Executives from oil and gas firms have revealed where they expect the Henry Hub natural gas price to be at various points in the future in the second quarter Dallas Fed Energy Survey, which was released recently. The survey asked participants what they expect Henry Hub natural gas prices to be in six months, one year, two years, and five years. Executives from 116 oil and gas firms answered this question and gave a mean response of $3.66 per million British thermal units (MMBtu) for the six month mark, $3.81 per MMBtu for the one year mark, $4.12 per MMBtu for the two year mark, and $4.50 per MMBtu for the five year mark, the survey showed. Executives from 117 oil and gas firms answered this question in the first quarter Dallas Fed Energy Survey and gave a mean response of $3.71 per MMBtu for the six month mark, $3.98 per MMBtu for the year mark, $4.30 per MMBtu for the two year mark, and $4.83 per MMBtu for the five year mark, that survey showed. The second quarter Dallas Fed Energy Survey also asked participants what they expect the Henry Hub price to be at the end of this year. Executives from 133 oil and gas firms answered this question and gave an average response of $3.66 per MMBtu, the survey highlighted. The low forecast was $1.75 per MMBtu, the high forecast was $5 per MMBtu, and the average Henry Hub natural gas daily spot price during the survey was $3.30 per MMBtu, the survey pointed out. Executives from 127 oil and gas firms answered this question in the first quarter Dallas Fed Energy Survey and gave an average response of $3.78 per MMBtu, that survey showed. The low forecast came in at $2 per MMBtu, the high forecast

Read More »

Trican to Acquire Tubing Services Firm Iron Horse

Calgary, Alberta-based Trican Well Service Ltd. said it has entered into an agreement to acquire privately owned fracturing and coiled tubing services provider Iron Horse Energy Services, which operates primarily in the Cardium, Charlie Lake, Mannville Stack, Viking, Montney and Shaunavon plays in the Western Canadian Sedimentary Basin (WCSB). Trican aims to acquire all of the issued and outstanding shares of Iron Horse for approximately $56.6 million (CAD 77.35 million) in cash and approximately 33.76 million common shares of Trican, the company said in a news release. Iron Horse “extends Trican’s fracturing footprint and adds industry-leading coiled tubing integrated fracturing expertise,” according to the release. The acquisition will add over four fracturing spreads and 10 coiled tubing units, which will augment Trican’s services offering throughout the WCSB across the drilling, completion, and production lifecycles, the company said. Following the acquisition, Iron Horse will operate as a wholly owned division of Trican, continuing to serve its existing customers while increasing its footprint with the support of Trican’s resources. Trican said it expects to retain all of the existing management and employees of Iron Horse. The acquisition is expected to close in the second half. Iron Horse Chairman and CEO Tom Coolen will be appointed to the board of directors of Trican after closing, according to the release. Other than Competition Act Approval, and TSX listing approval of the common shares of Trican to be issued pursuant to the Acquisition, no approval, order, consent of or filing with any government agency is required on the part of Iron Horse or Trican, in connection with the completion of the acquisition, Trican said. “Iron Horse is one of [the] few North American fracturing companies that has consistently demonstrated operational and financial performance that aligns with Trican. The acquisition will provide significant EBITDA, free cash

Read More »

Macquarie Strategists Forecast USA Crude Inventory Rise

In an oil and gas report sent to Rigzone by the Macquarie team on Tuesday, Macquarie strategists revealed that they are forecasting that U.S. crude inventories will be up by 2.7 million barrels for the week ending July 4. “This follows a 3.8 million barrel build in the prior week, with the crude balance realizing significantly looser than our expectations,” the strategists stated in the report. “For this week’s crude balance, from refineries, we model a minimal reduction in crude runs. Among net imports, we model a sharp reduction, with exports up (+0.7 million barrels per day) and imports down (-0.6 million barrels per day) on a nominal basis,” they added. In the report, the strategists warned that the timing of cargoes remains a source of potential volatility in this week’s crude balance. “From implied domestic supply (prod.+adj.+transfers), we look for a bounce (+1.2 million barrels) on a nominal basis this week,” the strategists went on to state in the report. “Rounding out the picture, we anticipate another small increase in SPR [Strategic Petroleum Reserve] stocks (+0.2 million barrels) this week,” they added. The Macquarie strategists also noted in the report that, “among products”, they “look for across the board builds (gasoline/distillate/ jet +1.0/+2.4/+1.1 million barrels)”. “We model implied demand for these three products at ~14.3 million barrels per day for the week ending July 4 amidst holiday effects,” they added. “On this front, while the week of July 4th typically sees a large reduction to distillate demand, the timing/magnitude of this impact could potentially be affected/reduced by the holiday falling on a Friday this year,” the strategists went on to note. In its latest weekly petroleum status report at the time of writing, which was released on July 2 and included data for the week ending June 27, the

Read More »

Hess Relinquishes Block 59 offshore Suriname after Failing to Find Partners

Hess Corp. decided to discontinue Block 59 exploration after failing to bring in new partners following Equinor ASA and Exxon Mobil Corp.’s withdrawal, Suriname’s national oil company said Tuesday. In July 2024 Norway’s majority state-owned Equinor and Texas-based ExxonMobil withdrew from the production sharing contract they signed July 2017 with New York City-based Hess. After 2D and 3D research that involved 6,000 kilometers (3,728.23 miles) and 9,000 square kilometers (3,474.92 square miles) respectively, ExxonMobil and Equinor deemed the risk for drilling an exploration well too high and transferred their stakes to Hess, according to Staatsolie Maatschappij Suriname NV. “Hess fulfilled its minimum work obligations and has decided not to move forward to the next phase of the exploration period, which concludes on 8 July 2025”, Staatsolie said in an online statement. Hess had failed to woo new partners to continue exploration, Staatsolie explained. PSC holders fully bear the costs and risks of exploration in the South American country, Staatsolie noted. Block 59 spanned about 11,480 square kilometers in waters 2,700-3,500 meters (8,858.27-11,482.94 feet) deep, according to Staatsolie. “Significant volumes are required for potential economically viable oilfield development in this block”, the statement said. “The area formerly designated as Block 59 will be incorporated into Staatsolie’s strategy to have as much of the offshore acreage under contract with international parties”, the statement said. “Currently, production sharing contracts are in place with a number of international oil and gas companies for the various blocks, covering approximately fifty percent of Suriname’s offshore”. In late 2024 ExxonMobil also exited Suriname’s Block 52 by transferring its 50 percent stake to partner Petroliam Nasional Bhd. (Petronas). “This withdrawal is part of ExxonMobil’s ongoing evaluation of assets in its global portfolio”, Staatsolie said in a press release November 20, 2024. “Staatsolie expects PETRONAS to continue the activities

Read More »

CoreWeave acquires Core Scientific for $9B to power AI infrastructure push

Such a shift, analysts say, could offer short-term benefits for enterprises, particularly in cost and access, but also introduces new operational risks. “This acquisition may potentially lower enterprise pricing through lease cost elimination and annual savings, while improving GPU access via expanded power capacity, enabling faster deployment of Nvidia chipsets and systems,” said Charlie Dai, VP and principal analyst at Forrester. “However, service reliability risks persist during this crypto-to-AI retrofitting.” This also indicates that struggling vendors such as Core Scientific and similar have a way to cash out, according to Yugal Joshi, partner at Everest Group. “However, it does not materially impact the availability of Nvidia GPUs and similar for enterprises,” Joshi added. “Consolidation does impact the pricing power of vendors.” Concerns for enterprises Rising demand for AI-ready infrastructure can raise concerns among enterprises, particularly over access to power-rich data centers and future capacity constraints. “The biggest concern that CIOs should have with this acquisition is that mature data center infrastructure with dedicated power is an acquisition target,” said Hyoun Park, CEO and chief analyst at Amalgam Insights. “This may turn out to create challenges for CIOs currently collocating data workloads or seeking to keep more of their data loads on private data centers rather than in the cloud.”

Read More »

CoreWeave achieves a first with Nvidia GB300 NVL72 deployment

The deployment, Kimball said, “brings Dell quality to the commodity space. Wins like this really validate what Dell has been doing in reshaping its portfolio to accommodate the needs of the market — both in the cloud and the enterprise.” Although concerns were voiced last year that Nvidia’s next-generation Blackwell data center processors had significant overheating problems when they were installed in high-capacity server racks, he said that a repeat performance is unlikely. Nvidia, said Kimball “has been very disciplined in its approach with its GPUs and not shipping silicon until it is ready. And Dell almost doubles down on this maniacal quality focus. I don’t mean to sound like I have blind faith, but I’ve watched both companies over the last several years be intentional in delivering product in volume. Especially as the competitive market starts to shape up more strongly, I expect there is an extremely high degree of confidence in quality.” CoreWeave ‘has one purpose’ He said, “like Lambda Labs, Crusoe and others, [CoreWeave] seemingly has one purpose (for now): deliver GPU capacity to the market. While I expect these cloud providers will expand in services, I think for now the type of customer employing services is on the early adopter side of AI. From an enterprise perspective, I have to think that organizations well into their AI journey are the consumers of CoreWeave.”  “CoreWeave is also being utilized by a lot of the model providers and tech vendors playing in the AI space,” Kimball pointed out. “For instance, it’s public knowledge that Microsoft, OpenAI, Meta, IBM and others use CoreWeave GPUs for model training and more. It makes sense. These are the customers that truly benefit from the performance lift that we see from generation to generation.”

Read More »

Oracle to power OpenAI’s AGI ambitions with 4.5GW expansion

“For CIOs, this shift means more competition for AI infrastructure. Over the next 12–24 months, securing capacity for AI workloads will likely get harder, not easier. Though cost is coming down but demand is increasing as well, due to which CIOs must plan earlier and build stronger partnerships to ensure availability,” said Pareekh Jain, CEO at EIIRTrend & Pareekh Consulting. He added that CIOs should expect longer wait times for AI infrastructure. To mitigate this, they should lock in capacity through reserved instances, diversify across regions and cloud providers, and work with vendors to align on long-term demand forecasts.  “Enterprises stand to benefit from more efficient and cost-effective AI infrastructure tailored to specialized AI workloads, significantly lower their overall future AI-related investments and expenses. Consequently, CIOs face a critical task: to analyze and predict the diverse AI workloads that will prevail across their organizations, business units, functions, and employee personas in the future. This foresight will be crucial in prioritizing and optimizing AI workloads for either in-house deployment or outsourced infrastructure, ensuring strategic and efficient resource allocation,” said Neil Shah, vice president at Counterpoint Research. Strategic pivot toward AI data centers The OpenAI-Oracle deal comes in stark contrast to developments earlier this year. In April, AWS was reported to be scaling back its plans for leasing new colocation capacity — a move that AWS Vice President for global data centers Kevin Miller described as routine capacity management, not a shift in long-term expansion plans. Still, these announcements raised questions around whether the hyperscale data center boom was beginning to plateau. “This isn’t a slowdown, it’s a strategic pivot. The era of building generic data center capacity is over. The new global imperative is a race for specialized, high-density, AI-ready compute. Hyperscalers are not slowing down; they are reallocating their capital to

Read More »

Arista Buys VeloCloud to reboot SD-WANs amid AI infrastructure shift

What this doesn’t answer is how Arista Networks plans to add newer, security-oriented Secure Access Service Edge (SASE) capabilities to VeloCloud’s older SD-WAN technology. Post-acquisition, it still has only some of the building blocks necessary to achieve this. Mapping AI However, in 2025 there is always more going on with networking acquisitions than simply adding another brick to the wall, and in this case it’s the way AI is changing data flows across networks. “In the new AI era, the concepts of what comprises a user and a site in a WAN have changed fundamentally. The introduction of agentic AI even changes what might be considered a user,” wrote Arista Networks CEO, Jayshree Ullal, in a blog highlighting AI’s effect on WAN architectures. “In addition to people accessing data on demand, new AI agents will be deployed to access data independently, adapting over time to solve problems and enhance user productivity,” she said. Specifically, WANs needed modernization to cope with the effect AI traffic flows are having on data center traffic. Sanjay Uppal, now VP and general manager of the new VeloCloud Division at Arista Networks, elaborated. “The next step in SD-WAN is to identify, secure and optimize agentic AI traffic across that distributed enterprise, this time from all end points across to branches, campus sites, and the different data center locations, both public and private,” he wrote. “The best way to grab this opportunity was in partnership with a networking systems leader, as customers were increasingly looking for a comprehensive solution from LAN/Campus across the WAN to the data center.”

Read More »

Data center capacity continues to shift to hyperscalers

However, even though colocation and on-premises data centers will continue to lose share, they will still continue to grow. They just won’t be growing as fast as hyperscalers. So, it creates the illusion of shrinkage when it’s actually just slower growth. In fact, after a sustained period of essentially no growth, on-premises data center capacity is receiving a boost thanks to genAI applications and GPU infrastructure. “While most enterprise workloads are gravitating towards cloud providers or to off-premise colo facilities, a substantial subset are staying on-premise, driving a substantial increase in enterprise GPU servers,” said John Dinsdale, a chief analyst at Synergy Research Group.

Read More »

Oracle inks $30 billion cloud deal, continuing its strong push into AI infrastructure.

He pointed out that, in addition to its continued growth, OCI has a remaining performance obligation (RPO) — total future revenue expected from contracts not yet reported as revenue — of $138 billion, a 41% increase, year over year. The company is benefiting from the immense demand for cloud computing largely driven by AI models. While traditionally an enterprise resource planning (ERP) company, Oracle launched OCI in 2016 and has been strategically investing in AI and data center infrastructure that can support gigawatts of capacity. Notably, it is a partner in the $500 billion SoftBank-backed Stargate project, along with OpenAI, Arm, Microsoft, and Nvidia, that will build out data center infrastructure in the US. Along with that, the company is reportedly spending about $40 billion on Nvidia chips for a massive new data center in Abilene, Texas, that will serve as Stargate’s first location in the country. Further, the company has signaled its plans to significantly increase its investment in Abu Dhabi to grow out its cloud and AI offerings in the UAE; has partnered with IBM to advance agentic AI; has launched more than 50 genAI use cases with Cohere; and is a key provider for ByteDance, which has said it plans to invest $20 billion in global cloud infrastructure this year, notably in Johor, Malaysia. Ellison’s plan: dominate the cloud world CTO and co-founder Larry Ellison announced in a recent earnings call Oracle’s intent to become No. 1 in cloud databases, cloud applications, and the construction and operation of cloud data centers. He said Oracle is uniquely positioned because it has so much enterprise data stored in its databases. He also highlighted the company’s flexible multi-cloud strategy and said that the latest version of its database, Oracle 23ai, is specifically tailored to the needs of AI workloads. Oracle

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »