Stay Ahead, Stay ONMINE

Five breakthroughs that make OpenAI’s o3 a turning point for AI — and one big challenge

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More The end of the year 2024 has brought reckonings for artificial intelligence, as industry insiders feared progress toward even more intelligent AI is slowing down. But OpenAI’s o3 model, announced just last week, has sparked a […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


The end of the year 2024 has brought reckonings for artificial intelligence, as industry insiders feared progress toward even more intelligent AI is slowing down. But OpenAI’s o3 model, announced just last week, has sparked a fresh wave of excitement and debate, and suggests big improvements are still to come in 2025 and beyond.

This model, announced for safety testing among researchers, but not yet released publicly, achieved an impressive score on the important ARC metric. The benchmark was created by François Chollet, a renowned AI researcher and creator of the Keras deep learning framework, and is specifically designed to measure a model’s ability to handle novel, intelligent tasks. As such, it provides a meaningful gauge of progress toward truly intelligent AI systems.

Notably, o3 scored 75.7% on the ARC benchmark under standard compute conditions and 87.5% using high compute, significantly surpassing previous state-of-the-art results, such as the 53% scored by Claude 3.5.

This achievement by o3 represents a surprising advancement, according to Chollet, who had been a critic of the ability of large language models (LLMs) to achieve this sort of intelligence. It highlights innovations that could accelerate progress toward superior intelligence, whether we call it artificial general intelligence (AGI) or not.

AGI is a hyped term, and ill-defined, but it signals a goal: intelligence capable of adapting to novel challenges or questions in ways that surpass human abilities.

OpenAI’s o3 tackles specific hurdles in reasoning and adaptability that have long stymied large language models. At the same time, it exposes challenges, including the high costs and efficiency bottlenecks inherent in pushing these systems to their limits. This article will explore five key innovations behind the o3 model, many of which are underpinned by advancements in reinforcement learning (RL). It will draw on insights from industry leaders, OpenAI’s claims, and above all Chollet’s important analysis, to unpack what this breakthrough means for the future of AI as we move into 2025.

The five core innovations of o3

1. “Program synthesis” for task adaptation

OpenAI’s o3 model introduces a new capability called “program synthesis,” which enables it to dynamically combine things that it learned during pre-training—specific patterns, algorithms, or methods—into new configurations. These things might include mathematical operations, code snippets, or logical procedures that the model has encountered and generalized during its extensive training on diverse datasets. Most significantly, program synthesis allows o3 to address tasks it has never directly seen in training, such as solving advanced coding challenges or tackling novel logic puzzles that require reasoning beyond rote application of learned information. François Chollet describes program synthesis as a system’s ability to recombine known tools in innovative ways—like a chef crafting a unique dish using familiar ingredients. This feature marks a departure from earlier models, which primarily retrieve and apply pre-learned knowledge without reconfiguration — and it’s also one that Chollet had advocated for months ago as the only viable way forward to better intelligence. 

At the heart of o3’s adaptability is its use of Chains of Thought (CoTs) and a sophisticated search process that takes place during inference—when the model is actively generating answers in a real-world or deployed setting. These CoTs are step-by-step natural language instructions the model generates to explore solutions. Guided by an evaluator model, o3 actively generates multiple solution paths and evaluates them to determine the most promising option. This approach mirrors human problem-solving, where we brainstorm different methods before choosing the best fit. For example, in mathematical reasoning tasks, o3 generates and evaluates alternative strategies to arrive at accurate solutions. Competitors like Anthropic and Google have experimented with similar approaches, but OpenAI’s implementation sets a new standard.

3. Evaluator model: A new kind of reasoning

O3 actively generates multiple solution paths during inference, evaluating each with the help of an integrated evaluator model to determine the most promising option. By training the evaluator on expert-labeled data, OpenAI ensures that o3 develops a strong capacity to reason through complex, multi-step problems. This feature enables the model to act as a judge of its own reasoning, moving large language models closer to being able to “think” rather than simply respond.

4. Executing Its own programs

One of the most groundbreaking features of o3 is its ability to execute its own Chains of Thought (CoTs) as tools for adaptive problem-solving. Traditionally, CoTs have been used as step-by-step reasoning frameworks to solve specific problems. OpenAI’s o3 extends this concept by leveraging CoTs as reusable building blocks, allowing the model to approach novel challenges with greater adaptability. Over time, these CoTs become structured records of problem-solving strategies, akin to how humans document and refine their learning through experience. This ability demonstrates how o3 is pushing the frontier in adaptive reasoning. According to OpenAI engineer Nat McAleese, o3’s performance on unseen programming challenges, such as achieving a CodeForces rating above 2700, showcases its innovative use of CoTs to rival top competitive programmers. This 2700 rating places the model at “Grandmaster” level, among the top echelon of competitive programmers globally.

O3 leverages a deep learning-driven approach during inference to evaluate and refine potential solutions to complex problems. This process involves generating multiple solution paths and using patterns learned during training to assess their viability. François Chollet and other experts have noted that this reliance on ‘indirect evaluations’—where solutions are judged based on internal metrics rather than tested in real-world scenarios—can limit the model’s robustness when applied to unpredictable or enterprise-specific contexts.

Additionally, o3’s dependence on expert-labeled datasets for training its evaluator model raises concerns about scalability. While these datasets enhance precision, they also require significant human oversight, which can restrict the system’s adaptability and cost-efficiency. Chollet highlights that these trade-offs illustrate the challenges of scaling reasoning systems beyond controlled benchmarks like ARC-AGI.

Ultimately, this approach demonstrates both the potential and limitations of integrating deep learning techniques with programmatic problem-solving. While o3’s innovations showcase progress, they also underscore the complexities of building truly generalizable AI systems.

The big challenge to o3

OpenAI’s o3 model achieves impressive results but at significant computational cost, consuming millions of tokens per task — and this costly approach is model’s biggest challenge. François Chollet, Nat McAleese, and others highlight concerns about the economic feasibility of such models, emphasizing the need for innovations that balance performance with affordability.

The o3 release has sparked attention across the AI community. Competitors such as Google with Gemini 2 and Chinese firms like DeepSeek 3 are also advancing, making direct comparisons challenging until these models are more widely tested.

Opinions on o3 are divided: some laud its technical strides, while others cite high costs and a lack of transparency, suggesting its real value will only become clear with broader testing. One of the biggest critiques came from Google DeepMind’s Denny Zhou, who implicitly attacked the model’s reliance on reinforcement learning (RL) scaling and search mechanisms as a potential “dead end,” arguing instead that a model should be able to learn to reason from simpler fine-tuning processes.

What this means for enterprise AI

Whether or not it represents the perfect direction for further innovation, for enterprises, o3’s new-found adaptability shows that AI will in one way or another continue to transform industries, from customer service and scientific research, in the future.

Industry players will need some time to digest what o3 has delivered here. For enterprises concerned about o3’s high computational costs, OpenAI’s upcoming release of the scaled-down “o3-mini” version of the model provides a potential alternative. While it sacrifices some of the full model’s capabilities, o3-mini promises a more affordable option for businesses to experiment with — retaining much of the core innovation while significantly reducing test-time compute requirements.

It may be some time before enterprise companies can get their hands on the o3 model. OpenAI says the o3-mini is expected to launch by the end of January. The full o3 release will follow after, though the timelines depend on feedback and insights gained during the current safety testing phase. Enterprise companies will be well advised to test it out. They’ll want to ground the model with their data and use cases and see how it really works.

But in the mean time, they can already use the many other competent models that are already out and well tested, including the flagship o4 model and other competing models — many of which are already robust enough for building intelligent, tailored applications that deliver practical value.

Indeed, next year, we’ll be operating on two gears. The first is in achieving practical value from AI applications, and fleshing out what models can do with AI agents, and other innovations already achieved. The second will be sitting back with the popcorn and seeing how the intelligence race plays out — and any progress will just be icing on the cake that has already been delivered.

For more on o3’s innovations, watch the full YouTube discussion between myself and Sam Witteveen below, and follow VentureBeat for ongoing coverage of AI advancements.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

A CSO’s perspective: 8 cyber predictions for 2025

As we step into 2025, the cyberthreat landscape is once again more dynamic and challenging than the year before. In 2024, we witnessed a remarkable acceleration in cyberattacks of all types, many fueled by advancements in generative AI. For security leaders, the stakes are higher than ever. In this post,

Read More »

Ericsson unveils genAI assistant for 5G network operations

Telecommunications and networking provider Ericsson recently launched its generative AI-based virtual assistant that uses large language model (LLM) technology to read, understand, and generate new content to provide personalized answers for network operators configuring wireless 5G networks, troubleshooting problems, and creating policies. Ericsson’s AI-based NetCloud Assistant, or ANA, is LLM-based

Read More »

BP to slash nearly 8,000 jobs as it looks to cut costs

London-listed supermajor BP (LON:BP) has confirmed plans to cut 8,000 jobs as it continues its plans to “simplify and focus” the company. The oil giant plans to reduce its internal headcount by 4,700 while reducing contractor numbers by 3,000. It is understood that staff were told about BPs plans on Thursday morning as CEO Murray Auchincloss issued a statement to employees. He said: “I understand and recognise the uncertainty this brings for everyone whose job may be at risk, and also the effect it can have on colleagues and teams. “We have a range of support available, and please continue to show care for each other, be considerate, and keep putting safety first – especially during times of change.” Last year the company boss announced plans to cut costs by $2 billion (£1.64bn) by the end of next year and this move is set to play a part in the penny-pinching drive. In the third quarter of 2024, BP saw profits slump by 18% to $2.3 billion (£1.8bn). Analysts had expected the drop in net income which was 30% less than the same period last year. © Supplied by BPWorkers from the Seagull tieback project on the BP’s ETAP platform. The supermajor employs 90,000 people across its global operations. BP was asked what number of the 7,700 job cuts will come from the UK. This comes as Unite the Union announced that its No Ban Without a Plan campaign received the backing of more than half of MSPs. The union initiative which aims to save the jobs of offshore workers as the UK transitions to renewable energy has received the backing of 65 MSPs and 6 MPs on a list with 71 political backers. However, Unite has shared that none of its supporters are representatives of the Green or Liberal

Read More »

USA EIA Reveals Latest Brent Oil Price Forecasts

The U.S. Energy Information Administration (EIA) revealed its latest Brent spot price forecasts in its January short term energy outlook (STEO), which was published this week. According to the STEO, the EIA sees the 2025 Brent spot price averaging $74.31 per barrel and the 2026 Brent spot price coming in at $66.46 per barrel. The EIA’s previous STEO projected that the 2025 Brent spot price would average $73.58 per barrel. That STEO did not offer a Brent spot price projection for 2026. The EIA expects the Brent spot price to average $76.34 per barrel in the first quarter of this year, $75 per barrel in the second quarter, $74 per barrel in the third quarter, $72 per barrel in the fourth quarter, $68.97 per barrel in the first quarter of 2026, $67.33 per barrel in the second quarter, $65.68 per barrel in the third quarter, and $64 per barrel in the fourth quarter of 2026, the January STEO showed. In its December STEO, the EIA forecast that the Brent spot price would average $74 per barrel in the first quarter of this year, $74.33 per barrel in the second quarter, $74 per barrel in the third quarter, and $72 per barrel in the fourth quarter. The EIA highlighted in the STEO that its forecast was completed before the United States issued additional sanctions targeting Russia’s oil sector on January 10. “For all of 2024, the Brent price averaged $81 per barrel and in 2023 averaged $82 per barrel,” the EIA noted in its latest STEO. “Following some initial upward price pressure in early 2025, we expect that crude oil prices will generally decline from mid-2025 through the end of 2026 as growth in global oil production outpaces growth in oil demand,” it added. “In our forecast, increases in oil prices

Read More »

Marcos Signs Law to Establish Downstream Gas Industry in Philippines

President Ferdinand Marcos Jr. has signed legislation to establish a downstream gas industry in the Philippines by increasing the share of gas in the domestic energy mix and positioning the country as a transshipment hub for liquefied natural gas in the Asia-Pacific. While climate and environmental campaigners opposing such a measure have called for expanding renewable energy instead, the new law insists it only promotes gas as a transition fuel and does not contradict existing policy for the shift to a lower-carbon future. The Philippine Natural Gas Industry Development Act seeks to “develop natural gas as a reliable fuel for power plants capable of addressing the peaking, mid-merit, and baseload demand of the country to help achieve energy security, while progressively transitioning to renewable energy sources”, according to the text published Wednesday on the government’s Official Gazette and the Senate. Marcos signed it January 8. The government shall also facilitate the development of “non-power end-uses of natural gas which include commercial, industrial, residential, and transport applications that promote fuel diversity”, states the law, codified as Republic Act 12120. “In all cases, the State shall ensure the safe, secure, reliable, transparent, competitive, and environmentally responsible operation of the PDNGI [Philippine downstream natural gas industry] value chain, and ensure alignment with the State’s policy on transitioning to a low-carbon future, consistent with the sustainable development goals on increasing the share of renewable energy in the country’s energy mix”, the law says. The law prioritizes locally-produced gas over imports so long as this restriction “is consistent with the State’s policy of ensuring energy security and consumer welfare”. However, the coal-reliant Southeast Asian country has only one active gas field out of two commercial discoveries, according to the Philippine Department of Energy (DOE). And Malampaya, offshore Palawan island, is depleting. Co-developer Shell PLC exited the

Read More »

Trident Seals Chevron Buy, TotalEnergies Swap in Congo-Brazzaville

Trident Energy has announced the completion of a suite of transactions involving the acquisition of Chevron Overseas (Congo) Ltd. and a swap with TotalEnergies SE covering several fields in the Republic of the Congo. With the sale of its local subsidiary, United States energy giant Chevron Corp. transferred its 31.5 percent non-operating stakes in the Moho-Bilondo block and the Nkossa and Nsoko II fields, as well as a 15.75 percent operating stake in the Lianzi field, to London-based Trident Energy. Trident Energy also obtained TotalEnergies’ operating 53.5 percent stakes in Nkossa and Nsoko II, raising its ownership in the two fields to 85 percent. State-owned Societe Nationale des Petroles du Congo (SNPC) has the remaining 15 percent. In exchange the French energy major acquired an additional 10 percent stake in the Moho-Bilondo block, which consists of the Moho-Bilondo field and the Moho Nord field. Operator TotalEnergies now holds a 63.5 percent stake in the license. Trident Energy has 21.5 percent. SNPC owns the remaining 15 percent. Trident Energy expects the new acquisitions to add about 30,000 barrels of oil per day (bopd) to its production. In the Moho-Bilondo block, Moho Nord is producing with a capacity of 140,000 bopd, according to information on TotalEnergies’ website. Meanwhile the mature fields Nkossa and Nsoko II, located 70 kilometers (43.5 miles) off the coast, have a combined average output of 15,000 barrels of oil equivalent a day, according to TotalEnergies. The transactions are “significant” for Trident Energy, “which has proven expertise in extending field life and unlocking production from mid-life assets as demonstrated by their takeovers in Brazil and Equatorial Guinea”, Trident Energy said in an online statement announcing completion. Trident Energy chief executive Jean-Michel Jacoulot said, “We look forward to working with TotalEnergies Congo, the SNPC and the Congolese government to generate further

Read More »

Analysts Look at USA Sanctions Effect

Crude oil’s strong start to the year was reinforced on January 10 with the announcement of the latest, and most extensive yet, round of U.S. sanctions on Russia. That’s what analysts at Standard Chartered Bank, including the company’s commodities research head Paul Horsnell, said in a report sent to Rigzone by Horsnell this week, adding that the new restrictions “roughly triple the number of directly sanctioned Russian crude oil tankers, enough to affect around 900,000 barrels per day”. “We do not expect Russia to be able to maintain the full extent of the flow, even with an increase in the use of shadow fleet tankers and ship-to-ship transfers, with perhaps an average 500,000 barrels per day of displacements over the next six months,” the Standard Chartered Bank analysts added in the report. “The global market had already tightened over the past three months, and the dislocation of Russian exports adds a further layer of prompt demand,” they went on to state. In a market analysis sent to Rigzone on Tuesday, Maria Agustina Patti, Financial Markets Strategist Consultant to Exness, said the U.S. sanctions “are expected to reduce Russian oil exports, potentially cutting up to 700,000 barrels per day from global supply”. “However, actual disruptions might be smaller, as Russia and its key buyers explore alternative shipping arrangements,” Patti added in the analysis. A research note sent to Rigzone late Tuesday by the JPM Commodities Research team said the estimated value of open interest across energy markets “increased by four percent week on week ($29 billion) to $668 billion”. “The increase was predominantly driven by crude oil and petroleum products which experienced healthy inflows of $12 billion during the week across all trader types,” the note added. “This was further supported by strong price action across WTI/Brent crude oil markets which rallied by

Read More »

Teesworks energy-from-waste incinerator group hits back at MP’s criticism

A group representing local councils who are building an energy-from-waste (EfW) incinerator in North East England has hit back after a local MP criticised the project and called for it to be halted. The Tees Valley Energy Recovery Facility (TVERF) is a joint venture between seven North East councils – Darlington, Hartlepool, Redcar & Cleveland, Stockton-on-Tees, Middlesbrough, Durham County and Newcastle City – which will burn over 400,000  tonnes of unrecyclable waste each year to generate electricity on the Teesworks site. Labour MP for Redcar Anna Turley wrote to the councils responsible stating her opposition to the project last week on the grounds that residents in Grangetown, adjacent to the planned incinerator, had expressed concerns about the health implications of the new facility being built. In response, the TVERF partners described recent media coverage and opposition to waste incineration as “scaremongering.” The project partners criticised a BBC investigation published in October, saying it was “highly selective” and did not “reflect the operational realities” of energy-from-waste facilities. They were also critical of local activists, who have been included in local press coverage. In September, healthcare professionals were among those who protested outside Newcastle Civic Centre against plans to send the city’s waste to Teesside. One of the protestors, Dr Matthew Keegan, said at the time: “This project will result in waste from across the whole region getting dumped on an area that has already been overburdened by pollution, inequality and government neglect.” His sentiments were echoed in Ms Turley’s criticisms. “Communities across Grangetown,” she wrote, “feel they are being asked to bear a disproportionate burden in the region’s waste management strategy.” She continued: “Grangetown, in particular, has shouldered more than its fair share of industrial developments. “It is time to explore alternative solutions that better balance the needs for the region

Read More »

Qualcomm purloins Intel’s chief Xeon designer with eyes toward data center development

If Intel was hoping for a turnaround in 2025, it will have to wait at least a little bit longer. The chief architect for Intel’s Xeon server processors has defected to chip rival Qualcomm, which is making yet another run at entering the data center market. Sailesh Kottapalli, a 28-year Intel veteran and a senior fellow and chief architect for the company’s Xeon processors, made the announcement on LinkedIn on January 13, stating that he joined Qualcomm as a senior vice president. “My journey took me through roles as a validation engineer, logic designer, full-chip floor planner, post-silicon debug engineer, micro architect, and architect,” he wrote. “I worked on CPU cores, memory, IO, and platform aspects of the system, spanning multiple architectures across x86 and Itanium, and products including CPU and GPU, most importantly shaping the Xeon product line.”

Read More »

8 Trends That Will Shape the Data Center Industry In 2025

What lies ahead for the data center industry in 2025? At Data Center Frontier, our eyes are always on the horizon, and we’re constantly talking with industry thought leaders to get their take on key trends. Our Magic 8 Ball prognostications did pretty well last year, so now it’s time to look ahead at what’s in store for the industry over the next 12 months, as we identify eight themes that stand to shape the data center business going forward. We’ll be writing in more depth about many of these trends, but this list provides a view of the topics that we believe will be most relevant in 2025. A publication about the future frontiers of data centers and AI shouldn’t be afraid to put it’s money where its mouth is, and that’s why we used AI tools to help research and compose this year’s annual industry trends forecast. The article is meant to be a bit encyclopedic in the spirit of a digest, less than an exactly prescriptive forecast – although we try to go there as well. The piece contains some dark horse trends. Do we think immersion cooling is going to explode this year, suddenly giving direct-to-chip a run for its money? Not exactly. But do we think that, given the enormous and rapidly expanding parameters of the AI and HPC boom, the sector for immersion cooling could see some breakthroughs this year? Seems reasonable. Ditto for the trends forecasting natural gas and quantum computing advancements. Such topics are definitely on the horizon and highly visible on the frontier of data centers, so we’d better learn more about them, was our thought. Because as borne out by recent history, data center industry trends that start at the bleeding edge (pun intended – also, on the list) sometimes

Read More »

Podcast: Data Center and AI Sustainability Imperatives with iMasons Climate Accord Executive Director, Miranda Gardiner

Miranda was a featured speaker at last September’s inaugural Data Center Frontier Trends Summit. The call for speakers is now open for this year’s event, which will be held again in Reston, Virginia from Aug. 26-28. DCF Show Podcast Quotes from Miranda Gardiner, Executive Director, iMasons Climate Accord On Her Career Journey and Early Passion for Sustainability:   – “My goals have always been kind of sustainability, affordable housing. I shared a story last week on a panel that my mother even found a yearbook of me from my elementary school years. The question that year was like, what do you hope for the future? And mine was there’d be no pollution and everyone would have a home.” On Transitioning to Data Centers:   – “We started to see this mission-critical focus in facilities like data centers, airports, and healthcare buildings. For me, connecting sustainability into the performance of the building made data centers the perfect match.” Overview of the iMasons Climate Accord:   – “The iMasons Climate Accord is an initiative started in 2022. The primary focus is emission reductions, and the only requirement to join is having an emission reduction strategy.”   – “This year, we refined our roadmap to include objectives such as having a climate strategy, incentivizing low-GHG materials like green concrete, and promoting equity by supporting small, women-owned, and minority-owned businesses.” On Industry Collaboration and Leadership:   – “This year, through the Climate Accord, we issued a call to action on the value of environmental product declarations (EPDs). It was signed by AWS, Digital Realty, Google, Microsoft, Schneider Electric, and Meta—talk about a big initiative and impact!” On EPDs and Carbon Disclosure:   – “EPDs provide third-party verification of materials coming into buildings. Pairing that with the Open Compute Project’s carbon disclosure labels on equipment creates vast opportunities for transparency and

Read More »

Accelsius and iM Data Centers Demo Next-Gen Cooling and Sustainability at Miami Data Center

Miami Data Center Developments Update Miami has recently witnessed several significant developments and investments in its data center sector, underscoring the city’s growing importance as a digital infrastructure hub. Notable projects include: Project Apollo:  A proposed 15-megawatt (MW), two-story, 75,000-square-foot data center in unincorporated Miami-Dade County. With an estimated investment of $150 million, construction is slated to commence between 2026 and 2027. The development team has prior experience with major companies such as Amazon, Meta, and Iron Mountain.  RadiusDC’s Acquisition of Miami I:  In August 2024, RadiusDC acquired the Miami I data center located in the Sweetwater area. Spanning 170,000 square feet across two stories, the facility currently offers 3.2MW of capacity, with plans to expand to 9.2 MW by the first half of 2026. The carrier-neutral facility provides connectivity to 11 fiber optic and network service providers.  Iron Mountain’s MIA-1 Data Center: Iron Mountain is developing a 150,000-square-foot, 16 MW data center on a 3.4-acre campus in Central North West Miami. The facility, known as MIA-1, is scheduled to open in 2026 and aims to serve enterprises, cloud providers, and large-scale users in South Florida. It will feature fiber connections to other Iron Mountain facilities and a robust pipeline of carriers and software-defined networks.  EDGNEX’s Investment Plans:  As of this month, Dubai, UAE-based EDGNEX has announced plans to invest $20 billion in the U.S. data center market, with the potential to double this investment. This plan includes a boutique condo project in Miami, estimated to have a $1 billion gross development value, indicating a significant commitment to the region’s digital infrastructure.  All of these developments highlight Miami’s strategic position as a connectivity hub, particularly serving as a gateway to Latin America and the Caribbean. The city’s data center market is characterized by steady growth, with a focus on retail colocation and

Read More »

Tract Capital Unveils Fleet Data Centers, Specializing In 500 MW+ Build-to-Suit Megacampuses

Tract Capital has announced the launch of Fleet Data Centers, a new platform dedicated to the development of mega-scale data center campuses with capacities of 500 MW or more, specifically designed for single-user customers.  The initiative is led by Grant van Rooyen, CEO of Tract Capital and Executive Chairman of Fleet Data Centers, and Chris Vonderhaar, the newly appointed President of Fleet Data Centers.  Vonderhaar brings extensive experience to the role, having served as Vice President of Demand and Supply Management at Google Cloud and as a senior leader at Amazon Web Services (AWS) for over a decade, where he oversaw the design, planning, construction, and operation of AWS’s global data center platform.  The Fleet leadership team also includes veterans from hyperscalers, wholesale data center providers, network infrastructure firms, and equipment vendors, with a collective track record of deploying dozens of gigawatts of data center capacity across hundreds of facilities globally. A Two Prong Strategy Defining two distinct strategies, Fleet is the mega-campus vertical development arm of Tract Capital, an alternative asset manager specializing in scaling digital infrastructure, which also operates Tract to refine development sites at ground level for data centers in terms of lining up power, fiber, zoning and entitlements.  Fleet Data Centers will aim to address the next phase of hyperscale data center growth by offering customized gigawatt-level campuses that provide predictability, flexibility, and scalability for hyperscalers navigating increasing infrastructure demands. This new venture from Tract Capital underscores the growing need for innovative, large-scale digital infrastructure solutions, particularly as hyperscalers face mounting challenges in scaling their global platforms to meet the demands of the digital age. The unveiling of Fleet is just another example of the way Tract Capital has consistently demonstrated its expertise in accelerating the scaling of responsible technology infrastructure, combining operational capabilities from industry

Read More »

Call for Speakers: Second Annual Data Center Frontier Trends Summit, Aug. 26-28, Reston, VA

Data Center Frontier (DCF) is excited to announce the Call for Speakers for our highly anticipated second annual Data Center Frontier Trends Summit, set to take place from August 26-28, 2025 in Reston, Virginia.  This premier industry event will once again bring together the brightest minds and leaders in the data center and digital infrastructure sectors to explore cutting-edge trends shaping the future of the industry.   Submit Speaking Proposals Here The DCF Trends Summit focuses on delivering deep insights and actionable knowledge for professionals navigating the evolving challenges and opportunities in data center innovation, energy efficiency, sustainability, and advanced technology integration. This year’s event will feature keynote speakers, expert panels, and interactive discussions on topics such as AI workloads, modular and edge computing, renewable energy strategies, and the global expansion of hyperscale facilities.   Call for Papers Details The DCF Trends Summit welcomes paper submissions on a wide range of relevant topics, including but not limited to: Emerging Trends:  AI, machine learning, and edge computing in data center operations. Power: Utility and substation power, renewables and behind-the-meter onsite, battery backup, energy storage. Sustainability:  Innovations in energy efficiency, renewable energy integration, and sustainable design. Technology Innovations:  Next-gen cooling systems, advanced automation, and breakthroughs in network infrastructure. National & Global Perspectives:  Regional market dynamics for site selection and regulation plus strategies for addressing evolving customer needs and workforce development.   View the Full DCF Trends ‘Topics of Interest’ Listing Industry professionals, researchers, and thought leaders are encouraged to submit papers that reflect their expertise, insights, and forward-looking perspectives. Submissions should align with the core themes of the Summit and provide actionable takeaways for attendees.   The deadline for paper submissions is January 29, 2025. All speakers will receive complimentary registration and the opportunity to share their work with a diverse audience

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »