Stay Ahead, Stay ONMINE

Aker BP, SLB, Stimwell Extend Alliance for Five More Years

Aker BP ASA has extended its alliance agreement with SLB (Schlumberger Limited) and Stimwell Services Ltd. for another five years. The alliance formed in 2019 has supported Aker BP’s operated assets in meeting production targets. Aker BP said in a media release the alliance will work on further accelerating and boosting oil production. The alliance […]

Aker BP ASA has extended its alliance agreement with SLB (Schlumberger Limited) and Stimwell Services Ltd. for another five years. The alliance formed in 2019 has supported Aker BP’s operated assets in meeting production targets.

Aker BP said in a media release the alliance will work on further accelerating and boosting oil production.

The alliance has established new standards for safe, efficient, and cost-effective operations through collaboration, digitalization, and innovative technology, according to Aker BP. Key achievements include simultaneous operations with jack-up rigs, a decreased backlog of locked-in barrels, and the first-ever autonomous intervention operation globally. Currently, both planning and execution utilize digital workflows, leading to enhanced productivity, reduced risk, and higher success rates, the company said.

“Strategic partnerships are essential to shaping the future of our industry. At Aker BP, we remain committed to the alliance model, which creates value through long-term collaboration. It enables us to increase productivity, maintain world-class performance, and deliver oil and gas with low cost and low emissions. This is how we position ourselves as the E&P company of the future”, Aker BP CEO Karl Johnny Hersvik said.

The alliance strategy aims to transform offshore well intervention.  Over the next five years, the partners aim to deliver top-quartile performance while developing future-proof capabilities, enhancing digital integration between subsurface and operations, expanding Aker BP’s Integrated Operations Centre for remote operations, and accelerating new technology deployment, Aker BP added.

Furthermore, the alliance will focus on bringing wells onstream across Aker BP’s projects, using a newly upgraded stimulation vessel to optimize the Valhall PWP wells, contributing significantly to future production.

“Increasing production and recovery from maturing assets is a top priority across the industry, and this alliance demonstrates how we can drive progress together through the power of partnership. The complex challenges facing our industry will increasingly require deep collaboration and trust across teams, and this alliance has been a cornerstone example of how powerful driving innovation together can be”, SLB CEO Olivier Le Peuch added.

“Stimulation has been critical in unlocking and increasing the recovery from tight reservoirs such as Valhall. During the last five years, the alliance working together managed to successfully develop very tight areas of the field by using innovative technology which significantly reduced the execution time and CO2 footprint and making it economic”, Sami Haidar, Managing Director at StimWell Servies, said.

To contact the author, email [email protected]

What do you think? We’d love to hear from you, join the conversation on the

Rigzone Energy Network.

The Rigzone Energy Network is a new social experience created for you and all energy professionals to Speak Up about our industry, share knowledge, connect with peers and industry insiders and engage in a professional community that will empower your career in energy.


MORE FROM THIS AUTHOR

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Cisco, Google Cloud offer enterprises new way to connect SD-WANs

Looking back, the rapid adoption of SaaS and cloud applications led to a WAN transformation and the emergence of SD-WAN via direct internet access, Sambi asserted. “Then, to enhance application performance, enterprises built colocation-based cloud on-ramps, which, while improving latency, introduced complexity and costs. This evolution led to a proliferation

Read More »

Fortinet embeds AI capabilities across Security Fabric platform

“By embedding FortiAI across the Fortinet Security Fabric platform, including new agentic AI capabilities, we’re empowering our customers to reduce the workload on their security and network analysts while improving the efficiency, speed, and accuracy of their security and networking operations,” said Michael Xie, founder, president, and chief technology officer

Read More »

Tailscale secures $160 million for its WireGuard-based VPN development

Building on WireGuard’s foundation At the heart of Tailscale’s technology is WireGuard, a modern VPN protocol that offers significant security and performance advantages over legacy solutions.  WireGuard is an open-source technology built in a way that minimizes the attack surface while providing greater performance than older VPN approaches. While WireGuard

Read More »

UALink releases inaugural GPU interconnect specification

UALink’s primary target for now is to provide an alternative to Nvdia’s high-bandwidth, low-latency, direct interconnect technology for CPU, GPU-to-GPU connectivity, NVLink. NVLink is primarily used in InfiniBand-based networks. Given the spec’s Ethernet heritage, UALink is seen in most circles as working hand-in-hand with the Ultra Ethernet Consortium to help

Read More »

Britain Utility Regulator Opens Applications for New LDES Investment Scheme

Britian’s Office of Gas and Electricity Markets (Ofgem) has opened applications under a new “cap and floor” scheme for so-called long-duration electricity storage (LDES) projects. The scheme seeks to produce the first LDES facilities in Britain in 40 years by easing hurdles such as high costs. Currently Britain has 2.8 gigawatts (GW) of LDES capacity, from four pumped storage hydro facilities in Scotland and Wales, according to Ofgem. The government sees LDES as key to achieving the United Kingdom’s goal of clean power by 2030 and net-zero emissions by 2050. “Long Duration Electricity Storage facilities provide vital back-up for the renewable power system – working like giant batteries that store electricity created by wind and solar farms, then release it to the grid when needed”, Ofgem said in an online statement. The inaugural application window is offering 2.7-7.7 GW, an indicative range that the National Energy Systems Operator (NESO) recommended to be installed by 2035. The official target for 2050 is to reach 20 GW of LDES capacity. “Producing one gigawatt constantly for a year is enough to power 2.65 million homes”, Ofgem said. Developers applying in window 1 have two tracks. One is for projects that can go onstream 2030 and the other is for projects that will be delivered 2033. “Ofgem will prioritize decisions on projects that can deliver by 2030 if necessary”, the statement said. The cap-and-floor scheme boosts “investor confidence with the security of providing minimum revenue for LDES operators to manage high start-up costs and long build times”, Ofgem said. “The scheme will ensure value for money by driving down costs, only allowing efficient projects with a storage capacity of more than eight hours. “At the same time, the regime protects consumers with a cap on profits, meaning any excess revenues flow directly back to customers

Read More »

Uncertainty is the Only Certainty, Commodities Analyst Warns

Uncertainty is the only certainty. That’s what Ole R. Hvalbye, Commodities Analyst at Skandinaviska Enskilda Banken AB (SEB), said in an oil report sent to Rigzone by the SEB team on Thursday, adding that Brent “spike[d]… $8 before sliding this morning”. “Brent crude prices have sharply rebounded since yesterday evening, reversing much of the steep losses seen during Tuesday afternoon and early Wednesday trading,” Hvalbye said in the report. “From a low of $58.4 per barrel, Brent surged overnight to $66.1, marking an immediate and notable $7.7 per barrel recovery. The brent price has declined to the current $63.8 per barrel during the morning,” he added. In the report, Hvalbye noted that market fundamentals remain sidelined as tariff rhetoric continues to dominate headlines. “The latest rally is largely driven by geopolitical developments, as the U.S. president announced a pause on scheduled additional tariffs for most countries, while simultaneously escalating measures against China,” he said. “In response to China’s retaliatory tariffs, the U.S. raised its tariff rate on Chinese goods to a high 125 percent,” he added. “Meanwhile, the president signaled a 90-day pause on ‘reciprocal’ tariffs targeting other trade partners, including the EU, as a ‘reward’ for not retaliating. This establishes a temporary tariff floor of 10 percent for all countries excluding China,” he continued. Hvalbye highlighted in the report that the market responded swiftly, noting that global equity and commodity markets surged, rebounding from recent lows. “Asian markets rallied this morning, while in the U.S., the S&P 500 closed up 9.5 percent, its strongest day since 2008. The Nasdaq Composite climbed more than 12 percent, its largest single-day gain since 2001,” Hvalbye said. “Crude prices are trading slightly down this morning, consolidating after the sharp rebound. Still, concerns over global oil demand persist, particularly in light of the growing

Read More »

USA Crude Flows to China Trickle to Near Zero After Tariff Blitz

The flow of oil from the world’s biggest producer to its largest importer is set to thin to virtually zero as a trade war between the two powerhouse economies escalates. After climbing in recent years, oil shipments from the US to China have been on the decline for much of 2025, thanks to US President Donald Trump’s successive rounds of tariffs at a time when the domestic refining sector is already under pressure. Beijing’s retaliation, raising tariffs on US imports to 84%, plus the latest US hike in duties on China to 125%, have only darkened the picture further. US flows are by no means vital to China — crude flows from the US to China in the early months of this year added up to roughly 1% of the Asian nation’s total imports, according to data from analytics firm Vortexa Ltd. —  but the collapse of oil purchases is indicative of a wider breakdown of trade relations between the world’s two largest economies. “With China imposing 84% tariffs on goods from the US, the cost of US crude would be almost double — $51 a barrel more expensive, based on $61 WTI,” said Ivan Mathews, head of APAC analysis for Vortexa. “This makes running US crude uneconomical for Chinese refiners.” US crude imports to China will “likely dwindle to zero in the coming months if the current tariff levels stay,” he added. Some of that crude would instead be going to other buyers in Asia. In recent days, Indian refiners have purchased cargoes to take advantage of lower prices, including grades from the US, as have oil processors in Japan. China, meanwhile, will likely fill the gap with supply from Middle Eastern producers like Oman or the United Arab Emirates, although it could also ramp up its buying of sensitive crude,

Read More »

EnergyHub partners with GM to use EVs as grid resources

EnergyHub, a distributed energy resource management services provider, has partnered with GM Energy, a General Motors business unit, to integrate electric vehicles and home energy storage solutions into utility and market programs across the country, the companies announced on April 2. “Our partnership with GM is a significant step forward in expanding the reach and impact of our utility clients’ EV and stationary storage programs,” Jeff Huron, senior manager of EV strategy and business development at EnergyHub, said in a statement.  EV adoption curves have flattened somewhat but sales are still growing. General Motors announced April 1 that its U.S. sales increased 17% overall in the first quarter of this year, compared with Q1 2024, while EV sales jumped 94%. EnergyHub said its EV managed charging programs can help utilities maintain grid stability through demand response and dynamic load shaping. “These capabilities enable peak demand management and aid efforts to ensure EVs and other DERs do not overload local distribution infrastructure,” it said. “Direct integrations with automakers like GM allow us to create the best possible experience for customers while ensuring utilities are prepared for the increasing demand from electric vehicles and distributed energy resources,” Huron said. Beginning later this year, Chevrolet, GMC and Cadillac EV owners will be able to enroll in select EnergyHub programs through their local utility. “This partnership enables smart or managed charging functionality, optimizing charging schedules to support bulk and distribution grid stability while maintaining customer mobility preferences,” EnergyHub said. GM has been expanding its EV partnerships. In March, GM Energy announced it joined Pacific Gas & Electric Co.’s vehicle-to-everything bidirectional charging pilot program for EVs in Northern and Central California.

Read More »

Trump aims to boost coal, in part by ordering power plants to stay open

Dive Brief: President Donald Trump signed four executive orders on Tuesday aimed at supporting coal mining and coal-fired power plants. One executive order directs U.S. Department of Energy Secretary Chris Wright to develop a process within 90 days for issuing emergency orders to keep power plants operating in areas of the country deemed to have potential grid reliability problems. The executive orders will likely have little effect on coal-fired generation or carbon markets given “constraints on the use of emergency authorities and the orders’ symbolic nature,” according to Capstone analysts. “While the order directs the [DOE] to focus on funding coal projects, we expect to see a limited pool of applicants.” Dive Insight: Coal-fired power plant capacity totaled 201 GW, or 15% of all U.S. capacity, at the start of this year, down from 305 GW, or 26% of U.S. capacity, at the end of 2015, according to the Federal Energy Regulatory Commission. With power demand growing, partly driven by data center development, the Trump administration aims to reverse the decline in coal-fired generation. Trump signed three other executive orders Tuesday. One order gives coal-fired power plants an extra two years to meet the Environmental Protection Agency’s Mercury and Air Toxics Standards, which was toughened by the Biden administration. The Trump administration aims to reverse the updated MATS rule. Another executive order directs the U.S. attorney general to challenge state climate and energy-related laws and policies that are deemed unconstitutional, preempted by federal law or unenforceable. Trump also aims to boost coal mining with another executive order, which also gives DOE, the Interior Department and Commerce Department 60 days to identify regions where coal-fired infrastructure is available to support AI data centers. During an executive order signing ceremony, Trump said he was directing Wright to keep open the 380-MW, coal-fired Cholla power

Read More »

Aramco, Sinopec Plan Petrochemical Complex at Saudi Refinery Site

Saudi Arabian Oil Co. (Aramco) and China Petroleum & Chemical Corp. (Sinopec) have signed a deal for the potential expansion of their Yanbu refinery to include petrochemical facilities. The “venture framework agreement”, signed in the 10th anniversary of Yanbu Aramco Sinopec Refining Co. (Yasref), “seeks to advance engineering studies for the development of a fully-integrated petrochemical complex”, the Saudi oil giant said in an online statement. Yanbu Aramco Sinopec Refining, 62.5 percent owned by Aramco and 37.5 percent by Sinopec, manages the refinery in the Yanbu Industrial City on the west coast of Saudi Arabia. Spanning about 5.2 million square meters (55.97 million square feet), the refinery processes 400,000 barrels a day of Arabian heavy crude oil, according to Yasref. The potential expansion involves the construction of a mixed feed steam cracker with a capacity of 1.8 million metric tons per annum (MMtpa) and a 1.5-MMtpa aromatics complex with associated downstream derivatives. “The planned Yasref expansion aligns with our downstream strategy to unlock the full potential of our resources, including converting up to four million barrels per day of crude oil into petrochemicals by 2030”, said Aramco president for downstream Mohammed Y Al Qahtani. “The planned expansion project solidifies our commitment to product innovation and diversification”, Aramco president and chief executive Amin H. Nasser said. “As we look forward to strengthening our collaboration with Sinopec in making Yasref a leading refining and petrochemicals joint venture, we aim to contribute to growing Saudi Arabia’s position as a global leader in energy and chemicals”. Zhao Dong, president of China’s state-owned Sinopec, said, “We expect the Yasref expansion project to unlock new dimensions of collaborative potential as we navigate the energy transition”. Besides Yasref, Aramco and Sinopec are also partners in Fujian Refining & Petrochemical Co., Sinopec SABIC Tianjin Petrochemical Co., Sinopec Senmei (Fujian)

Read More »

China’s rare earth export controls threaten enterprise IT hardware supply chains

“AI-first infrastructure rollouts — particularly those involving GPUs, edge accelerators, and high-efficiency cooling — are directly in the crosshairs,” Gogia noted. “So are quantum computing R&D efforts and high-reliability storage systems where thermal and magnetic materials matter.” China, responsible for 70% of global rare earth mining output and 87% of refined supply, poses a serious threat to enterprise IT hardware supply chains with these restrictions — especially for companies with AI-optimized server lines. AI chip production under threat The impact on semiconductor manufacturing comes at a critical time when enterprise demand for AI chips is soaring. Companies including Nvidia, AMD, Intel, and TSMC rely on rare earth elements during the manufacturing of advanced chips. “We see the greatest exposure in private data center expansion projects, AI inferencing at the edge, and next-gen device manufacturing, including specialized industrial IoT and robotics,” noted Gogia. Major cloud providers have been aggressively expanding their AI compute capacity, with substantial hardware refreshes planned for late 2025. These plans may now face delays or cost increases as chip manufacturers grapple with supply constraints. Pricing pressures to be felt in 3-6 months The immediate impact is expected to be limited as manufacturers work through existing inventory, but pricing pressure could emerge within 3-6 months, experts feel.

Read More »

DARPA backs multiple quantum paths in benchmarking initiative

Nord Quantique plans to use the money to expand its team, says Julien Camirand Lemyre, the company’s president, CTO and co-founder. That’s an opportunity to accelerate the development of the technology, he says. “By extension, what this will mean for enterprise users is that quantum solutions to real-world business problems will be available sooner, due to that acceleration,” he says. “And so enterprise customers need to also accelerate how they are thinking about adoption because the advantages quantum will provide will be tangible.” Lemyre predicts that useful quantum computers will be available for enterprises before the end of the decade. “In fact, there has been tremendous progress across the entire quantum sector in recent years,” he says. “This means industry needs to begin thinking seriously about how they will integrate quantum computing into their operations over the medium term.” “We’re seeing, with the deployment of programs like the QBI in the US and investments of billions of dollars from  public and private investors globally, an increasing maturity of quantum technologies,” said Paul Terry, CEO at Photonic, which is betting on optically-linked silicon spin qubits.  “Our architecture has been designed from day one to build modular, scalable, fault-tolerant quantum systems able to be deployed in data centers,” he said. He’s not the only one to mention fault-tolerance. DARPA stressed fault-tolerance in its announcement, and its selections point to the importance of error correction for the future of quantum computing. The biggest problem with today’s quantum computers is that the number of errors increases faster than the number of qubits, making them impossible to scale up. Quantum companies are working on a variety of approaches to reduce the error rates low enough that quantum computers can get big enough to actually to real work.

Read More »

Zayo’s Fiber Bet: Scaling Long-Haul and Metro Networks for AI Data Centers

Zayo Group Holdings Inc. has emerged as one of the most aggressive fiber infrastructure players in North America, particularly in the context of AI-driven growth. With a $4 billion investment in AI-related long-haul fiber expansion, Zayo is positioning itself as a critical enabler of the AI and cloud computing boom. The company is aggressively expanding its long-haul fiber network, adding over 5,000 route miles to accommodate the anticipated 2-6X increase in AI-driven data center capacity by 2030. This initiative comes as AI workloads continue to push the limits of existing network infrastructure, particularly in long-haul connectivity. New Fiber Routes The new routes include critical connections between key AI data center hubs, such as Chicago-Columbus, Las Vegas-Reno, Atlanta-Ashburn, and Columbus-Indianapolis, among others. Additionally, Zayo is overbuilding seven existing routes to further enhance network performance, resiliency, and low-latency connectivity. This new development is a follow-on to 15 new long haul routes representing over 5300 route miles of new and expanded capacity deployed over the last five years. These route locations were selected based on expected data center growth, power availability, existing capacity constraints, and specific regional considerations. The AI Data Center Sector: A Significant Driver of Fiber Infrastructure The exponential growth of AI-driven data center demand means that the U.S. faces a potential bandwidth shortage. Zayo’s investments look to ensure that long-haul fiber capacity keeps pace with this growth, allowing AI data centers to efficiently transmit data between key markets. This is especially important as data center development locations are being driven more by land and power availability rather than proximity to market. Emerging AI data center markets get the high speed fiber they need, especially as they are moving away from expensive power regions (e.g., California, Virginia) to lower-cost locations (e.g., Ohio, Nevada, Midwest). Without the high-speed networking capabilities offered by

Read More »

Crusoe Adds 4.5 GW Natural Gas to Fuel AI, Expands Abilene Data Center to 1.2 GW

Crusoe and the Lancium Clean Campus: A New Model for Power-Optimized Compute Crusoe Energy’s 300-megawatt deployment at the Lancium Clean Campus in Abilene is a significant marker of how data center strategies are evolving to integrate more deeply with energy markets. By leveraging demand flexibility, stranded power, and renewable energy, Crusoe is following a path similar to some of the most forward-thinking projects in the data center industry. But it’s also pushing the model further—fusing AI and high-performance computing (HPC) with the next generation of power-responsive infrastructure. Here’s how Crusoe’s strategy compares to some of the industry’s most notable power-driven data center deployments: Google’s Oklahoma Data Center: Proximity to Renewable Growth A close parallel to Crusoe’s energy-centric site selection strategy is Google’s Mayes County data center in Oklahoma. Google sited its facility there to take advantage of abundant wind energy, aligning with the local power grid’s renewable capacity. Similarly, Crusoe is tapping into Texas’s deregulated energy market, optimizing for low-cost renewable power and the ability to flexibly scale compute operations in response to grid conditions. Google has also been an industry leader in time-matching workloads to renewable energy availability, something that Crusoe is enabling in real time through grid-responsive compute orchestration. Sabey Data Centers in Quincy: Low-Cost Power as a Foundation Another instructive comparison is Sabey Data Centers’ Quincy, Washington, campus, which was built around one of the most cost-effective power sources in the U.S.—abundant hydroelectric energy. Sabey’s long-term strategy has been to co-locate power-intensive compute infrastructure near predictable, low-cost energy sources. Crusoe’s project applies a similar logic but adapts it for a variable grid environment. Instead of relying on a fixed low-cost power source like hydro, Crusoe dynamically adjusts to real-time energy availability, a strategy that could become a model for future power-aware, AI-driven workloads. Compass and Aligned: Modular, Energy-Adaptive

Read More »

Executive Roundtable: Data Center Site Selection and Market Evolution in a Constrained Environment

For the third installment of our Executive Roundtable for the First Quarter of 2025, we asked our panel of seasoned industry experts about how the dynamics of data center site selection have never been more complex—or more critical to long-term success. In an industry where speed to market is paramount, operators must now navigate an increasingly constrained landscape in the age of AI, ultra cloud and hyperscale expansion, marked by fierce competition for land, tightening power availability, and evolving local regulations.  Traditional core markets such as Northern Virginia, Dallas, and Phoenix remain essential, but supply constraints and permitting challenges are prompting developers to rethink their approach. As hyperscalers and colocation providers push the boundaries of site selection strategy, secondary and edge markets are emerging as viable alternatives, driven by favorable energy economics, infrastructure investment, and shifting customer demand.  At the same time, power procurement is now reshaping the equation. With grid limitations and interconnection delays creating uncertainty in major hubs, operators are exploring new solutions, from direct utility partnerships to on-site generation with renewables, natural gas, and burgeoning modular nuclear concepts. The question now is not just where to build but how to ensure long-term operational resilience. As data center demand accelerates, operators face mounting challenges in securing suitable land, reliable power, and regulatory approvals in both established and emerging markets.  And so we asked our distinguished executive panel for the First Quarter of 2025, with grid capacity constraints, zoning complexities, and heightened competition shaping development decisions, how are companies refining their site selection strategies in Q1 2025 to balance speed to market, scalability, and sustainability? And, which North American regions are showing the greatest potential as the next wave of data center expansion takes shape?

Read More »

Podcast: iMasons CEO Santiago Suinaga on the Future of Sustainable AI Data Centers

For this episode of the DCF Show podcast, host Matt Vincent, Editor in Chief of Data Center Frontier, is joined by Santiago Suinaga, CEO of Infrastructure Masons (iMasons), to explore the urgent challenges of scaling data center construction while maintaining sustainability commitments, among other pertinent industry topics. The AI Race and Responsible Construction “Balancing scale and sustainability is key because the AI race is real,” Suinaga emphasizes. “Forecasted capacities have skyrocketed to meet AI demand. Hyperscale end users and data center developers are deploying high volumes to secure capacity in an increasingly constrained global market.” This surge in demand pressures the industry to build faster than ever before. Yet, as Suinaga notes, speed and sustainability must go hand in hand. “The industry must embrace a build fast, build smart mentality. Leveraging digital twin technology, AI-driven design optimization, and circular economy principles is critical.” Sustainability, he argues, should be embedded at every stage of new builds, from integrating low-carbon materials to optimizing energy efficiency from the outset. “We can’t afford to compromise sustainability for speed. Instead, we must integrate renewable energy sources and partner with local governments, utilities, and energy providers to accelerate responsible construction.” A key example of this thinking is peak shaving—using redundant infrastructure and idle capacities to power the grid when data center demand is low. “99.99% of the time, this excess capacity can support local communities, while ensuring the data center retains prioritized energy supply when needed.” Addressing Embodied Carbon and Supply Chain Accountability Decarbonization is a cornerstone of iMasons’ efforts, particularly through the iMasons Climate Accord. Suinaga highlights the importance of tackling embodied carbon—the emissions embedded in data center construction materials and IT hardware. “We need standardized reporting metrics and supplier accountability to drive meaningful change,” he says. “Greater transparency across the supply chain can be

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »