Stay Ahead, Stay ONMINE

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more.

The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them.

In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks.

Going all-in on red teaming pays practical, competitive dividends

It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks.

Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find.

What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle design to combine human expertise and contextual intelligence on one side with AI-based techniques on the other.

“When automated red teaming is complemented by targeted human insight, the resulting defense strategy becomes significantly more resilient,” writes OpenAI in the first paper (Ahmad et al., 2024).

The company’s premise is that using external testers to identify the most high-impact real-world scenarios, while also evaluating AI outputs, leads to continuous model improvements. OpenAI contends that combining these methods delivers a multi-layered defense for their models that identify potential vulnerabilities quickly. Capturing and improving models with the human contextual intelligence made possible by a human-in-the-middle design is proving essential for red-teaming AI models.

Why red teaming is the strategic backbone of AI security

Red teaming has emerged as the preferred method for iteratively testing AI models. This kind of testing simulates a variety of lethal and unpredictable attacks and aims to identify their most potent and weakest points. Generative AI (gen AI) models are difficult to test through automated means alone, as they mimic human-generated content at scale. The practices described in OpenAI’s two papers seek to close the gaps automated testing alone leaves, by measuring and verifying a model’s claims of safety and security.

In the first paper (“OpenAI’s Approach to External Red Teaming”) OpenAI explains that red teaming is “a structured testing effort to find flaws and vulnerabilities in an AI system, often in a controlled environment and collaboration with developers” (Ahmad et al., 2024). Committed to leading the industry in red teaming, the company had over 100 external red teamers assigned to work across a broad base of adversarial scenarios during the pre-launch vetting of GPT-4 prior to launch.

Research firm Gartner reinforces the value of red teaming in its forecast, predicting that IT spending on gen AI will soar from $5 billion in 2024 to $39 billion by 2028. Gartner notes that the rapid adoption of gen AI and the proliferation of LLMs is significantly expanding these models’ attack surfaces, making red teaming essential in any release cycle.

Practical insights for security leaders

Even though security leaders have been quick to see the value of red teaming, few are following through by making a commitment to get it done. A recent Gartner survey finds that while 73% of organizations recognize the importance of dedicated red teams, only 28% actually maintain them. To close this gap, a simplified framework is needed that can be applied at scale to any new model, app, or platform’s red teaming needs.

In its paper on external red teaming OpenAI defines four key steps for using a human-in-the-middle design to make the most of human insights:

  • Defining testing scope and teams: Drawing on subject matter experts and specialists across key areas of cybersecurity, regional politics, and natural sciences, OpenAI targets risks that include voice mimicry and bias. The ability to recruit cross-functional experts is, therefore, crucial. (To gain an appreciation for how committed OpenAI is to this methodology and its implications for stopping deepfakes, please see our article “GPT-4: OpenAI’s shield against $40B deepfake threat to enterprises.”)
  • Selecting model versions for testing, then iterating them across diverse teams: Both of OpenAI’s papers emphasize that cycling red teams and models using an iterative approach delivers the most insightful results. Allowing each red team to cycle through all models is conducive to greater team learning of what is and isn’t working.
  • Clear documentation and guidance: Consistency in testing requires well-documented APIs, standardized report formats, and explicit feedback loops. These are essential elements for successful red teaming.
  • Making sure insights translate into practical and long-lasting mitigations: Once red teams log vulnerabilities, they drive targeted updates to models, policies and operational plans — ensuring security strategies evolve in lockstep with emerging threats.

Scaling adversarial testing with GPT-4T: The next frontier in red teaming

AI companies’ red teaming methodologies are demonstrating that while human expertise is resource-intensive, it remains crucial for in-depth testing of AI models.

In OpenAI’s second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning” (Beutel et al., 2024), OpenAI addresses the challenge of scaling adversarial testing using an automated, multi-pronged approach that combines human insights with AI-generated attack strategies.

The core of this methodology is GPT-4T, a specialized variant of the GPT-4 model engineered to produce a wide range of adversarial scenarios.

Here’s how each component of the methodology contributes to a stronger adversarial testing framework:

  • Goal diversification. OpenAI describes how it is using GPT-4T to create a broad spectrum of scenarios, starting with initially benign-seeming prompts and progressing to more sophisticated phishing campaigns. Goal diversification focuses on anticipating and exploring the widest possible range of potential exploits. By using GPT-4T’s capacity for diverse language generation, OpenAI contends that red teams avoid tunnel vision and stay focused on probing for vulnerabilities that manual-only methods miss.
  • Reinforcement learning (RL). A multi-step RL framework rewards the discovery of new and previously unseen vulnerabilities. The purpose is to train the automated red team by improving each iteration. This enables security leaders to refocus on genuine risks rather than sifting through volumes of low-impact alerts. It aligns with Gartner’s projection of a 30% drop in false positives attributable to gen AI in application security testing by 2027. OpenAI writes, “Our multi-step RL approach systematically rewards the discovery of newly identified vulnerabilities, driving continuous improvement in adversarial testing.”
  • Auto-generated rewards: OpenAI defines this as a system that tracks and updates scores for partial successes by red teams, assigning incremental rewards for identifying each unprotected weak area of a model.

Securing the future of AI: Key takeaways for security leaders

OpenAI’s recent papers show why a structured, iterative process that combines internal and external testing delivers the insights needed to keep improving models’ accuracy, safety, security and quality.

Security leaders’ key takeaways from these papers should include: 

Go all-in and adopt a multi-pronged approach to red teaming. The papers emphasize the value of combining external, human-led teams with real-time simulations of AI attacks generated randomly, as they reflect how chaotic intrusion attempts can be. OpenAI contends that while humans excel at spotting context-specific gaps, including biases, automated systems identify weaknesses that emerge only under stress testing and repeated sophisticated attacks.

Test early and continuously throughout model dev cycles. The white papers make a compelling argument against waiting for production-ready models and instead beginning testing with early-stage versions. The goal is to find emerging risks and retest later to make sure the gaps in models were closed before launch.

Whenever possible, streamline documentation and feedback with real-time feedback loops. Standardized reporting and well-documented APIs, along with explicit feedback loops, help convert red team findings into actionable, trackable mitigations. OpenAI emphasizes the need to get this process in place before beginning red teaming, to accelerate fixes and remediation of problem areas.

Using real-time reinforcement learning is critically important, as is the future of AI red teaming. OpenAI makes the case for automating frameworks that reward discoveries of new attack vectors as a core part of the real-time feedback loops. The goal of RL is to create a continuous loop of improvement. 

Don’t settle for anything less than actionable insights from the red team process. It’s essential to treat every red team discovery or finding as a catalyst for updating security strategies, improving incident response plans, and revamping guidelines as required.

Budget for the added expense of enlisting external expertise for red teams. A central premise of OpenAI’s approach to red teaming is to actively recruit outside specialists who have informed perspectives and knowledge of advanced threats. Areas of expertise valuable to AI-model red teams include deepfake technology, social engineering, identity theft, synthetic identity creation, and voice-based fraud. “Involving external specialists often surfaces hidden attack paths, including sophisticated social engineering and deepfake threats.” (Ahmad et al., 2024)

Papers:

Beutel, A., Xiao, K., Heidecke, J., & Weng, L. (2024). “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning.” OpenAI.

Ahmad, L., Agarwal, S., Lampe, M., & Mishkin, P. (2024). “OpenAI’s Approach to External Red Teaming for AI Models and Systems.” OpenAI.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

At Cisco Live, it’s all about AI for networking and security

Key features of the Cisco Deep Network Model include: Purpose-built for networking: 20% more precise reasoning for troubleshooting, configuration, and automation. Trusted training: Fine-tuned on 40+ years of expertise and expert-vetted for accuracy. Continuous learning: Evolves with live telemetry and real-world Cisco TAC and CX services insights. Another new Agentic

Read More »

Kyndryl, AWS unwrap AI-driven mainframe migration service

Kyndryl is tapping into agentic AI technology from AWS to offer enterprise customers another way to integrate with or move mainframe applications to the AWS Cloud platform. Specifically, Kyndryl will offer an implementation service based on the recently announced AWS Transform package, which analyzes mainframe code bases, decomposes them into

Read More »

Solar industry posts record Q1 growth but projects longer-term decline

Dive Brief: The U.S. doubled its solar cell manufacturing capacity and added 8.6 GW of solar module manufacturing capacity in the first quarter of 2025, marking the third-largest quarter for new solar manufacturing capacity on record, according to a report by Wood Mackenzie for the Solar Energy Industries Association. Despite the strong first quarter figures, Wood Mackenzie expects the solar industry to contract about 2% annually between 2025-2030, adding an average 43 GW of new solar generation per year in that time. New solar installations are expected to decline 7% between 2025-2027. Wood Mackenzie’s projections for the solar industry do not take the proposed wind-down of clean energy tax credits that has passed the House into account. Cutting the tax credits could trigger project cancellations and a possible energy shortage, according to Sean Gallagher, senior vice president of policy for SEIA. Dive Insight: Despite growing demand for energy, the solar industry faces a rocky road over the next few years — particularly if the Senate concurs with renewable energy tax credit cuts that have already passed the House, according to this week’s report from Wood Mackenzie and SEIA. Solar manufacturing posted particularly strong growth in the first quarter of 2025, though Wood Mackenzie notes that the growth upstream manufacturing of solar components, especially polysilicon and wafers, “remains slow or non-existent.” New solar generation capacity totaled 10.8 GW, 7% lower than first quarter installations in 2024 and 43% lower than the fourth quarter of 2024 — but still the fourth largest quarter for deployment on record, according to the report. The first quarter records don’t appear to represent an attempt by the industry to wrap up projects before the potential application of new tariffs or cuts to applicable tax credits, Gallagher said. Most of the projects that came online in the

Read More »

US Moves Some Diplomats as Iran Tensions Rise

The US ordered some staff to leave its embassy in Baghdad, officials said, after Iran threatened to strike American assets in the Middle East in the event it’s attacked over its nuclear program. The decision to reduce staffing in Iraq was “based on our latest analysis,” according to the US State Department. Defense Secretary Pete Hegseth authorized family members of US military stationed across the region to leave, according to a Pentagon statement. The State Department also said US government employees and family members in Israel are restricted from traveling outside major cities such as Tel Aviv and Jerusalem until further notice. None of the statements cited a specific threat. But they came after the New York Post published an interview with President Donald Trump in which he said he was less confident the US will reach a deal with Iran. The countries are negotiating an agreement that would curb the Islamic Republic’s nuclear activities in return for sanctions relief. Trump has consistently said he wants an agreement with Iran and to avoid a war, but that the US could resort to military action if Tehran doesn’t accept a deal. “I sincerely hope it won’t come to that and that the talks reach a resolution,” Iran’s Defense Minister Aziz Nasirzadeh said on Wednesday. “But if they don’t, and conflict is imposed on us, the other side will undoubtedly suffer greater losses.” Iran announced the start of military drills on Thursday, “with a focus on enemy movements.” The head of its Islamic Revolutionary Guard Corps said the force was “ready for any scenario.” The same day, the Islamic Republic said it would establish a new uranium-enrichment center in response to a decision by the United Nations atomic watchdog to censure it. Oil prices jumped on the reports the US would move

Read More »

JERA Enters Long-Term Deals to Buy Up To 5.5 MMtpa of US LNG

Six liquefied natural gas (LNG) projects in the United States have secured agreements with JERA Co. Inc. for the sale of up to 5.5 million metric tons per annum (MMtpa) over 20 years to the Japanese power utility. JERA signed a sale and purchase agreement (SPA) for about two MMtpa from NextDecade Corp.’s Rio Grande LNG in Brownsville, Texas. The agreement is for train 5; NextDecade has yet to make an FID (final investment decision) on the fourth and fifth trains. The three-train first phase of the project is under construction. Rio Grande LNG holds a permit to export to FTA and non-FTA countries. Another SPA involves around one MMtpa from Kimmeridge’s proposed Commonwealth LNG in Calcasieu Pass, Louisiana. In February the Department of Energy (DOE) granted the project a conditional permit to export to countries with no free trade agreement (FTA) with the U.S. Kimmeridge expects to make an FID in the third quarter of 2025 and start production 2029. JERA also signed an SPA with Venture Global Inc.’s CP2 LNG in Cameron Paris, Louisiana, for approximately one MMtpa. The project recently started site work after completing clearance from the Federal Energy Regulatory Commission. In March the DOE granted a conditional permit for CP2 LNG to export to non-FTA countries. The project has already secured authorization for its export volume when it received FTA approval April 2022. Meanwhile Sempra’s Port Arthur LNG phase 2 in Jefferson County, Texas secured heads of agreement with JERA for about 1.5 MMtpa. Late last month the DOE granted Port Arthur LNG Phase a non-FTA permit, marking the resumption of the issuance of final orders for LNG export following the previous administration’s pause. JERA also penned heads of agreement with Cheniere Energy Inc.’s Corpus Christi LNG and Sabine Pass LNG. Train 3 of Corpus

Read More »

GOP lawmakers reiterate asks for clean energy credit tweaks in reconciliation bill

Dive Brief: A group of 13 House Republicans sent a letter to their Senate counterparts Friday asking the Senate to make further alterations to changes made to the Inflation Reduction Act’s clean energy tax credits, as the chamber takes up its portion of the budget reconciliation process. The Republicans were led by Virginia Rep. Jen Kiggans, who said in a June 6 press release that though she supported the House-passed “One Big Beautiful Bill Act” as it stands, “there remains significant room for improvement in preserving the clean energy tax credits.” The requested changes largely mirror those requested by Kiggans and her colleagues prior to the House’s passage of the bill. The group’s letter cites more than $14 billion in clean energy projects that have been canceled this year, with signatories stating that “project cancellations will continue to snowball” if Congress does not provide “business certainty as these provisions are phased out.” Dive Insight: Kiggans and a nearly identical group of House Republicans wrote to their House colleagues last month seeking positive changes to the clean energy tax credit phaseout plan that came out of committee work on the reconciliation bill. Instead, the House-passed version of the bill imposes a faster phaseout timeline, including a repeal of tax credits for projects that don’t begin construction within 60 days of the bill’s enactment. The Republican lawmakers in support of IRA’s clean energy credits asked that the Senate alter that provision — along with a foreign entity of concern provision that was again called “overly prescriptive” — and maintain tax credit transferability throughout the credits’ lifetime. The group said that though they were “proud to have worked to ensure that the bill did not include a full repeal of the clean energy tax credits,” they remain “deeply concerned about those provisions.” The

Read More »

EPA proposes slashing power plant carbon, mercury emission limits

The Environmental Protection Agency on Wednesday proposed eliminating greenhouse gas emissions limits for power plants as well as repealing a toughening of the mercury and air toxics standard for coal- and oil-fired generators. “These Biden-era regulations have imposed massive costs on coal-, oil-, and gas-fired power plants, raising the cost of living for American families, imperiling the reliability of our electric grid, and limiting American energy prosperity,” the EPA said in a press release. The EPA estimates that its proposal to repeal Obama and Biden era GHG rules would save the power sector $19 billion over 20 years and reverting to the 2012 MATS rule would save $1.2 billion over a decade. Cooperative utilities, independent power producers, the National Mining Association and others praised the EPA’s decision. “Besides being illegal, this Biden-era [GHG] rule would have shut down the nation’s fleet of coal power plants despite warnings from experts and officials that more than half of the U.S. is at risk of electricity shortages,” said Michelle Bloodworth, CEO of America’s Power, a trade group for coal-fired power plant owners. In part, opponents of the EPA’s GHG rule argued that it relied on carbon capture and sequestration technology that isn’t commercially available. “EEI appreciates EPA’s acknowledgement that carbon capture and storage technologies are not yet viable for widespread deployment,” said Alex Bond, executive director of legal and clean energy policy at the Edison Electric Institute, a trade group for investor-owned utilities. “Electric companies need standards for natural gas facilities that are attainable to plan and permit new facilities, along with flexible regulatory approaches that help maintain dispatchable generation.” EEI continues to support the EPA’s authority to regulate GHG emissions under the Clean Air Act, according to Bond. Power plants account for about a quarter of U.S. carbon emissions, making them the second largest

Read More »

Meta-Constellation virtual PPA could be first of many deals for existing reactor output: experts

Dive Brief: Meta’s 20-year nuclear power purchase agreement with Constellation Energy could mark the start of a trend as profitable technology companies seek clean, firm electricity to power their AI ambitions, experts say. On Wednesday, Amazon Web Services and Talen Energy followed Meta’s June 3 announcement with news of a similar deal of their own. Investment bank Jefferies estimated the Meta-Constellation deal’s value around $70/MWh, or a roughly $20/MWh premium to Illinois’ energy-plus-capacity market compensation from 2028 and as much as $15/MWh higher than regional wind power purchase agreements. Constellation did not disclose the price. Nuclear plants in Illinois and other states with “pay to stay” subsidies are ripe for similar deals in the near future, energy consultancy ClearView Energy partners said Monday. Constellation’s 1,870-MW Quad Cities Clean Energy Center is one to watch as its reactors come up for relicensing in 2029 and 2031, ClearView said. Dive Insight: Meta will purchase the clean energy attributes of Constellation’s 1,092-MW Clinton Clean Energy Center in central Illinois beginning in June 2027, as its home state’s 10-year, ratepayer-funded Zero Emissions Credit program winds down. The ZEC program saved Clinton from premature closure in 2017 “after years of financial losses,” Constellation said on June 3. The virtual power purchase agreement will support a 30-MW power uprate and a 20-year operating license extension, which will allow the plant to operate until at least 2047. Meta’s power purchases will support its “clean energy goals and operations in the region,” rather than power data centers directly, Constellation said. Constellation called the deal “a market-based solution” that will replace the ZEC subsidy and ensure long-term plant operation without ratepayer support. “[Meta] figured out that supporting the relicensing and expansion of existing plants is just as impactful as finding new sources of energy,” Constellation CEO Joe Dominguez said

Read More »

Micron joins HBM4 race with 36GB 12-high stack, eyes AI and data center dominance

Race to power the next generation of AI By shipping samples of the HMB4 to the key customers, Micron has joined SK hynix in the HBM4 race. In March this year, SK hynix shipped the 12-Layer HBM4 samples to customers. SK hynix’s HBM4 has implemented bandwidth capable of processing more than 2TB of data per second, processing data equivalent to more than 400 full-HD movies (5GB each) in a second, said the company. “HBM competitive landscape, SK hynix has already sampled and secured approval of HBM4 12-high stack memory early Q1’2025 to NVIDIA for its next generation Rubin product line and plans to mass produce HBM4 in 2H 2025,” said Danish Faruqui, CEO, Fab Economics. “Closely following, Micron is pending Nvidia’s tests for its latest HBM4 samples, and Micron plans to mass produce HBM4 in 1H 2026. On the other hand, the last contender, Samsung is struggling with Yield Ramp on HBM4 Technology Development stage, and so has to delay the customer samples milestones to Nvidia and other players while it earlier shared an end of 2025 milestone for mass producing HBM4.” Faruqui noted another key differentiator among SK hynix, Micron, and Samsung: the base die that anchors the 12-high DRAM stack. For the first time, both SK hynix and Samsung have introduced a logic-enabled base die on 3nm and 4nm process technology to enable HBM4 product for efficient and faster product performance via base logic-driven memory management. Both Samsung and SK hynix rely on TSMC for the production of their logic-enabled base die. However, it remains unclear whether Micron is using a logic base die, as the company lacks in-house capability to fabricate at 3nm.

Read More »

Cisco reinvigorates data center, campus, branch networking with AI demands in mind

“We have a number of … enterprise data center customers that have been using bi-directional optics for many generations, and this is the next generation of that feature,” said Bill Gartner, senior vice president and general manager of Cisco’s optical systems and optics business. “The 400G lets customer use their existing fiber infrastructure and reduces fiber count for them so they can use one fiber instead of two, for example,” Gartner said. “What’s really changed in the last year or so is that with AI buildouts, there’s much, much more optics that are part of 400G and 800G, too. For AI infrastructure, the 400G and 800G optics are really the dominant optics going forward,” Gartner said. New AI Pods Taking aim at next-generation interconnected compute infrastructures, Cisco expanded its AI Pod offering with the Nvidia RTX 6000 Pro and Cisco UCS C845A M8 server package. Cisco AI Pods are preconfigured, validated, and optimized infrastructure packages that customers can plug into their data center or edge environments as needed. The Pods include Nvidia AI Enterprise, which features pretrained models and development tools for production-ready AI, and are managed through Cisco Intersight. The Pods are based on Cisco Validated Design principals, which offer customers pre-tested and validated network designs that provide a blueprint for building reliable, scalable, and secure network infrastructures, according to Cisco. Building out the kind of full-scale AI infrastructure compute systems that hyperscalers and enterprises will utilize is a huge opportunity for Cisco, said Daniel Newman, CEO of The Futurum Group. “These are full-scale, full-stack systems that could land in a variety of enterprise and enterprise service application scenarios, which will be a big story for Cisco,” Newman said. Campus networking For the campus, Cisco has added two new programable SiliconOne-based Smart Switches: the C9350 Fixed Access Smart Switches and C9610

Read More »

Qualcomm’s $2.4B Alphawave deal signals bold data center ambitions

Qualcomm says its Oryon CPU and Hexagon NPU processors are “well positioned” to meet growing demand for high-performance, low-power compute as AI inferencing accelerates and more enterprises move to custom CPUs housed in data centers. “Qualcomm’s advanced custom processors are a natural fit for data center workloads,” Qualcomm president and CEO Cristiano Amon said in the press release. Alphawave’s connectivity and compute technologies can work well with the company’s CPU and NPU cores, he noted. The deal is expected to close in the first quarter of 2026. Complementing the ‘great CPU architecture’ Qualcomm has been amassing Client CPUs have been a “big play” for Qualcomm, Moor’s Kimball noted; the company acquired chip design company Nuvia in 2021 for $1.4 billion and has also announced that it will be designing data center CPUs with Saudi AI company Humain. “But there was a lot of data center IP that was equally valuable,” he said. This acquisition of Alphawave will help Qualcomm complement the “great CPU architecture” it acquired from Nuvia with the latest in connectivity tools that link a compute complex with other devices, as well as with chip-to-chip communications, and all of the “very low level architectural goodness” that allows compute cores to deliver “absolute best performance.” “When trying to move data from, say, high bandwidth memory to the CPU, Alphawave provides the IP that helps chip companies like Qualcomm,” Kimball explained. “So you can see why this is such a good complement.”

Read More »

LiquidStack launches cooling system for high density, high-powered data centers

The CDU is serviceable from the front of the unit, with no rear or end access required, allowing the system to be placed against the wall. The skid-mounted system can come with rail and overhead piping pre-installed or shipped as separate cabinets for on-site assembly. The single-phase system has high-efficiency dual pumps designed to protect critical components from leaks and a centralized design with separate pump and control modules reduce both the number of components and complexity. “AI will keep pushing thermal output to new extremes, and data centers need cooling systems that can be easily deployed, managed, and scaled to match heat rejection demands as they rise,” said Joe Capes, CEO of LiquidStack in a statement. “With up to 10MW of cooling capacity at N, N+1, or N+2, the GigaModular is a platform like no other—we designed it to be the only CDU our customers will ever need. It future-proofs design selections for direct-to-chip liquid cooling without traditional limits or boundaries.”

Read More »

Enterprises face data center power design challenges

” Now, with AI, GPUs need data to do a lot of compute and send that back to another GPU. That connection needs to be close together, and that is what’s pushing the density, the chips are more powerful and so on, but the necessity of everything being close together is what’s driving this big revolution,” he said. That revolution in new architecture is new data center designs. Cordovil said that instead of putting the power shelves within the rack, system administrators are putting a sidecar next to those racks and loading the sidecar with the power system, which serves two to four racks. This allows for more compute per rack and lower latency since the data doesn’t have to travel as far. The problem is that 1 mW racks are uncharted territory and no one knows how to manage the power, which is considerable now. ”There’s no user manual that says, hey, just follow this and everything’s going to be all right. You really need to push the boundaries of understanding how to work. You need to start designing something somehow, so that is a challenge to data center designers,” he said. And this brings up another issue: many corporate data centers have power plugs that are like the ones that you have at home, more or less, so they didn’t need to have an advanced electrician certification. “We’re not playing with that power anymore. You need to be very aware of how to connect something. Some of the technicians are going to need to be certified electricians, which is a skills gap in the market that we see in most markets out there,” said Cordovil. A CompTIA A+ certification will teach you the basics of power, but not the advanced skills needed for these increasingly dense racks. Cordovil

Read More »

HPE Nonstop servers target data center, high-throughput applications

HPE has bumped up the size and speed of its fault-tolerant Nonstop Compute servers. There are two new servers – the 8TB, Intel Xeon-based Nonstop Compute NS9 X5 and Nonstop Compute NS5 X5 – aimed at enterprise customers looking to upgrade their transaction processing network infrastructure or support larger application workloads. Like other HPE Nonstop systems, the two new boxes include compute, software, storage, networking and database resources as well as full-system clustering and HPE’s specialized Nonstop operating system. The flagship NS9 X5 features support for dual-fabric HDR200 InfiniBand interconnect, which effectively doubles the interconnect bandwidth between it and other servers compared to the current NS8 X4, according to an HPE blog detailing the new servers. It supports up to 270 networking ports per NS9 X system, can be clustered with up to 16 other NS9 X5s, and can support 25 GbE network connectivity for modern data center integration and high-throughput applications, according to HPE.

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »