Stay Ahead, Stay ONMINE

Nvidia: Latest news and insights

Nvidia: ‘Graphics 3.0’ will drive physical AI productivity August 15, 2025: Nvidia has floated the idea of “Graphics 3.0” with the hope of making AI-generated graphics central to physical productivity. The concept revolves around graphics created by genAI tools. Nvidia say AI-generated graphics could help in training robots to do their jobs in the physical […]

Nvidia: ‘Graphics 3.0’ will drive physical AI productivity

August 15, 2025: Nvidia has floated the idea of “Graphics 3.0” with the hope of making AI-generated graphics central to physical productivity. The concept revolves around graphics created by genAI tools. Nvidia say AI-generated graphics could help in training robots to do their jobs in the physical world or by helping AI assistants automate the creation of equipment and structures.

Nvidia launches Blackwell-powered RTX Pro GPUs for compact AI workstations

August 12, 2025: Nvidia announced two new professional GPUs, the RTX Pro 4000 Small Form Factor (SFF) and the RTX Pro 2000. Built on its Blackwell architecture, Nvidia’s new GPUs aim to deliver powerful AI capabilities in compact desktop and workstation deployments.

Nvidia’s new genAI model helps robots think like humans

August 11, 2025: Nvidia has developed a genAI model to help robots make human-like decisions by analyzing surrounding scenes. The Cosmos Reason model in robots can take in information from video and graphics input, analyze the data, and use its understanding to make decisions.

Nvidia patches critical Triton server bugs that threaten AI model security

August 5, 2025: A surprising attack chain in Nvidia’s Triton Inference Server, starting with a seemingly minor memory-name leak, could allow full remote server takeover without user authentication.

China demands ‘security evidence’ from Nvidia over H20 chip backdoor fears

August 4, 2025: China escalated pressure on Nvidia with the state-controlled People’s Daily publishing an opinion piece titled “Nvidia, how can I trust you?” — a day after regulators summoned company officials over alleged security vulnerabilities in H20 artificial intelligence chips.

Nvidia to restart H20 exports to China, unveils new export-compliant GPU

July 15, 2025: Nvidia will restart H20 AI chip sales to China and release a new GPU model compliant with export rules, a move that could impact global AI hardware strategies for enterprise IT teams. Nvidia has applied for US approval to resume sales and says that the government has indicated licenses will be granted and deliveries could begin soon.

Nvidia GPUs are vulnerable to Rowhammer attacks

July 15, 2025: Nvidia has issued a security reminder to application developers, computer manufacturers, and IT leaders that modern memory chips in graphic processors are potentially susceptible to so-called Rowhammer exploits after Canadian university researchers proved that an Nvidia A6000 GPU could be successfully compromised with a similar attack.

Nvidia hits $4T market cap as AI, high-performance semiconductors hit stride

July 11, 2025: Nvidia became the first publicly traded company to surpass a $4 trillion market capitalization value, 13 months after surpassing the $3 trillion mark. This makes Nvidia the world’s most valuable company ahead of Apple and Microsoft.

New Nvidia technology provides instant answers to encyclopedic-length questions

Jul 8, 2025: Have a question that needs to process an encyclopedia-length dataset? Nvidia says its new technique can answer it instantly. Built leveraging the company’s Blackwell processor’s capabilities, the new “Helix Parallelism” method allows AI agents to process millions of words — think encyclopedia-length — and support up to 32x more users at a time.

Nvidia doubles down on GPUs as a service

July 8, 2025: Nvidia’s recent initiative to dive deeper into the GPU-as-a-service (GPUaaS) model marks a significant and strategic shift that reflects an evolving landscape within the cloud computing market. 

Nvidia, Perplexity to partner with EU and Middle East AI firms to build sovereign LLMs

June 12, 2025: Nvidia and AI search firm Perplexity said they are joining hands with model builders and cloud providers across Europe and the Middle East to refine sovereign large-language models (LLMs) and accelerate enterprise AI uptake in local industries.

Nvidia: ‘Sovereign AI’ will change digital work

June 11, 2025: Nvidia executives think sovereign AI has the potential to change digital work as generative AI (genAI) aligns with national priorities and local regulations.

AWS cuts prices of some EC2 Nvidia GPU-accelerated instances

June 9, 2025: AWS has reduced the prices of some of its Nvidia GPU-accelerated instances to attract more AI workloads while competing with rivals, such as Microsoft and Google, as demand for GPUs and the cost of securing them continues to grow.

Nvidia aims to bring AI to wireless

June 6, 2025: Nvidia hopes to maximize RAN infrastructure use (traditional networks average a low 30% to 35%), use AI to rewrite the air interface, and enhance performance and efficiency through radio signal processing. The longer-term goal is to seamlessly process AI traffic at the network edge to create new monetization opportunities for service providers.

Oracle to spend $40B on Nvidia chips for OpenAI data center in Texas

May 26, 2025: Oracle is reportedly spending about $40 billion on Nvidia’s high-performance computer chips to power OpenAI’s new data center in Texas, marking a pivotal shift in the AI infrastructure landscape that has significant implications for enterprise IT strategies.

Nvidia eyes China rebound with stripped-down AI chip tailored to export limits

May 26, 2025: Nvidia plans to launch a lower-cost AI chip for China in June, aiming to protect market share under the US export controls and signal a broader shift toward affordable, segmented products that could impact global enterprise AI spending.

Nvidia introduces ‘ridesharing for AI’ with DGX Cloud Lepton

May 19, 2025: Nvidia introduced DGX Cloud Lepton, an AI-centric cloud software program that makes it easier for AI factories to rent out their hardware to developers who wish to access performant compute globally.

May 19, 2025: Nvidia kicked off the Computex systems hardware tradeshow with the news it has opened the NVLink interconnect technology to the competition with the introduction of NVLink Fusion. NVLink is a high-speed interconnect born out of its Mellanox networking group which lets multiple GPUs in a system or rack share compute and memory resources, thus making many GPUs appear to the system as a single processor.

AMD, Nvidia partner with Saudi startup to build multi-billion dollar AI service centers

May 15, 2025: As part of the avalanche of business deals coming from President Trump’s Middle East tour, both AMD and Nvidia have struck multi-billion dollar deals with an emerging Saudi AI firm. The deals served as the coming out party for Humain, a state-backed artificial intelligence (AI) company that operates under the Kingdom’s Public Investment Fund (PIF) and is chaired by Crown Prince Mohammed bin Salman. 

Nvidia, ServiceNow engineer open-source model to create AI agents

May 6, 2025: Nvidia and ServiceNow have created an AI model that can help companies create learning AI agents to automate corporate workloads..The open-source Apriel model, available generally in the second quarter on HuggingFace, will help create AI agents that can make decisions around IT, human resources and customer-service functions.

Nvidia AI supercluster targets agents, reasoning models on Oracle Cloud

April 29, 2025: The move marks the first wave of liquid-cooled Nvidia GB200 NVL72 racks in OCI data centers, involving thousands of Nvidia Grace CPUs and Blackwell GPUs. 

Nvidia says NeMo microservices now generally available

April 23, 2025: Nvidia announced the general availability of neural module (NeMo) microservices, a modular platform for building and customizing gen AI models and AI agents.NeMo microservices integrate with partner platforms to provide features including prompt tuning, supervised fine-tuning, and knowledge retrieval tools.

Nvidia expects ban on chip exports to China to cost $5.5B

April 16, 2025: Nvidia now expects new US government restrictions on exports of its H20 chip to China will cost the company as much as $5.5 billion.

Incomplete patching leaves Nvidia, Docker exposed to DOS attacks

April 15, 2025: A critical race condition bug affecting the Nvidia Container Toolkit, which received a fix in September, might still be open to attacks owing to incomplete patching.

Nvidia lays out plans to build AI supercomputers in the US

April 14, 2025: There was mixed reaction from industry analysts over an announcement that Nvidia plans to produce AI supercomputers entirely in the US. The company said in a blog post that, together with its manufacturing partners, it has commissioned more than one million square feet (92,900 square meters) of manufacturing space to build and test Nvidia Blackwell chips in Arizona and AI supercomputers in Texas.

Potential Nvidia chip shortage looms as Chinese customers rush to beat US sales ban

April 2, 2025: The AI chip shortage could become even more dire as Chinese customers are purportedly looking to hoard Nvidia chips ahead of a proposed US sales ban. According to inside sources, Chinese companies including ByteDance, Alibaba Group, and Tencent Holdings have ordered at least $16 billion worth of Nvidia’s H20 server chips for running AI workloads in just the first three months of this year.

Nvidia’s Blackwell raises the bar with new MLPerf Inference V5.0 results

April 2, 2025: Nvidia released a set of MLPerf Inference V5.0 benchmark results for its Blackwell GPU, the successor to Hopper, saying that its GB200 NVL72 system, a rack-scale offering designed for AI reasoning, set a series of performance records.

5 big takeaways from Nvidia GTC

March 25, 2025: Now that the dust has settled from Nvidia’s GTC 2025, a few industry experts weighed in on some core big picture developments from the conference. Here are five of their top observations.

Nvidia wants to be a one-stop enterprise technology shop

March 24, 2025: After last week’s Nvidia GTC 2025 event, a new, fuller picture of the vendor emerged. Analysts agree that Nvidia is not just a graphics chip provider anymore. It’s a full-stack solution provider, and GPUs are just one of many parts.

Nvidia launches AgentIQ toolkit to connect disparate AI agents

March 21, 2025: As enterprises look to adopt agentic AI to boost the efficiency of their applications, Nvidia introduced a new open-source software library — AgentIQ toolkit — to help developers connect disparate agents and agent frameworks. The toolkit, according to Nvidia, packs in a variety of tools, including ones to weave in RAG, search, and conversational UI into agentic AI applications.

Nvidia launches research center to accelerate quantum computing breakthrough

March 21, 2025: In a move to help accelerate the timeline for practical, real-world quantum applications, Nvidia is establishing the Nvidia Accelerated Quantum Research Center. “Quantum computing will augment AI supercomputers to tackle some of the world’s most important problems,” Nvidia CEO Jensen Huang said.

Nvidia, xAI and two energy giants join genAI infrastructure initiative

March 19, 2025: An industry generative artificial intelligence (genAI) alliance, the AI Infrastructure Partnership (AIP), on Wednesday announced that xAI, Nvidia, GE Vernova, and NextEra Energy were joining BlackRock, Microsoft, and Global Infrastructure Partners as members.

IBM broadens access to Nvidia technology for enterprise AI

March 19, 2025: New collaborations between IBM and Nvidia have yielded a content-aware storage capability for IBM’s hybrid cloud infrastructure, expanded integration between watsonx and Nvidia NIM, and AI services from IBM Consulting that use Nvidia Blueprints.

Nvidia’s silicon photonics switches bring better power efficiency to AI data centers

March 19, 2025: Amid the flood of news from Nvidia’s annual GTC event, one item stood out. Nvidia introduced new silicon photonics network switches that integrate network optics into the switch using a technique called co-packaged optics (CPO), replacing traditional external pluggable transceivers. While Nvidia alluded to its new switches providing a cost savings, the primary benefit is to reduce power consumption with an improvement in network resiliency.

What is Nvidia Dynamo and why it matters to enterprises?

March 19, 2025: Chipmaker Nvidia has released a new open-source inferencing software — Dynamo, at its GTC 2025 conference, that will allow enterprises to increase throughput and reduce cost while using large language models on Nvidia GPUs.

Nvidia, xAI and two energy giants join genAI infrastructure initiative

March 19, 2025:  AI Infrastructure Partnership (AIP) announced that xAI, Nvidia, GE Vernova, and NextEra Energy joined the AIP. But given that no financial commitments or any other details were released, will it make a difference?

HPE, Nvidia broaden AI infrastructure lineup

March 19, 2025: HPE news from Nvidia GTC includes a new Private Cloud AI developer kit, Nvidia AI blueprints, GPU optimization capabilities, and servers built with Nvidia Blackwell Ultra and Blackwell architecture.

Cisco, Nvidia team to deliver secure AI factory infrastructure

March 18, 2025: Cisco and Nvidia have expanded their partnership to create their most advanced AI architecture package to date, designed to promote secure enterprise AI networking.

Nvidia’s ‘hard pivot’ to AI reasoning bolsters Llama models for agentic AI

March 18, 2025: The company has post-trained its new Llama Nemotron family of reasoning models to improve multistep math, coding, reasoning, and complex decision-making. The enhancements aim to provide developers and enterprises with a business-ready foundation for creating AI agents that can work independently or as part of connected teams.

Nvidia details its GPU, CPU, and system roadmap for the next three years

March 18, 2025: Nvidia CEO Jensen Huang shared previously unreleased specifications for its Rubin graphics processing unit (GPU), due in 2026, the Rubin Ultra coming in 2027, and announced the addition of a new GPU called Feynman to the mix for 2028.

Oracle, Nvidia partner to add AI software into OCI services

March 18, 2025: Nvidia’s AI Enterprise stack will be available natively through the OCI Console and will be available anywhere in OCI’s distributed cloud while providing enterprises access to over 160 AI tools for training and inference, including NIM microservices, the companies said in a joint statement at Nvidia’s annual GTC conference.

Nvidia GTC 2025: What to expect from the AI leader

March 3, 2025: Last year, Nvidia’s GTC 2024 grabbed headlines with the introduction of the Blackwell architecture and the DGX systems powered by it. With Nvidia GTC 2025 right around the corner, the tech world is eager to see what Nvidia – and its partners and competitors – will unveil next. 

Cisco, Nvidia expand AI partnership to include Silicon One technology

February 25, 2025; Cisco and Nvidia have expanded their collaboration to support enterprise AI implementations by tying Cisco’s Silicon One technology to Nvidia’s Ethernet networking platform. The extended agreement is designed to offer customers yet another way to support AI workloads across the data center and strengthens both companies’ strategies to expand the role of Ethernet networking for AI in the enterprise.

Nvidia forges healthcare partnerships to advance AI-driven genomics, drug discovery

February 14, 2025: Through new partnerships with industry leaders, Nvidia aims to advance practical use cases for AI in healthcare and life sciences. It’s a logical move: Healthcare has the most significant upside, particularly in patient care, among all the industries applicable to AI. 

Nvidia partners with cybersecurity vendors for real-time monitoring

February 12, 2025: Nvidia partnered with leading cybersecurity firms to provide real-time security protection using its accelerator and networking hardware in combination with its AI software. Under the agreement, Nvidia will provide integration of its BlueField and Morpheus hardware with cyber defenses software from Armis, Check Point Software Technologies, CrowdStrike, Deloitte and World Wide Technology .

Nvidia claims near 50% boost in AI storage speed

February 7, 2025: Nvidia is touting a near 50% improvement in storage read bandwidth thanks to intelligence in its Spectrum-X Ethernet networking equipment, according to the vendor’s technical blog post. Spectrum-X is a combination of the company’s Spectrum-4 Ethernet switch and BlueField-3 SuperNIC smart networking card, which supports RoCE v2 for remote direct memory access (RDMA) over Converged Ethernet.

Nvidia unveils preview of DeepSeek-R1 NIM microservice

February 3, 2025: The chipmaker stock plummeted 17% after Chinese AI developer DeepSeek unveiled its DeepSeek-R1 LLM. Last week, Nvidia announced the DeepSeek-R1 model is now available as a preview Nvidia inference microservice (NIM) on build.nvidia.com.

Nvidia unveils preview of DeepSeek-R1 NIM microservice

January 31, 2025: Nvidia stock plummeted 17% after Chinese AI developer, DeepSeek, unveiled its DeepSeek-R1 LLM. Later the same week, the chipmaker turned around and announced the DeepSeek-R1 model is available as a preview Nvidia inference microservice (NIM) on build.nvidia.com.

Nvidia intros new guardrail microservices for agentic AI

January 16, 2025: Nvidia added new Nvidia inference microservices (NIMs) for AI guardrails to its Nvidia NeMo Guardrails software tools. The new microservices aim to help enterprises improve accuracy, security, and control of agentic AI applications, addressing a key reservation IT leaders have about adopting the technology.

Nvidia year in review

January 10, 2025: Last year was Nvidia’s year. Its command of mindshare and market share was unequaled among tech vendors. Here’s a recap of some of the key Nvidia events of 2024 that highlight just how powerful the world’s most dominant chip player is.

Nvidia launches blueprints to help jumpstart AI projects

January 8, 2025: Nvidia recently issued designs for AI factories after hyping up the idea for several months. Now it has come out with AI blueprints, essentially prebuilt templates that give developers a jump start on creating AI systems.

Nvidia’s Project DIGITS puts AI supercomputing chips on the desktop

January 6, 2025: Nvidia is readying a tiny desktop device called Project DIGITS, a “personal AI supercomputer” with a lightweight version of the Grace Blackwell platform found in its most powerful servers; it’s aimed at data scientists, researchers, and students who will be able to prototype, tune, and run large genAI models.

Nvidia unveils generative physical AI platform, agentic AI advances at CES

January 6, 2025: At CES in Las Vegas, Nvidia trumpeted a slew of AI announcements, with an emphasis on generative physical AI that promises a new revolution in factory and warehouse automation. “AI requires us to build an entirely new computing stack to build AI factories, accelerated computing at data center scale,” Rev Lebaredian, vice president of omniverse and simulation technology at Nvidia.

Verizon, Nvidia team up for enterprise AI networking

December 30, 2024: Verizon and Nvidia partnered to build AI services for enterprises that run workloads over Verizon’s 5G private network. The new offering, 5G Private Network with Enterprise AI, will run a range of AI applications and workloads over Verizon’s private 5G network with Mobile Edge Compute (MEC). MEC is a colocated infrastructure that is a part of Verizon’s public wireless network, bringing compute and storage closer to devices and endpoints for ultra-low latency.

Nvidia’s Run:ai acquisition waved through by EU

December 20, 2024: Nvidia will face no objections to its plan to acquire Israeli AI orchestration software vendor Run:ai Labs in Europe, after the European Commission gave the deal its approval today. But Nvidia may not be out of the woods yet. Competition authorities in other markets are closely examining the company’s acquisition strategy.

China launches anti-monopoly probe into Nvidia amid rising US-China chip tensions

December 10, 2024: China has initiated an investigation into Nvidia over alleged violations of the country’s anti-monopoly laws, signaling a potential escalation in the ongoing tech and trade tensions between Beijing and Washington.

Nvidia Blackwell chips face serious heating issues

November 18, 2024: Nvidia’s next-generation Blackwell data center processors have significant problems with overheating when installed in high-capacity server racks, forcing redesigns of the racks themselves, according to a report by The Information. These issues have reportedly led to design changes, meaning delays in shipping product and raising concern that its biggest customers, including Google, Meta, and Microsoft, will be able to deploy Blackwell servers according to their schedules.

Nvidia to power India’s AI factories with tens of thousands of AI chips

October 24, 2024: Nvidia plans to deploy thousands of Hopper GPUs in India to create AI factories and collaborate with Reliance Industries to develop AI infrastructure.. Yotta Data Services, Tata Communications, E2E Networks, and Netweb will lead the AI factories — large-scale data centers for producing AI. Nvidia added that the expansion will provide nearly 180 exaflops of computing power.

Nvidia contributes Blackwell rack design to Open Compute Project

October 15, 2024: Nvidia contributed to the Open Compute Project its Blackwell GB200 NVL72 electro-mechanical designs – including the rack architecture, compute and switch tray mechanicals, liquid cooling and thermal environment specifications, and Nvidia NVLink cable cartridge volumetrics –.

As global AI energy usage mounts, Nvidia claims efficiency gains of up to 100,000X

October 08, 2024: As concerns over AI energy consumption ratchet up, chip maker Nvidia is defending what it calls a steadfast commitment to sustainability. The company reports that its GPUs have experienced a 2,000X reduction in energy use over the last 10 years in training and a 100,000X energy reduction over that same time in generating tokens.

Accenture forms new Nvidia business group focused on agentic AI adoption

October 4, 2024: Accenture and Nvidia announced an expanded partnership focused on helping customers rapidly scale AI adoption. Accenture said the new group will use Accenture’s AI Refinery platform — built on the Nvidia AI stack, including Nvidia AI Foundry, Nvidia AI Enterprise, and Nvidia Omniverse — to help clients create a foundation for use of agentic AI.

IBM expands Nvidia GPU options for cloud customers

October 1, 2024: IBM expanded access to Nvidia GPUs on IBM Cloud to help enterprise customers advance their AI implementations, including large language model (LLM) training. IBM Cloud users can now access Nvidia H100 Tensor Core GPU instances in virtual private cloud and managed Red Hat OpenShift environments.

Oracle to offer 131,072 Nvidia Blackwell GPUs via its cloud

September 12, 2024: Oracle started taking pre-orders for 131,072 Nvidia Blackwell GPUs in the cloud via its Oracle Cloud Infrastructure (OCI) Supercluster to aid large language model (LLM) training and other use cases, the company announced at the CloudWorld 2024 conference.  The launch of an offering that provides these many Blackwell GPUs, also known as Grace Blackwell (GB) 200, is significant as enterprises globally are faced with the unavailability of high-bandwidth memory (HBM) — a key component used in making GPUs.

Why is the DOJ investigating Nvidia?

September 11, 2024: After a stock sell-off following its quarterly earnings report, Nvidia’s pain was aggravated by news that the Department of Justice is escalating its investigation into the company for anticompetitive practices. According to a Bloomberg report, the DOJ sent a subpoena to Nvidia as part of a probe into alleged antitrust practices.

Cisco, HPE, Dell announce support for Nvidia’s pretrained AI workflows

September 4, 2024: Cisco, HPE, and Dell are using Nvidia’s new AI microservices blueprints to help enterprises streamline the deployment of generative AI applications. Nvidia’s announced its NIM Agent Blueprints, a catalogue of pretrained, customizable AI workflows that are designed to provide a jump-start for developers creating AI applications. NIM Agent Blueprints target a number of use cases, including customer service, virtual screening for computer-aided drug discovery, and a multimodal PDF data extraction workflow for retrieval-augmented generation (RAG) that can ingest vast quantities of data.

Nvidia reportedly trained AI models on YouTube data

August 4, 2024: Nvidia scraped huge amounts of data from YouTube to train its AI models, even though neither Youtube nor individual YouTube channels approved the move, according to leaked documents. Among other things, Nvidia reportedly used the YouTube data to train its deep learning model Cosmos, an algorithm for automated driving, a human-like AI avatar, and Omniverse, a tool for building 3D worlds.

Can Intel’s new chips compete with Nvidia in the AI universe?

June 9, 2024: Intel is aiming its next-generation X86 processors at AI tasks, even though the chips won’t actually run AI workloads themselves.mAt Computex, Intel announced its Xeon 6 processor line, talking up what it calls Efficient-cores (E-cores) that it said will deliver up to 4.2 times the performance of Xeon 5 processors. The first Xeon 6 CPU is the Sierra Forest version (6700 series) a more performance-oriented line, Granite Rapids with Performance cores (P-cores or 6900 series), will be released next quarter.

Everyone but Nvidia joins forces for new AI interconnect

May 30, 2024: A clear sign of Nvidia’s dominance is when Intel and AMD link arms to deliver a competing product. That’s what happened when AMD and Intel – along with Broadcom, Cisco, Google, Hewlett Packard Enterprise, Meta and Microsoft – formed the Ultra Accelerator Link (UALink) Promoter Group to develop high-speed interconnections between AI processors.

Nvidia to build supercomputer for federal AI research

May 15, 2024: The U.S. government will use an Nvidia DGX SuperPOD to provide researchers and developers access to much more computing power than they have had in the past to produce generative AI advances in areas such as climate science, healthcare and cybersecurity.

Nvidia, Google Cloud team to boost AI startups

April 11, 2024: Alphabet’s Google Cloud unveiled a slew of new products and services at Google Cloud Next 2024, among them a program to help startups and small businesses build generative AI applications and services. The initiative brings together the Nvidia Inception program for startups and the Google for Startups Cloud Program.

Nvidia GTC 2024 wrap-up: Blackwell not the only big news

March 29, 2024: Nvidia’s GDC is in our rearview mirror, and there was plenty of news beyond the major announcement of the Blackwell architecture and the massive new DGX systems powered by it. Here’s a rundown of some of the announcements you might have missed.

Nvidia expands partnership with hyperscalers to boost AI training and development

March 19, 2024: Nvidia extended its existing partnerships with hyperscalers Amazon Web Services (AWS), Google Cloud Platform, Microsoft Azure, and Oracle Cloud Infrastructure, to make available its latest GPUs and foundational large language models and to integrate its software across their platforms.

Nvidia launches Blackwell GPU architecture

March 18, 2024: Nvidia kicked off its GTC 2024 conference with the formal launch of Blackwell, its next-generation GPU architecture due at the end of the year. Blackwell uses a chiplet design, to a point. Whereas AMD’s designs have several chiplets, Blackwell has two very large dies that are tied together as one GPU with a high-speed interlink that operates at 10 terabytes per second, according to Ian Buck, vice president of HPC at Nvidia.

Cisco, Nvidia target secure AI with expanded partnership

February 9, 2024: Cisco and Nvidia expanded their partnership to offer integrated software and networking hardware that promises to help customers more easily spin up infrastructure to support AI applications. The agreement deepens both companies’ strategy to expand the role of Ethernet networking for AI workloads in the enterprise. It also gives both companies access to each other’s sales and support systems.

Nvidia and Equinix partner for AI data center infrastructure

January 9, 2024: Nvidia partnered with data center giant Equinix to offer what the vendors are calling Equinix Private AI with Nvidia DGX, a turnkey solution for companies that are looking to get into the generative AI game but lack the data center infrastructure and expertise to do it.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

F5 brings new visibility and AI controls to Big-IP, NGINX

The demand came from a gap that general-purpose observability tools were not filling. Customers running tools like Datadog and New Relic told F5 they needed something different.  F5 Insight pulls from technology acquired through the Threat Stack and Fletch acquisitions and runs on F5’s AI data fabric. It includes an

Read More »

Tech layoffs surpass 45,000 in early 2026

Layoffs spread across tech sectors Beyond Amazon, Meta, and Block, several technology vendors and platform companies have also announced sizable layoffs this year. According to the RationalFX report: Semiconductor and electronics company ams OSRAM has announced 2,000 layoffs. Telecommunications vendor Ericsson has announced 1,900 job cuts. Semiconductor equipment manufacturer ASML

Read More »

Eridu exits stealth with $200M to rebuild AI networking

That gap is not static. Promode Nedungadi, Chief Technology Officer, said the architectural and algorithmic trends driving AI are making the network problem harder, not easier. Techniques like mixture-of-experts models and the disaggregation of inference into separate prefill and decode stages all require more data movement. “Every one of those

Read More »

United States to Release 172 Million Barrels of Oil From the Strategic Petroleum Reserve

WASHINGTON—U.S. Secretary of Energy Chris Wright released the following statement regarding the International Energy Agency (IEA) and the U.S. Strategic Petroleum Reserve (SPR): “Earlier today, 32 member nations of the International Energy Agency unanimously agreed to President Trump’s request to lower energy prices with a coordinated release of 400 million barrels of oil and refined products from their respective reserves.  “As part of this effort, President Trump authorized the Department of Energy to release 172 million barrels from the Strategic Petroleum Reserve, beginning next week. This will take approximately 120 days to deliver based on planned discharge rates.  “President Trump promised to protect America’s energy security by managing the Strategic Petroleum Reserve responsibly and this action demonstrates his commitment to that promise. Unlike the previous administration, which left America’s oil reserves drained and damaged, the United States has arranged to more than replace these strategic reserves with approximately 200 million barrels within the next year—20% more barrels than will be drawn down—and at no cost to the taxpayer.  “For 47 years, Iran and its terrorist proxies have been intent on killing Americans. They have manipulated and threatened the energy security of America and its allies. Under President Trump, those days are coming to an end.  “Rest assured, America’s energy security is as strong as ever.”                                                                                         ###

Read More »

Occidental Petroleum, 1PointFive STRATOS DAC plant nears startup in Texas Permian basin

Occidental Petroleum Corp. and its subsidiary 1PointFive expect Phase 1 of the STRATOS direct air capture (DAC) plant in Texas’ Permian basin to come online in this year’s second quarter. In a post to LinkedIn, 1PointFive said Phase 1 “is in the final stage of startup” and that Phase 2, which incorporates learnings from research and development and Phase 1 construction activities, “will also begin commissioning in Q2, with operational ramp-up continuing through the rest of the year.” Once fully operational, STRATOS is designed to capture up to 500,000 tonnes/year (tpy) of CO2. As part of the US Environmental Protection Agency (EPA) Class VI permitting process and approval, it was reported that STRATOS is expected to include three wells to store about 722,000 tpy of CO2 in saline formations at a depth of about 4,400 ft. The company said a few activities before start-up remain, including ramping up remaining pellet reactors, completing calciner final commissioning in parallel, and beginning CO2 injection. Start-up milestones achieved include: Completed wet commissioning with water circulation. Received Class VI permits to sequester CO2. Ran CO2 compression system at design pressure. Added potassium hydroxide (KOH) to capture CO2 from the atmosphere. Building pellet inventory. Burners tested on calciner.  

Read More »

Brava Energia weighs Phase 3 at Atlanta to extend production plateau

Just 2 months after bringing its flagship Atlanta field onstream with the new FPSO Atlanta, Brazil’s independent operator Brava Energia SA is evaluating a potential third development phase that could add roughly 25 million bbl of reserves and help sustain peak production longer than originally planned. The Phase 3 project, still at an early technical and economic evaluation stage, focuses on the Atlanta Nordeste area; a separate, shallower reservoir discovered in 2006 by Shell’s 9-SHEL-19D-RJS well. According to André Fagundes, vice-president of research (Brazil) at Welligence Energy Analytics, Phase 2 has four wells still to be developed: two expected in 2027 and two in 2029. Phase 3 would involve drilling two additional wells in 2031, bringing total development to 12 producing wells. Until recently, full-field development was understood to comprise 10 wells, but Brava has since updated guidance to reflect a 12-well development concept. Atlanta field upside The primary objective is clear. “We believe its main objective is to extend the production plateau,” Fagundes said. Welligence estimates incremental recovery could reach 25 MMbbl, increasing the field’s overall recovery factor by roughly 1.5%. Lying outside Atlanta’s main Cretaceous reservoir, Atlanta Nordeste represents a genuine upside opportunity, Fagundes explained. The field benefits from strong natural aquifer support, and no water or gas injection is anticipated. Water-handling constraints that affected early production using the Petrojarl I—limited to 11,500 b/d of water treatment—are no longer a bottleneck. FPSO Atlanta can process up to 140,000 b/d of water. Reservoir performance to date has been solid, albeit with difficulties. Recurrent electric submersible pump (ESP) failures and processing limits on the previous FPSO complicated full validation of original reservoir models. With the new 50,000-b/d FPSO in operation since late 2024, reservoir deliverability has become the main constraint. Phase 3 wells would also use ESPs and require additional subsea

Read More »

California Resources eyes ‘measured’ capex ramp on way to 12% production growth thanks to Berry buy

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: var(–color-primary-main); } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style { font-size: 0.75rem; opacity: .6; } #onetrust-pc-sdk [id*=btn-handler], #onetrust-pc-sdk [class*=btn-handler] { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-policy a, #onetrust-pc-sdk a, #ot-pc-content a { color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-pc-sdk .ot-active-menu { border-color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-accept-btn-handler, #onetrust-banner-sdk #onetrust-reject-all-handler, #onetrust-consent-sdk #onetrust-pc-btn-handler.cookie-setting-link { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-consent-sdk .onetrust-pc-btn-handler { color: #c19a06 !important; border-color: #c19a06 !important; } The leaders of California Resources Corp., Long Beach, plan to have the company’s total production average 152,000-157,000 boe/d in 2026, with each quarter expected to be in that range. That output would equate to an increase of more than 12% from the operator’s 137,000 boe/d during fourth-quarter 2025, due mostly to the mid-December acquisition of Berry Corp. Fourth-quarter results folded in 14 days of Berry production and included 109,000 b/d of oil, with the company’s assets in the San Joaquin and Los Angeles basins accounting for 99,000 b/d of that total. The company dilled 31 new wells during the quarter and 76 in all of 2025—all in the San Joaquin—but that number will grow significantly to about 260 this year as state officials have resumed issuing permits following the passage last fall of a bill focused on Kern County production. Speaking to analysts after CRC reported fourth-quarter net income of $12 million on $924 million in revenues, president and chief executive officer Francisco Leon and chief financial officer Clio Crespy said the goal is to manage 2026 output decline to roughly 0.5% per quarter while operating four rigs and

Read More »

Petro-Victory Energy spuds São João well in Brazil

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: var(–color-primary-main); } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style { font-size: 0.75rem; opacity: .6; } #onetrust-pc-sdk [id*=btn-handler], #onetrust-pc-sdk [class*=btn-handler] { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-policy a, #onetrust-pc-sdk a, #ot-pc-content a { color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-pc-sdk .ot-active-menu { border-color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-accept-btn-handler, #onetrust-banner-sdk #onetrust-reject-all-handler, #onetrust-consent-sdk #onetrust-pc-btn-handler.cookie-setting-link { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-consent-sdk .onetrust-pc-btn-handler { color: #c19a06 !important; border-color: #c19a06 !important; } Petro-Victory Energy Corp. has spudded the SJ‑12 well at São João field in Barreirinhas basin, on the Brazilian equatorial margin, Maranhão.  Drilling and testing SJ‑12 is aimed at proving enough gas can be produced to sell locally. The well forms part of the single non‑associated gas well commitment under a memorandum of understanding signed in 2024 with Enava. São João contains 50.1 bcf (1.4 billion cu m) non‑associated gas resources. Petro‑Victory 100% owns and operates São João field.

Read More »

Opinion Poll: Strait of Hormuz disruptions

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: var(–color-primary-main); } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style { font-size: 0.75rem; opacity: .6; } #onetrust-pc-sdk [id*=btn-handler], #onetrust-pc-sdk [class*=btn-handler] { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-policy a, #onetrust-pc-sdk a, #ot-pc-content a { color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-pc-sdk .ot-active-menu { border-color: #c19a06 !important; } #onetrust-consent-sdk #onetrust-accept-btn-handler, #onetrust-banner-sdk #onetrust-reject-all-handler, #onetrust-consent-sdk #onetrust-pc-btn-handler.cookie-setting-link { background-color: #c19a06 !important; border-color: #c19a06 !important; } #onetrust-consent-sdk .onetrust-pc-btn-handler { color: #c19a06 !important; border-color: #c19a06 !important; } 388041610 © Ahmad Efendi | Dreamstime.com US, Israel, and Iran flags <!–> ]–> <!–> –> Oil & Gas Journal wants to hear your thoughts about how the collaborative strike on Iran by the US and Israel and disruptions through the Strait of Hormuz may impact oil prices.  

Read More »

Datalec targets rapid infrastructure deployment with new modular data centers

“We are engineering the data center with a new lens bringing pre-engineered system designs that are flexible and adaptable that enables a tailored solution for clients,” said John Lever, director of modular solutions at Datalec. The systems are flexible enough that these solutions cater for all types of data center, from standard server technology to AI and high-density compute. Datalec also provides “bolt-on” solutions, including a ‘digital wrapper’ including digital twinning and lifecycle and global support, Lever says. Another way Datalec says it differentiates from competing modular designs is a larger share of work is done offsite in a controlled manufacturing environment, which cuts onsite construction time, improves safety and limits disruption to live facilities, Lever says. The company competes with other modular data center vendors including Schneider Electric, Vertiv, Flex many others. DPI’s says its services are aimed at colocation providers, hyperscale and AI infrastructure teams, and large enterprises that need to add capacity quickly, safely and cost effectively across multiple regions.

Read More »

Study finds significant savings from direct current power for AI workloads

The result is a 50% to 80% reduction in copper usage, due to fewer conductors and less parallel cabling, and an 8% to 12% reduction in annual energy-related OpEx through lower conversion and distribution losses. By reducing conductor count, cabling, and redundant power components, 800VDC enables meaningful savings at both build-out and operational stages. AI-first facilities can see a $4 million to $8 million in CapEx savings per 10 MW build by reducing upstream AC. For a one-gigawatt data center, you’re saving a couple million pounds of copper wire, he said. Burke says an all-DC data center is best done with a whole new facility rather than retrofitting old facilities. “[DC] is going to be in a lot of greenfield data centers that are going to be built, and data centers that are going to go to higher compute power are also going to DC,” he said. He did recommend all-DC retrofits for existing data centers that are going to employ high power computing with GPUs. Enteligent’s unnamed and as yet unreleased product is a converter that takes 800 volts and partitions it to 50 volts for the computing servers. The company will provide a new power supply, power shelf that converts 800 volts DC to 50 volts DC much more efficiently than any current power supplies. Burke said the company is doing NDA level testing and pilot programs now with its product, but it will be making a formal announcement within the next few weeks. There are a number of players in the DC arena focusing on different parts of the power supply market including Vertiv, Rutherford, Siemens, Eaton and many more.

Read More »

Cisco blends Splunk analytics, security with core data center management

With the integration, data center teams can gather and act on events, alarms, health scores, and inventory through open APIs, Cisco stated. It also offers pre-built and customizable dashboards for inventory, health, fabric state, anomalies, and advisories as well as correlates telemetry across fabrics and technology tiers for actionable insights, according to Cisco. “This isn’t just another connector or API call. This is an embedded, architectural integration designed to transform how you monitor, troubleshoot, and secure your data center fabric. By bringing the power of Splunk directly into the Data Center Networking environment, we are enabling teams to solve complex problems faster, maintain strict data sovereignty, and dramatically reduce operational costs,” wrote Usha Andra is a senior product marketing leader and Anant Shah, senior product manager, both with Cisco Data Center Networking in a blog about the integration.  “Traditionally, network monitoring involves a trade-off. You either send massive amounts of raw logs to a centralized data lake, incurring high ingress and storage costs. Or you rely on sampled data that misses critical microbursts and anomalies,” Andra and Shah wrote.  “Native Splunk integration changes the paradigm by running Splunk capabilities directly within the Cisco Nexus Dashboard. This allows for the streaming of high-fidelity telemetry, including anomalies, advisories, and audit logs, directly to Splunk analytics.”

Read More »

Execution, Power, and Public Trust: Rich Miller on 2026’s Data Center Reality and Why He Built Data Center Richness

DCF founder Rich Miller has spent much of his career explaining how the data center industry works. Now, with his latest venture, Data Center Richness, he’s also examining how the industry learns. That thread provided the opening for the latest episode of The DCF Show Podcast, where Miller joined present Data Center Frontier Editor in Chief Matt Vincent and Senior Editor David Chernicoff for a wide-ranging discussion that ultimately landed on a simple conclusion: after two years of unprecedented AI-driven announcements, 2026 will be the year reality asserts itself. Projects will either get built, or they won’t. Power will either materialize, or it won’t. Communities will either accept data center expansion – or they’ll stop it. In other words, the industry is entering its execution phase. Why Data Center Richness Matters Now Miller launched Data Center Richness as both a podcast and a Substack publication, an effort to experiment with formats and better understand how professionals now consume industry information. Podcasts have become a primary way many practitioners follow the business, while YouTube’s discovery advantages increasingly make video versions essential. At the same time, Miller remains committed to written analysis, using Substack as a venue for deeper dives and format experimentation. One example is his weekly newsletter distilling key industry developments into just a handful of essential links rather than overwhelming readers with volume. The approach reflects a broader recognition: the pace of change has accelerated so much that clarity matters more than quantity. The topic of how people learn about data centers isn’t separate from the industry’s trajectory; it’s becoming part of it. Public perception, regulatory scrutiny, and investor expectations are now shaped by how stories are told as much as by how facilities are built. That context sets the stage for the conversation’s core theme. Execution Defines 2026 After

Read More »

Nomads at the Frontier: PTC 2026 Signals the Digital Infrastructure Industry’s Moment of Execution

Each January, the Pacific Telecommunications Council conference serves as a barometer for where digital infrastructure is headed next. And according to Nomad Futurist founders Nabeel Mahmood and Phillip Koblence, the message from PTC 2026 was unmistakable: The industry has moved beyond hype. The hard work has begun. In the latest episode of The DCF Show Podcast, part of our ongoing ‘Nomads at the Frontier’ series, Mahmood and Koblence joined Data Center Frontier to unpack the tone shift emerging across the AI and data center ecosystem. Attendance continues to grow year over year. Conversations remain energetic. But the character of those conversations has changed. As Mahmood put it: “The hype that the market started to see is actually resulting a bit more into actions now, and those conversations are resulting into some good progress.” The difference from prior years? Less speculation. More execution. From Data Center Cowboys to Real Deployments Koblence offered perhaps the sharpest contrast between PTC conversations in 2024 and those in 2026. Two years ago, many projects felt speculative. Today, developers are arriving with secured power, customers, and construction underway. “If 2024’s PTC was data center cowboys — sites that in someone’s mind could be a data center — this year was: show me the money, show me the power, give me accurate timelines.” In other words, the market is no longer rewarding hypothetical capacity. It is demanding delivered capacity. Operators now speak in terms of deployments already underway, not aspirational campuses still waiting on permits and power commitments. And behind nearly every conversation sits the same gating factor. Power. Power Has Become the Industry’s Defining Constraint Whether discussions centered on AI factories, investment capital, or campus expansion, Mahmood and Koblence noted that every conversation eventually returned to energy availability. “All of those questions are power,” Koblence said.

Read More »

Land and Expand: Early 2026 Megaprojects Reflect a Power-First Ethos

Vantage — Lighthouse (Port Washington, Wisconsin) Although the on-site ceremonial groundbreaking occurred in 2025, Vantage Data Centers’ Lighthouse campus in Port Washington, Wisconsin, remained one of the most closely watched AI infrastructure developments entering 2026, with updated local materials posted February 19 reinforcing the project’s scale and timeline. Announced in October 2025 in partnership with OpenAI and Oracle, Lighthouse is positioned as the Midwest anchor site within the companies’ broader Stargate expansion, which targets up to 4.5 gigawatts of additional AI capacity globally. Current plans call for four hyperscale data centers delivering nearly 902 MW of IT load on a site encompassing roughly 672 acres, with construction expected to run through 2028. From a Land and Expand perspective, the project exemplifies the new generation of AI campuses involving large-scale land banking paired with phased delivery designed to stay ahead of hyperscale demand curves. Just as notable is the project’s power and community framework. Vantage is working with WEC Energy Group’s We Energies on a dedicated rate structure under which the developer will underwrite 100% of the power infrastructure investment, a model explicitly designed to shield existing customers from rate increases. The utility partnership also includes plans to enable nearly 2 gigawatts of new zero-emission energy capacity, with approximately 70% allocated to the Lighthouse campus and the remainder supporting broader grid needs. Water and environmental positioning are also central to the project narrative. Lighthouse is designed around a closed-loop liquid cooling system intended to minimize water consumption, alongside local restoration investments aimed at achieving water positivity. Vantage has also committed to preserving significant portions of the site’s natural landscape while pursuing LEED certification for the campus. Economically, the development is expected to generate more than 4,000 primarily union construction jobs and over 1,000 long-term operational roles, while Vantage has pledged at

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »