Stay Ahead, Stay ONMINE

Inside Nvidia’s ‘grid-to-chip’ vision: How Vera Rubin and Spectrum-XGS push toward AI giga-factories

Vera Rubin MGX brings together Nvidia’s Vera CPUs and Rubin CPX GPUs, all using the same open MGX rack footprint as Blackwell. The system allows for numerous configurations and integrations. “MGX is a flexible, modular building block-based approach to server and rack scale design,” Delaere said. “It allows our ecosystem to create a wide range […]

Vera Rubin MGX brings together Nvidia’s Vera CPUs and Rubin CPX GPUs, all using the same open MGX rack footprint as Blackwell. The system allows for numerous configurations and integrations.

“MGX is a flexible, modular building block-based approach to server and rack scale design,” Delaere said. “It allows our ecosystem to create a wide range of configurations, and do so very quickly.”

Vera Rubin MGX will deliver almost eight times more performance than Nvidia’s GB 300 for certain types of calculation, he said. The architecture is liquid-cooled and cable-free, allowing for faster assembly and serviceability. Operators can quickly mix and match components such as CPUs, GPUs, or storage, supporting interoperability, Nvidia said.

Matt Kimball, principal data center analyst at Moor Insights and Strategy, highlighted the modularity and cleanness of the MGX tray design.

“This simplifies the manufacturing process significantly,” he said. For enterprises managing tens or even hundreds of thousands of racks, “this design enables a level of operational efficiency that can deliver real savings in time and cost.”

Nvidia is also showing innovation with cooling, Kimball said. “Running cooling to the midplane is a very clean design and more efficient.”

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Intel details new efficient Xeon processor line

The new chips will be able to support up to 12-channel DDR5 memory with speeds of up to 8000 MT/s, a substantial increase over the 8 channels of 6400MT/s in the prior generation. In addition to that, the platform will support up to 6 UPI 2.0 links with up to

Read More »

OPEC Keeps Oil Outlook Unchanged

OPEC kept its outlook for the oil market unchanged as the group and its allies continue to add barrels to the global market. Global oil demand will grow by 1.3 million barrels a day this year and by 1.4 million in 2026, the same as previously forecast, the Organization of the Petroleum Exporting Countries said in its monthly market report on Monday.  OPEC+ — representing the group and its allies — is adding barrels to a global market that’s expected to be in surplus next year. The group increased output by 630,000 barrels a day in September, according to the report, citing information from secondary sources. Brent oil futures have slumped 18% in the past year, with a big part of the drop coming since April, when the OPEC+ hikes were first announced, to the surprise of traders. The projected demand for OPEC+ crude was also unchanged from last month’s report, at 42.5 million barrels a day this year. Next year, demand for the group’s crude is expected to rise to 43.1 million barrels a day. This month, the OPEC+ alliance agreed to revive just 137,000 barrels a day of halted supply, a slower pace than earlier this year.  Other key figures from the monthly report: September output average about 43.05 million barrels a day. Oil demand in the OECD is expected to rise by around 130,000 barrels a day, with OECD Americas leading the increase. Non-OECD demand is expected to grow by around 1.2 million barrels a day. Transport fuels are expected to be the key growth source for oil demand this year, with jet fuel/kerosene adding 380,000 barrels a day of demand and diesel 300,000 barrels a day. Non-OPEC production is expected to grow by about 800,000 barrels a day this year, and 600,000 next year, also unchanged from last

Read More »

USA EIA Lowers Henry Hub Price Projections for 2025 and 2026

In its latest short term energy outlook (STEO), which was released on October 7, the U.S. Energy Information Administration (EIA) lowered its Henry Hub natural gas spot price forecast for both 2025 and 2026. According to this STEO, the EIA now sees the commodity coming in at $3.42 per million British thermal units (MMBtu) in 2025 and $3.94 per MMBtu in 2026. In its previous STEO, which was released in September, the EIA projected that the Henry Hub natural gas spot price would average $3.52 per MMBtu this year and $4.28 per MMBtu next year. The EIA’s latest STEO sees the commodity averaging $3.33 per MMBtu in the fourth quarter of 2025, $3.86 per MMBtu in the first quarter of next year, $3.31 per MMBtu in the second quarter, $3.91 per MMBtu in the third quarter, and $4.68 per MMBtu in the fourth quarter of 2026. In its September STEO, the EIA forecast that the Henry Hub natural gas spot price would average $3.04 per MMBtu in the third quarter of this year, $3.72 per MMBtu in the fourth quarter, $4.25 per MMBtu in the first quarter of 2026, $3.64 per MMBtu in the second quarter, $4.26 per MMBtu in the third quarter, and $4.99 per MMBtu in the fourth quarter. The EIA’s October STEO highlighted that the commodity came in at $3.03 per MMBtu in the third quarter of 2025. Both the October and the September STEOs showed that the Henry Hub natural gas spot price averaged $4.15 per MMBtu in the first quarter of this year, $3.19 per MMBtu in the second quarter, and $2.19 per MMBtu overall in 2024. In its October STEO, the EIA said it expects the Henry Hub spot price “to increase from around $3.00 per MMBtu in September to $4.10 per MMBtu by

Read More »

Serica to Acquire BP Stake in Culzean Field offshore UK

BP PLC signed an agreement to sell adjoining production and exploration assets on the United Kingdom’s side of the North Sea that include the Culzean gas and condensate field to Serica Energy PLC for at least $232 million, Serica said Monday. The transaction consists of BP’s 32 percent non-operating stake in the P111 license, which contains Culzean, and the adjacent P2544 exploration block, Serica said in a press release. The acquisition is subject to a 30-day preemption period during which BP’s partners – operator TotalEnergies SE (49.99 percent) and NEO NEXT Energy Ltd (18.01 percent) – may exercise their option to acquire BP’s stake on the same terms as those agreed by Serica, Serica said. “Should this transaction complete, it would deliver a step-change for Serica, adding material production and cash flows from the largest producing gas field in the UK”, Serica chief executive Chris Cox said. “Culzean is a world-class asset, delivering gas from a modern platform with exceptionally high uptime and low emissions”. BP’s share of production from Culzean was about 25,500 barrels of oil equivalent a day in the first half of 2025, Serica noted. The transaction involves an upfront cash consideration of $232 million, “subject to customary working capital adjustments and partially offset by the receipt of a payment reflecting interim post-tax cashflows between the economic date of the transaction [September 1, 2025] and the completion date, expected around the end of 2025”, Serica said. Serica may make two additional cash payments contingent on “successful results and production from a large exploration opportunity on the P2544 license and changes to the UK ring-fence fiscal regime”, Serica said. “The company can fund the consideration through a combination of interim cashflows from the Culzean interest and existing financial resources (including cash and undrawn amounts under the existing $525 million Reserve Based

Read More »

Sintana to Acquire Challenger

The boards of Sintana Energy Inc and Challenger Energy Group PLC have agreed on the terms of Sintana’s acquisition of the entire issued and to-be-issued share capital of Challenger for an implied diluted price of around GBP 44.72 million/ CAD 83.63 million ($59.72 million) or 16.61 pence per share. London-listed Challenger’s main assets are a 40 percent stake in Chevron Corp-operated AREA OFF-1 and a 100 percent ownership in AREA OFF-3, both exploration blocks offshore Uruguay. “Combined, these represent a total license holding of approximately 27,800 square kilometers (net to Challenger approximately 19,000 sqkm), making Challenger one of the largest offshore acreage holders in Uruguay and the only ‘junior’ with a position in offshore Uruguay and the broader offshore region (including northern Argentina and southern Brazil)”, Sintana said in a statement on its website. Challenger also holds legacy exploration assets in The Bahamas, for which the government has not responded to a request for license renewal, Sintana noted. Sintana, whose stock trades in Toronto, owns onshore and offshore exploration licenses in Namibia and Colombia’s Magdalena Basin. “The combination of Challenger and Sintana is expected to create a leading exploration platform spanning the Southern Atlantic conjugate margin, with a combined portfolio offering high-impact exposure to two of the world’s currently most active and emerging hydrocarbon exploration geographies with a diversified portfolio of licenses at various levels of maturity, underpinned by partnerships with majors that provide significant financial and operational support to reach material milestones”, the statement said. The expanded Sintana would own interests in eight licenses in Namibia – including the Mopane discovery – and Uruguay, as well as legacy assets in The Bahamas and Colombia, Sintana said. Under the transaction, Challenger shareholders would receive about 0.4705 common shares of Sintana for each Challenger ordinary share held. Immediately after the completion of the merger, Challenger

Read More »

Oil Prices Rebound After Sharp Declines

Oil prices are rebounding modestly after last week’s sharp declines driven by U.S.-China tensions and demand destruction fears, XMArabia Analyst Nadir Belbarka said in a statement sent to Rigzone on Monday. “President Trump’s plan for 100 percent tariffs on Chinese goods from November 1, alongside geopolitical risks in Ukraine, Gaza, and Russia, has heightened volatility,” Belbarka said in the statement. “Late-session reassurances, including a potential Xi meeting, stabilized sentiment, but the oil market remains fragile,” the analyst warned. “Price action is driven by macro expectations rather than supply fundamentals, with markets sensitive to Washington’s signals,” Belbarka added. In the statement, Belbarka noted that, “with a thin economic calendar, oil prices hinge on geopolitical tone and positioning flows”. “Trump’s aggressive stance keeps crude volatile, with investors eyeing U.S. storage data and potential OPEC/IEA [International Energy Agency] updates,” Belbarka said. Belbarka went on to state that oil’s near-term path will likely be shaped more by political signals than supply-demand fundamentals until fresh data emerges. In a separate market analysis sent to Rigzone today, Samer Hasn, Senior Market Analyst at XS.com also highlighted that oil prices rebounded today, “rising by more than two percent after suffering sharp losses that brought them to their lowest levels since May”.  “The upward move in oil comes as investors digested signs of trade relief following recent remarks by U.S. President Donald Trump, alongside stronger than expected Chinese trade data,” Hasn said in the analysis. A statement posted on the Donald J. Trump Truth Social page on October 10 stated, “it has just been learned that China has taken an extraordinarily aggressive position on Trade in sending an extremely hostile letter to the World, stating that they were going to, effective November 1st, 2025, impose large scale Export Controls on virtually every product they make, and some not

Read More »

Lack of Spare OPEC+ Capacity Could Lead to Price Spike

OPEC+’s spare capacity will be dwindling as it keeps bringing oil back into the market, even as total capacity for the group as a whole is growing. That’s what Ed Morse, Senior Adviser and Commodities Strategist at Hartree Partners, and previously the Global Head of Commodity Research at Citi Group, told Rigzone in an exclusive interview recently. Morse also highlighted a couple of issues with this spare capacity in the interview. “One main issue is the definition,” he said. “Is it oil that can be brought to market in one month, three months or six months? And is it a production level that can be sustained for a while – and is that ‘while’ six months, 12 months or even longer,” he added. “And then there is the issue of the domestic and international political risks to deliverable capacity,” Morse continued. Morse went on to tell Rigzone in the interview that some have spare OPEC+ capacity “at a fairly robust level”. The Hartree Partners strategist noted that, “from a definition of what can be brought to market in a four to six week period of time”, his own judgment is that OPEC+ capacity is about 2.75 million barrels per day. “Given what could happen in Russia and Iran alone, that isn’t a lot of oil to calm the market,” Morse warned. So, what does that mean for the oil market? Responding to this question, Morse told Rigzone that oil market consequences depend on when a disruption to supply would take place.  “The current market is physically range bound and weakening, with OPEC+ producers exporting around two million barrels per day more now than in mid-summer given the tangible and actual increases in the group’s production and the end of summer burn for power generation,” he said. “It remains backwardated but time

Read More »

Inside Nvidia’s ‘grid-to-chip’ vision: How Vera Rubin and Spectrum-XGS push toward AI giga-factories

Vera Rubin MGX brings together Nvidia’s Vera CPUs and Rubin CPX GPUs, all using the same open MGX rack footprint as Blackwell. The system allows for numerous configurations and integrations. “MGX is a flexible, modular building block-based approach to server and rack scale design,” Delaere said. “It allows our ecosystem to create a wide range of configurations, and do so very quickly.” Vera Rubin MGX will deliver almost eight times more performance than Nvidia’s GB 300 for certain types of calculation, he said. The architecture is liquid-cooled and cable-free, allowing for faster assembly and serviceability. Operators can quickly mix and match components such as CPUs, GPUs, or storage, supporting interoperability, Nvidia said. Matt Kimball, principal data center analyst at Moor Insights and Strategy, highlighted the modularity and cleanness of the MGX tray design. “This simplifies the manufacturing process significantly,” he said. For enterprises managing tens or even hundreds of thousands of racks, “this design enables a level of operational efficiency that can deliver real savings in time and cost.” Nvidia is also showing innovation with cooling, Kimball said. “Running cooling to the midplane is a very clean design and more efficient.”

Read More »

Cisco seriously amps-up Silicon One chip, router for AI data center connectivity

Some say deep buffers shouldn’t be used to handle this type of traffic; the contention is that these buffers fill and drain, creating jitter in the workloads, and that slows things down, Chopra told Network World. “But the real source of that challenge is not the buffers. It’s a poor congestion management scheme and poor load balancing with AI workloads, which are completely deterministic and predictable. You can actually proactively figure out how to place flows across the network and avoid the congestion,” he said. The 8223’s deep-buffer design provides ample memory to temporarily store packets during congestion or traffic bursts, an essential feature for AI networks where inter-GPU communication can create unpredictable, high-volume data flows, according to Gurudatt Shenoy, vice president of Cisco Provider Connectivity. “Combined with its high-radix architecture, the 8223 allows more devices to connect directly, reducing latency, saving rack space, and further lowering power consumption. The result is a flatter, more efficient network topology supporting high-bandwidth, low-latency communication that is critical for AI workloads,” Shenoy wrote in a blog post. NOS options Notably, the first operating systems that the 8223 supports are the Linux Foundation’s Software for Open Networking in the Cloud (SONiC) and Facebook open switching system (FBOSS) – not Cisco’s own IOS XR.  IXR will be supported, too, but at a later date, according to Cisco.  SONiC decouples network software from the underlying hardware and lets it run on hundreds of switches and ASICs from multiple vendors while supporting a full suite of network features such as Border Gateway Protocol (BGP), remote direct memory access (RDMA), QoS, and Ethernet/IP. One of the keys to SONiC is its switch-abstraction interface, which defines an API to provide a vendor-independent way of controlling forwarding elements such as a switching ASIC, an NPU, or a software switch in a uniform

Read More »

Utilities Race to Meet Surging Data Center Demand With New Power Models

Over the last 18 months or so, the energy generation industry and its public utilities have been significantly impacted by the AI data center boom. It has been demonstrated across North America that the increase in demand for power, as driven by the demand for hyperscale and AI data centers, greatly exceeds the ability of the industry to actually generate and deliver power to meet the demand. We have covered many of the efforts being made to control the availability of power. In response, utilities and regulators have begun rethinking how to manage power availability through means such as: temporary moratoriums on new data center interconnections; the creation of new rate classes; cogeneration and load-sharing agreements; renewable integration; and power-driven site selection strategies.  But the bottom line is that in many locations utilities will need to change the way they work and how and where they spend their CAPEX budgets. The industry has already realized that their demand forecast models are hugely out of date, and that has had a ripple effect on much of the planning done by public utilities to meet the next generation of power demand. Most utilities now acknowledge that their demand forecasting models have fallen behind reality, triggering revisions to Integrated Resource Plans (IRPs) and transmission buildouts nationwide. This mismatch between forecast and actual demand is forcing a fundamental rethink of capital expenditure priorities and long-term grid planning. Spend More, Build Faster Utilities are sharply increasing CAPEX and rebalancing their resource portfolios—not just for decarbonization, but to keep pace with multi-hundred-megawatt data center interconnects. This trend is spreading across the industry, not confined to a few isolated utilities. Notable examples include: Duke Energy raised its five-year CAPEX plan to $83 billion (a 13.7% increase) and plans to add roughly 5 GW of natural gas capacity

Read More »

Duos Pairs Mobile Power and Modular Edge Data Centers for Rapid Texas Rollout

Duos Technology Group has launched the fifth of its AI edge data centers, part of a plan to deploy 15 units by the end of 2025. The projects are executed through Duos Edge AI, a subsidiary focused on modular, rapidly installed edge data centers (EDCs) in underserved markets, beginning with school districts and regional carrier hubs across Texas. The newest site is being deployed on-premises with the Dumas Independent School District in Dumas, Texas. High-Density Edge Design Duos’ EDCs emphasize very high rack densities (100 kW+ per rack), SOC 2 Type II compliance, N+1 power with dual generators, and a 90-day build/turn-up cycle. Each site is positioned approximately 12 miles from end users, cutting latency for real-time workloads. To meet the power demands of these edge deployments, Duos formed Duos Energy and partnered with Fortress/APR Energy to deliver behind-the-meter mobile gas turbines. This approach allows compute to go live in 90 days without waiting years for utility interconnection upgrades. The goal is straightforward: move power and compute close to demand, with rapid deployment. Duos’ modular pods are designed for exurban and rural locations as localized compute hubs for carriers, schools, healthcare systems, and municipal users. The rugged design pairs high-density racks with the short deployment cycle and proximity targeting, enabling a wide range of applications. With Dumas ISD now live, Duos has five sites in Texas, including Amarillo/Region 16, Victoria/Region 3, Dumas ISD, and multiple Corpus Christi locations. Mobile Power vs. Modular Compute While Duos doesn’t consistently describe its data center units as “mobile,” they are modular and containerized, engineered for rapid, site-agnostic deployment. The “mobile” label more explicitly applies to Duos’ power strategy—a turbine fleet that can be fielded or re-fielded to match demand. From an operator’s perspective, the combined proposition functions like a mobile platform: pre-integrated compute pods

Read More »

Report: AMD could be Intel’s next foundry customer

[ Related: More Intel news and insights ] AMD has lagged behind Nvidia in the AI business but has done well in the federal supercomputing business, holding numerous top spots with supercomputers like El Capitan and Frontier. Manufacturing its chips in the United States would be a good way to get the Trump administration off its back given its push for domestic manufacturing of semiconductors. The Trump administration is pushing for 50% of chips sold in America to be manufactured domestically, and tariffs on chips that are not. It also faces outbound restrictions. Earlier this year, AMD faced export restrictions GPUs meant for China as part of U.S. export controls against China’s AI business. “I believe this is a smart move by AMD to secure capacity in the local market without fighting against Nvidia and Apple and their deeper pockets for the limited capacity at TSMC,” said Alvi Nguyen, senior analyst with Forrester Research.” With the US investment in Intel, followed by Nvidia, this is can be seen as diversifying their supply chain and providing cheaper, locally sourced parts.” For Intel, this will continue a streak of good news it has enjoyed recently. “Having customers take up capacity at their foundries will go a long way in legitimizing their semiconductor processes and hopefully create the snowball effect of getting even more US-based customers,” said Nguyen. In recent weeks, Intel has partnered with Nvidia to jointly make PC and data center chips. Nvidia also took a $5B stake in Intel. Earlier the Trump administration made a $11.1B, or 10%, stake in Intel.

Read More »

AI Infra Summit 2025 Insights: AI Factories at the Core of the Fifth Industrial Revolution

NVIDIA’s AI Factory: Engineering the Future of Compute NVIDIA’s keynote illuminated the invisible but indispensable heart of the AI revolution—the AI factory. This factory blends hardware and software innovation to achieve performance breakthroughs that transcend traditional limits. Technologies such as disaggregated rack-scale GPUs and the novel 4-bit floating point numerical representation move beyond incremental improvements; they redefine what is achievable in energy efficiency and cost-effectiveness. The software ecosystem NVIDIA fosters, including open-source frameworks like Dynamo, enables unprecedented flexibility in managing inference workloads across thousands of GPUs. This adaptability is crucial given the diverse, dynamic demands of modern AI, where workloads can fluctuate dramatically in scale and complexity. The continuous leap in benchmark performance, often quadrupling hardware capabilities through software alone, continues to reinforce their accelerated innovation cycle. NVIDIA’s framing of AI factories as both technology platforms and business enablers highlights an important shift. The value computed is not merely in processing raw data but in generating economic streams through optimizing speed, reducing costs, and creating new AI services. This paradigm is central to understanding how the new industrial revolution will operate through highly efficient AI factories uniting production and innovation. AWS and the Cloud’s Role in Democratizing AI Power Amazon Web Services (AWS) represents a key pillar in making AI capabilities accessible across the innovation spectrum. AWS’ focus on security and fault tolerance reflects maturity in cloud design, ensuring trust and resilience are priorities alongside raw compute power. The evolution towards AI agents capable of specification-driven operations signifies a move beyond traditional computing paradigms towards contextual, autonomous AI services embedded deeply in workflows. Their launch of EC2 P6-B200 instances with next-generation Blackwell GPUs and specialized Trainium chips represents a continual drive to optimize AI training and inference at scale and out-of-box improvements in performance of 85% reduction in training time;

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »