Stay Ahead, Stay ONMINE

Alert to Kali Linux admins: Get the new signing key or no distro updates for you

This isn’t the first time Kali has had a signing key problem, noted Robert Beggs, head of Canadian penetration testing and incident response provider DigitalDefence. In 2018, a key was allowed to expire.  “It’s a minor blip,” he said in an interview, “that’s easy to overcome” by typing in a line of code, as detailed […]

This isn’t the first time Kali has had a signing key problem, noted Robert Beggs, head of Canadian penetration testing and incident response provider DigitalDefence. In 2018, a key was allowed to expire.

 “It’s a minor blip,” he said in an interview, “that’s easy to overcome” by typing in a line of code, as detailed in the Kali blog.

Loss of signing keys is “very uncommon” among application vendors, he said, “because this is an enterprise level project where someone should be managing a group of people together. The fact that it happened twice [at Kali] suggests they just don’t have central management. It [loss of the key] doesn’t make the product worse, doesn’t denigrate the excellent work they’re putting in. It just says that the central management piece is absent.”

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

HPE Aruba boosts NAC security, adds GreenLake ‘kill switch’

In addition, HPE Aruba tightened the integration between HPE Aruba Networking Central and HPE OpsRamp, the technology HPE bought in 2023 to manage hybrid and multicloud environments. OpsRamp monitors elements such as third-party switches, access points, firewalls, and routers. Tighter integration expands the ability to natively monitor third-party devices from vendors such as Cisco, Arista,

Read More »

Scottish suppliers not able to compete for renewables work due to policy

More than half of respondents to Scottish Renewables’ supply chain report said that domestic market conditions are making it more difficult to win work. Of those who answered the trade body’s questions, 60% did not think that the UK and Scottish governments are enabling the right market conditions for Scottish businesses to compete and secure contracts. This response was seen as “concerning” by Scottish Renewables’ director of energy transition and supply chain, Emma Harrick. She said: “If we want to unlock the full potential of this industry, that gap must be addressed. “And if we’re serious about delivering on our clean power potential and building a world-leading green economy, we need to think bigger than we ever have before. “This means urgently delivering an economic environment that maximises the entrepreneurial spirit of our clean power supply chain.” © Mhairi EdwardsThe arrival of Edda Brint_Edda Wind Service Operation Vessel (SOV) for Seagreen offshore wind farm at Montrose Port. In addition to this, 64% of those who answered Scottish Renewables’ questions said that project pipeline uncertainty within the renewables market is a significant barrier to scaling up their businesses. However, 36% said that “pipeline uncertainty does slow down progress but we are adapting”. Policy uncertainty was also reported as the biggest “challenge” facing Scotland’s renewable energy supply chain. “Navigating policy changes and regulatory uncertainties” was voted on by 36% of respondents as the “most pressing” issue facing their business. Securing enough “skilled talent for the growing industry” was the second biggest challenge at 28%, while 24% listed supply chain bottlenecks and materials shortages, and 12% mentioned attracting sufficient investment to “meet project demands”. More needed to crack supply chain ‘conundrum’ At a recent Aberdeen event, Scottish Renewables chief executive Claire Mack said that the green energy sector “has yet to crack the

Read More »

Vitol Steps Up Venezuela Oil Buying Before USA License Lapses

Vitol Group has stepped up oil purchases in Venezuela in the final weeks of a US license that allows the world’s largest independent oil trader to do business in the sanctioned nation. Vitol is set to load at least 2.5 million barrels of Venezuelan oil this month, according to documents seen by Bloomberg. It has more than doubled buying in April compared with March, when it took 1 million barrels, the data shows. Vitol’s license to operate in Venezuela hasn’t been previously reported. The company aims to move Venezuelan oil for as long as it can while remaining in compliance with sanctions, said a person with knowledge of the situation who asked not to be identified discussing non-public information. US President Donald Trump has ratcheted up pressure on Venezuelan President Nicolas Maduro’s regime, which he says poses an “extraordinary threat” to US national security. Trump is targeting Venezuela’s oil because it’s a major source of government revenue, and has revoked licenses for foreign energy companies operating there, including Chevron Corp., Repsol SA, Eni SpA, and Maurel & Prom. In contrast to Vitol, which trades with Venezuela and doesn’t operate any assets there, those other companies produce oil and gas in the country and are winding down operations ahead of the expiry of licenses issued by the US Treasury. The deadline for all the companies, including Vitol, to cease operating is May 27. Vitol’s latest purchases are being made under the terms of an agreement it signed in 2023, said another person familiar with matter. The trading company, which is headquartered in London, hasn’t inked any new Venezuelan deals since then, the person said. A spokesperson for Vitol declined to comment. A spokesperson for the White House National Security Council deferred to Treasury, which declined to comment. State oil producer Petroleos de

Read More »

Oil Slumps as Consumer Confidence Sinks

Oil dropped as the global trade war hurt the outlook for demand, with data indicating signs of strain in the US economy. West Texas Intermediate slipped 2.6% to settle around $60 a barrel, the lowest close in more than two weeks. The widely referenced US consumer confidence gauge weakened significantly in data released Tuesday, another sign of the pessimism stemming from President Donald Trump’s levies. Additional reports due this week, including manufacturing data out of China, will shed further light on the economic strength of the world’s biggest importer of crude. A handful of spreads along the futures curve traded in contango on Tuesday as traders brace for a “meaningful surplus” in the future, according to Morgan Stanley. At present, Brent’s nearer months are still pricier than those next in sequence, forming a rare configuration with little historical precedent, according to the firm. US crude is on track for the largest monthly loss since 2021, with prices battered by tit-for-tat tariffs between the US and its trading partners, as well as by OPEC+’s plans to revive more production. While many countries are entering into trade negotiations with Washington, Beijing says it has so far declined to engage. There is now an “elevated probability” that Saudi Arabia and other key OPEC+ nations opt to accelerate planned supply increases at the upcoming May 5 meeting, JPMorgan analysts led by Natasha Kaneva said in a note. “It is becoming clear to the market that growth over recent months has primarily been driven by stockpiling and pre-tariff hoarding,” said Arne Lohmann Rasmussen, chief analyst at A/S Global Risk Management. “That effect is fading, which could lead to a sharp decline in US growth as consumption starts drawing down existing inventories instead of boosting production.” Geopolitical tensions are showing signs of cooling, increasing the prospect

Read More »

Secretary of Energy Chris Wright to visit Rinnai America in Griffin, Georgia to Mark 100 Days of Unleashing American Energy

WASHINGTON—U.S. Secretary of Energy Chris Wright will travel to Griffin, Georgia, on Friday, May 2, to tour Rinnai America Corporation’s manufacturing facility. Secretary Wright will deliver remarks highlighting the Department’s progress in unleashing American energy dominance, protecting consumer freedom, and restoring energy leadership in the first 100 days of the Trump Administration.  Since taking office, Secretary Wright has postponed several burdensome Biden-era appliance mandates, including the rule targeting non-condensing tankless water heaters. Rinnai is the only company manufacturing non-condensing tankless water heaters in the United States. A Biden-era rule would have effectively banned these products—putting more than 200 Georgia jobs at risk, undermining U.S. manufacturing, and stripping away a cost-effective, high-efficiency option from American households. The visit marks the Department of Energy’s “First 100 Days of Unleashing American Energy Victories,” which includes: Halting restrictive appliance rules to defend consumer choice and protect domestic manufacturing jobs. Supporting energy-efficient innovation by preserving access to advanced home technologies. Reforming federal regulations to allow market-driven solutions and prevent government overreach in household energy use. Defending American competitiveness by ensuring manufacturers like Rinnai can continue producing energy-saving products in the U.S. Delivering relief to families by eliminating costly mandates that would have raised appliance prices across the board. Please contact [email protected] for press availability or media inquiries.

Read More »

Macquarie Strategists Forecast USA Crude Inventory Rise

In an oil and gas report sent to Rigzone late Monday by the Macquarie team, Macquarie strategists revealed that they are forecasting that U.S. crude inventories will be up by 4.3 million barrels for the week ending April 25. “This compares to our early look which anticipated a 7.0 million barrel build. On the product side of the ledger, in aggregate, our expectations are little changed,” the strategists noted in the report. “For this week’s crude balance, from refineries, we model crude runs slightly lower (-0.1 million barrels per day). Among net imports, we model a moderate increase, with exports (+0.5 million barrels per day) and imports (+1.0 million barrels per day) higher on a nominal basis,” they added. The strategists warned in the report that timing of cargoes remains a source of potential volatility in this week’s crude balance. “From implied domestic supply (prod.+adj.+transfers), we look for a small increase (+0.1 million barrels per day). Rounding out the picture, we anticipate a larger increase in SPR [Strategic Petroleum Reserve] stocks (+1.0 million barrels) this week,” the strategists went on to state. “Among products, we look for draws in gasoline (-2.3 million barrels) and distillate (-1.3 million barrels), with a build in jet (+0.5 million barrels). We model implied demand for these three products at 14.0 million barrels per day for the week ending April 25,” they continued. In an oil and gas report sent to Rigzone by the Macquarie team late Thursday, Macquarie strategists outlined that they saw “a large U.S. crude build” in the U.S. Energy Information Administration’s (EIA) next weekly petroleum status report. That report is scheduled to be released on April 30 and will include data for the week ending April 25. The EIA’s latest weekly petroleum status report at the time of writing was released

Read More »

Humber industry backs progress on Harbour Energy’s Viking CCS following HyNet agreement

Industrial bodies in the Humber are calling on the UK government to back Harbour Energy’s proposed Viking carbon capture and storage (CCS) project in the region. The backing comes after the UK government reached financial agreements with offshore CCS projects in Liverpool Bay and Teesside in recent months. Harbour Energy’s Viking CCS project was shortlisted for Track-2 status by the previous Conservative government in 2023 alongside the Acorn Project in Scotland. Since taking office last year, Labour has committed £22 billion in funding for the HyNet North West and East Coast Cluster Track-1 CCS projects. But the UK CCS sector has been calling on the government to “urgently” provide funding certainty for Viking and Acorn so they can move ahead. Humber industry backs Viking CCS In a joint statement, a group of Humber and UK trade bodies called on Labour to provide firm backing for Viking. The group includes the Northern Powerhouse Partnership, Hull and Humber Chamber of Commerce, Humber Energy Board and the Carbon Capture and Storage Association (CCSA). The trade bodies also backed Immingham gas plant operator VPI’s ambition to be the anchor emitter for Viking. © Supplied by VPIAn overhead view of the VPI Immingham plant in the Humber region. VPI is planning to invest £1.5 billion in a carbon capture facility at its Immingham plant, before using Harbour’s Viking project to transport and store CO2 emissions under the North Sea. The firm estimates the Immingham carbon capture project could create 1,500 jobs at peak construction, alongside a further 20,000 jobs for the wider Viking plans. Establishing Viking could also create a “huge” CO2 import opportunity for the UK, the trade bodies said. Viking will ‘secure the future’ of Humber industry Humber Energy Board executive director Jonathan Oxley said Viking CCS is a “key component” of efforts

Read More »

Nvidia AI supercluster targets agents, reasoning models on Oracle Cloud

Oracle has previously built an OCI Supercluster with 65,536 Nvidia H200 GPUs using the older Hopper GPU technology and no CPU that offers up to 260 exaflops of peak FP8 performance. According to the blog post announcing the availability, the Blackwell GPUs are available via Oracle’s public, government, and sovereign clouds, as well as in customer-owned data centers through its OCI Dedicated Region and Alloy offerings. Oracle joins a growing list of cloud providers that have made the GB200 NVL72 system available, including Google, CoreWeave and Lambda. In addition, Microsoft offers the GB200 GPUs, though they are not deployed as an NVL72 machine.

Read More »

Deep Data Center: Neoclouds as the ‘Picks and Shovels’ of the AI Gold Rush

In 1849, the discovery of gold in California ignited a frenzy, drawing prospectors from around the world in pursuit of quick fortune. While few struck it rich digging and sifting dirt, a different class of entrepreneurs quietly prospered: those who supplied the miners with the tools of the trade. From picks and shovels to tents and provisions, these providers became indispensable to the gold rush, profiting handsomely regardless of who found gold. Today, a new gold rush is underway, in pursuit of artificial intelligence. And just like the days of yore, the real fortunes may lie not in the gold itself, but in the infrastructure and equipment that enable its extraction. This is where neocloud players and chipmakers are positioned, representing themselves as the fundamental enablers of the AI revolution. Neoclouds: The Essential Tools and Implements of AI Innovation The AI boom has sparked a frenzy of innovation, investment, and competition. From generative AI applications like ChatGPT to autonomous systems and personalized recommendations, AI is rapidly transforming industries. Yet, behind every groundbreaking AI model lies an unsung hero: the infrastructure powering it. Enter neocloud providers—the specialized cloud platforms delivering the GPU horsepower that fuels AI’s meteoric rise. Let’s examine how neoclouds represent the “picks and shovels” of the AI gold rush, used for extracting the essential backbone of AI innovation. Neoclouds are emerging as indispensable players in the AI ecosystem, offering tailored solutions for compute-intensive workloads such as training large language models (LLMs) and performing high-speed inference. Unlike traditional hyperscalers (e.g., AWS, Azure, Google Cloud), which cater to a broad range of use cases, neoclouds focus exclusively on optimizing infrastructure for AI and machine learning applications. This specialization allows them to deliver superior performance at a lower cost, making them the go-to choice for startups, enterprises, and research institutions alike.

Read More »

Soluna Computing: Innovating Renewable Computing for Sustainable Data Centers

Dorothy 1A & 1B (Texas): These twin 25 MW facilities are powered by wind and serve Bitcoin hosting and mining workloads. Together, they consumed over 112,000 MWh of curtailed energy in 2024, demonstrating the impact of Soluna’s model. Dorothy 2 (Texas): Currently under construction and scheduled for energization in Q4 2025, this 48 MW site will increase Soluna’s hosting and mining capacity by 64%. Sophie (Kentucky): A 25 MW grid- and hydro-powered hosting center with a strong cost profile and consistent output. Project Grace (Texas): A 2 MW AI pilot project in development, part of Soluna’s transition into HPC and machine learning. Project Kati (Texas): With 166 MW split between Bitcoin and AI hosting, this project recently exited the Electric Reliability Council of Texas, Inc. planning phase and is expected to energize between 2025 and 2027. Project Rosa (Texas): A 187 MW flagship project co-located with wind assets, aimed at both Bitcoin and AI workloads. Land and power agreements were secured by the company in early 2025. These developments are part of the company’s broader effort to tackle both energy waste and infrastructure bottlenecks. Soluna’s behind-the-meter design enables flexibility to draw from the grid or directly from renewable sources, maximizing energy value while minimizing emissions. Competition is Fierce and a Narrower Focus Better Serves the Business In 2024, Soluna tested the waters of providing AI services via a  GPU-as-a-Service through a partnership with HPE, branded as Project Ada. The pilot aimed to rent out cloud GPUs for AI developers and LLM training. However, due to oversupply in the GPU market, delayed product rollouts (like NVIDIA’s H200), and poor demand economics, Soluna terminated the contract in March 2025. The cancellation of the contract with HPE frees up resources for Soluna to focus on what it believes the company does best: designing

Read More »

Quiet Genius at the Neutral Line: How Onics Filters Are Reshaping the Future of Data Center Power Efficiency

Why Harmonics Matter In a typical data center, nonlinear loads—like servers, UPS systems, and switch-mode power supplies—introduce harmonic distortion into the electrical system. These harmonics travel along the neutral and ground conductors, where they can increase current flow, cause overheating in transformers, and shorten the lifespan of critical power infrastructure. More subtly, they waste power through reactive losses that don’t show up on a basic utility bill, but do show up in heat, inefficiency, and increased infrastructure stress. Traditional mitigation approaches—like active harmonic filters or isolation transformers—are complex, expensive, and often require custom integration and ongoing maintenance. That’s where Onics’ solution stands out. It’s engineered as a shunt-style, low-pass filter: a passive device that sits in parallel with the circuit, quietly siphoning off problematic harmonics without interrupting operations.  The result? Lower apparent power demand, reduced electrical losses, and a quieter, more stable current environment—especially on the neutral line, where cumulative harmonic effects often peak. Behind the Numbers: Real-World Impact While the Onics filters offer a passive complement to traditional mitigation strategies, they aren’t intended to replace active harmonic filters or isolation transformers in systems that require them—they work best as a low-complexity enhancement to existing power quality designs. LoPilato says Onics has deployed its filters in mission-critical environments ranging from enterprise edge to large colos, and the data is consistent. In one example, a 6 MW data center saw a verified 9.2% reduction in energy consumption after deploying Onics filters at key electrical junctures. Another facility clocked in at 17.8% savings across its lighting and support loads, thanks in part to improved power factor and reduced transformer strain. The filters work by targeting high-frequency distortion—typically above the 3rd harmonic and up through the 35th. By passively attenuating this range, the system reduces reactive current on the neutral and helps stabilize

Read More »

New IEA Report Contrasts Energy Bottlenecks with Opportunities for AI and Data Center Growth

Artificial intelligence has, without question, crossed the threshold—from a speculative academic pursuit into the defining infrastructure of 21st-century commerce, governance, and innovation. What began in the realm of research labs and open-source models is now embedded in the capital stack of every major hyperscaler, semiconductor roadmap, and national industrial strategy. But as AI scales, so does its energy footprint. From Nvidia-powered GPU clusters to exascale training farms, the conversation across boardrooms and site selection teams has fundamentally shifted. It’s no longer just about compute density, thermal loads, or software frameworks. It’s about power—how to find it, finance it, future-proof it, and increasingly, how to generate it onsite. That refrain—“It’s all about power now”—has moved from a whisper to a full-throated consensus across the data center industry. The latest report from the International Energy Agency (IEA) gives this refrain global context and hard numbers, affirming what developers, utilities, and infrastructure operators have already sensed on the ground: the AI revolution will be throttled or propelled by the availability of scalable, sustainable, and dispatchable electricity. Why Energy Is the Real Bottleneck to Intelligence at Scale The major new IEA report puts it plainly: The transformative promise of AI will be throttled—or unleashed—by the world’s ability to deliver scalable, reliable, and sustainable electricity. The stakes are enormous. Countries that can supply the power AI craves will shape the future. Those that can’t may find themselves sidelined. Importantly, while AI poses clear challenges, the report emphasizes how it also offers solutions: from optimizing energy grids and reducing emissions in industrial sectors to enhancing energy security by supporting infrastructure defenses against cyberattacks. The report calls for immediate investments in both energy generation and grid capabilities, as well as stronger collaboration between the tech and energy sectors to avoid critical bottlenecks. The IEA advises that, for countries

Read More »

Colorado Eyes the AI Data Center Boom with Bold Incentive Push

Even as states work on legislation to limit data center development, it is clear that some locations are looking to get a bigger piece of the huge data center spending that the AI wave has created. It appears that politicians in Colorado took a look around and thought to themselves “Why is all that data center building going to Texas and Arizona? What’s wrong with the Rocky Mountain State?” Taking a page from the proven playbook that has gotten data centers built all over the country, Colorado is trying to jump on the financial incentives for data center development bandwagon. SB 24-085: A Statewide Strategy to Attract Data Center Investment Looking to significantly boost its appeal as a data center hub, Colorado is now considering Senate Bill 24-085, currently making its way through the state legislature. Sponsored by Senators Priola and Buckner and Representatives Parenti and Weinberg, this legislation promises substantial economic incentives in the form of state sales and use tax rebates for new data centers established within the state from fiscal year 2026 through 2033. Colorado hopes to position itself strategically to compete with neighboring states in attracting lucrative tech investments and high-skilled jobs. According to DataCenterMap.com, there are currently 53 data centers in the state, almost all located in the Denver area, but they are predominantly smaller facilities. In today’s era of massive AI-driven hyperscale expansion, Colorado is rarely mentioned in the same breath as major AI data center markets.  Some local communities have passed their own incentive packages, but SB 24-085 aims to offer a unified, statewide framework that can also help mitigate growing NIMBY (Not In My Backyard) sentiment around new developments. The Details: How SB 24-085 Works The bill, titled “Concerning a rebate of the state sales and use tax paid on new digital infrastructure

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »