Stay Ahead, Stay ONMINE

Kumo’s ‘relational foundation model’ predicts the future your LLM can’t see

Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more The generative AI boom has given us powerful language models that can write, summarize and reason over vast amounts of text and other types of data. But when it comes to high-value […]

Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more


The generative AI boom has given us powerful language models that can write, summarize and reason over vast amounts of text and other types of data. But when it comes to high-value predictive tasks like predicting customer churn or detecting fraud from structured, relational data, enterprises remain stuck in the world of traditional machine learning. 

Stanford professor and Kumo AI co-founder Jure Leskovec argues that this is the critical missing piece. His company’s tool, a relational foundation model (RFM), is a new kind of pre-trained AI that brings the “zero-shot” capabilities of large language models (LLMs) to structured databases.

“It’s about making a forecast about something you don’t know, something that has not happened yet,” Leskovec told VentureBeat. “And that’s a fundamentally new capability that is, I would argue, missing from the current purview of what we think of as gen AI.”

Why predictive ML is a “30-year-old technology”

While LLMs and retrieval-augmented generation (RAG) systems can answer questions about existing knowledge, they are fundamentally retrospective. They retrieve and reason over information that is already there. For predictive business tasks, companies still rely on classic machine learning. 

For example, to build a model that predicts customer churn, a business must hire a team of data scientists who spend a considerably long time doing “feature engineering,” the process of manually creating predictive signals from the data. This involves complex data wrangling to join information from different tables, such as a customer’s purchase history and website clicks, to create a single, massive training table.

“If you want to do machine learning (ML), sorry, you are stuck in the past,” Leskovec said. Expensive and time-consuming bottlenecks prevent most organizations from being truly agile with their data.

How Kumo is generalizing transformers for databases

Kumo’s approach, “relational deep learning,” sidesteps this manual process with two key insights. First, it automatically represents any relational database as a single, interconnected graph. For example, if the database has a “users” table to record customer information and an “orders” table to record customer purchases, every row in the users table becomes a user node, every row in an orders table becomes an order node, and so on. These nodes are then automatically connected using the database’s existing relationships, such as foreign keys, creating a rich map of the entire dataset with no manual effort.

Relational deep learning (source: Kumo AI)

Second, Kumo generalized the transformer architecture, the engine behind LLMs, to learn directly from this graph representation. Transformers excel at understanding sequences of tokens by using an “attention mechanism” to weigh the importance of different tokens in relation to each other. 

Kumo’s RFM applies this same attention mechanism to the graph, allowing it to learn complex patterns and relationships across multiple tables simultaneously. Leskovec compares this leap to the evolution of computer vision. In the early 2000s, ML engineers had to manually design features like edges and shapes to detect an object. But newer architectures like convolutional neural networks (CNN) can take in raw pixels and automatically learn the relevant features. 

Similarly, the RFM ingests raw database tables and lets the network discover the most predictive signals on its own without the need for manual effort.

The result is a pre-trained foundation model that can perform predictive tasks on a new database instantly, what’s known as “zero-shot.” During a demo, Leskovec showed how a user could type a simple query to predict whether a specific customer would place an order in the next 30 days. Within seconds, the system returned a probability score and an explanation of the data points that led to its conclusion, such as the user’s recent activity or lack thereof. The model was not trained on the provided database and adapted to it in real time through in-context learning

“We have a pre-trained model that you simply point to your data, and it will give you an accurate prediction 200 milliseconds later,” Leskovec said. He added that it can be “as accurate as, let’s say, weeks of a data scientist’s work.” 

The interface is designed to be familiar to data analysts, not just machine learning specialists, democratizing access to predictive analytics.

Powering the agentic future

This technology has significant implications for the development of AI agents. For an agent to perform meaningful tasks within an enterprise, it needs to do more than just process language; it must make intelligent decisions based on the company’s private data. The RFM can serve as a predictive engine for these agents. For example, a customer service agent could query the RFM to determine a customer’s likelihood of churning or their potential future value, then use an LLM to tailor its conversation and offers accordingly.

“If we believe in an agentic future, agents will need to make decisions rooted in private data. And this is the way for an agent to make decisions,” Leskovec explained.

Kumo’s work points to a future where enterprise AI is split into two complementary domains: LLMs for handling retrospective knowledge in unstructured text, and RFMs for predictive forecasting on structured data. By eliminating the feature engineering bottleneck, the RFM promises to put powerful ML tools into the hands of more enterprises, drastically reducing the time and cost to get from data to decision.

The company has released a public demo of the RFM and plans to launch a version that allows users to connect their own data in the coming weeks. For organizations that require maximum accuracy, Kumo will also offer a fine-tuning service to further boost performance on private datasets.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Chronosphere unveils logging package with cost control features

According to a study by Chronosphere, enterprise log data is growing at 250% year-over-year, and Chronosphere Logs helps engineers and observability teams to resolve incidents faster while controlling costs. The usage and volume analysis and proactive recommendations can help reduce data before it’s stored, the company says. “Organizations are drowning

Read More »

Cisco CIO on the future of IT: AI, simplicity, and employee power

AI can democratize access to information to deliver a “white-glove experience” once reserved for senior executives, Previn said. That might include, for example, real-time information retrieval and intelligent process execution for every employee. “Usually, in a large company, you’ve got senior executives, and you’ve got early career hires, and it’s

Read More »

AMI MegaRAC authentication bypass flaw is being exploitated, CISA warns

The spoofing attack works by manipulating HTTP request headers sent to the Redfish interface. Attackers can add specific values to headers like “X-Server-Addr” to make their external requests appear as if they’re coming from inside the server itself. Since the system automatically trusts internal requests as authenticated, this spoofing technique

Read More »

Pertamina International Shipping Posts Higher Annual Revenue, Profit

PT Pertamina International Shipping (PIS) has reported $3.48 billion in revenue for 2024, up 4.4 percent from 2023. Profit grew 69.3 percent from $329.9 million in 2023 to $558.6 million in 2024. “This strong financial performance proves that the business transformation we have carried out is on the right path and affirms PIS’s position as one of Asia’s reputable maritime logistics companies. This business growth not only marks corporate advancement but also increases our contribution to national energy security”, PIS Corporate Secretary Muhammad Baron said. Throughout 2024, PIS transported 161 billion liters (42.5 billion gallons) of energy. It added 10 new tankers, including four VLGCs: Pertamina Gas Caspia, Dahlia, Tulip, Bergenia, PIS Jawa, Kalimantan, Kerinci, Rinjani, Rokan and Natuna. It had 102 vessels by year-end, the company said. “PIS continues to strengthen its fleet and increase domestic cargo transportation capacity in line with growing national energy demand. PIS is targeting higher transport capacity to ensure energy availability and support Asta Cita’s national energy independence agenda”, Baron added. By 2024, PIS vessels operated 65 international routes, up from 11 in 2021. To meet rising global demand, PIS opened three international offices in Singapore, Dubai, and London through its subsidiary PIS Asia Pacific, increasing non-captive revenue from 4 percent in 2021 to 19 percent in 2024, the company said. “We are grateful that PIS’s achievements, driven by increasingly efficient business transformation, have had a positive impact on the development of the national maritime industry. This is part of PIS’s commitment to revitalize various domestic industries and drive Indonesia’s economy sustainably”, Baron said. To contact the author, email [email protected] What do you think? We’d love to hear from you, join the conversation on the Rigzone Energy Network. The Rigzone Energy Network is a new social experience created for you and all energy professionals

Read More »

DNOW Acquires MRC Global for $1.5B

DNOW Inc. has agreed to buy MRC Global Inc. in an all-stock transaction valued $1.5 billion, creating a premier energy and industrial solutions provider. In a joint statement, the companies said the combination brings together complementary portfolios, services, and supply chain solutions. The combined entity, which will retain the name DNOW, will have a footprint of more than 350 service and distribution locations across more than 20 countries, the statement said. Under the terms of the agreement, MRC Global shareholders will receive 0.9489 shares of DNOW common stock for each share of MRC Global common stock, representing an 8.5 percent premium to MRC Global’s 30-day volume-weighted average price of $12.77 as of June 25. Upon the completion of the transaction, DNOW and MRC Global shareholders will respectively own approximately 56.5 percent and approximately 43.5 percent of the resulting company. “The combination of DNOW and MRC Global will create a premier energy and industrial solutions provider with a balanced portfolio of businesses and a diversified customer base fortifying long-term profitability and cash flow generation”, David Cherechinsky, DNOW President and CEO, said. “MRC Global’s differentiated product offerings and complementary assets strengthen DNOW’s 160-year legacy as a worldwide supplier of energy and industrial products and packaged, engineered process and production equipment”. The two companies expect to generate $70 million of annual cost synergies within three years. Cherechinsky will take on the same role in the combined company. Mark Johnson will remain as Chief Financial Officer. The DNOW board will be expanded to 10 directors to accommodate two MRC Global board members. Dick Alario will remain as chairman of the board. The parties expect to close the transaction in the fourth quarter. The combined company will remain headquartered in Houston, Texas. To contact the author, email [email protected] What do you think? We’d love to

Read More »

Borouge Partners with Honeywell to Develop Autonomous Operations in UAE

Abu Dhabi-based petrochemicals company Borouge PLC has partnered with Honeywell to conduct a proof of concept for AI-powered autonomous operations. The company said in a media release that this collaboration has the potential to revolutionize its UAE plant operations. The collaboration between Borouge and Honeywell is set to deliver the petrochemical industry’s first AI-driven control room designed for full-scale, real-time operation, establishing a new standard for the future of AI in petrochemicals, Borouge said. “Borouge’s AI, Digitalization, and Technology (AIDT) transformation program is setting new standards in operations, innovation, and business performance. By collaborating with global AI leaders such as Honeywell, we are accelerating growth, driving efficiency, and enhancing shareholder value. This project further strengthens Borouge’s competitive edge as we continue to deliver on our ambitious AIDT roadmap,” Hazeem Sultan Al Suwaidi, Chief Executive Officer of Borouge, said. The companies agreed to bring their expertise in process technology and autonomous control capabilities to identify new opportunities to deploy Agentic AI solutions and advanced machine learning algorithms, Borouge said. “Our collaboration with Borouge is a clear example of how joint efforts can accelerate innovation across industry. By integrating AI and automation technologies into core operations, we are helping unlock new levels of efficiency, safety, and performance. This agreement shows how advanced technologies, applied with purpose, can reshape industrial operations at scale”, George Bou Mitri, President of Honeywell Industrial Automation in the Middle East, Turkey, Africa and Central Asia, said. Borogue said the initiative seeks to implement proof-of-concept technologies that will improve its operations across its Ruwais facilities in the UAE. By embracing autonomous operations, Borouge said it can optimize production, cut energy consumption, and boost safety, all while driving down costs, at what will be the world’s largest petrochemical site. Borouge expects its AIDT program to bring in $575 million in

Read More »

ICYMI: ENERGY SECRETARY: It’s Time to Stop Subsidizing Solar and Wind in Perpuity

New York Post June 27, 2025 “How the Big Beautiful Bill will lower energy costs, shore up the electric grid — and unleash American prosperity” By Chris Wright How much would you pay for an Uber if you didn’t know when it would pick you up or where it was going to drop you off? Probably not much. Yet this is the same effect that variable generation sources like wind and solar have on our power grids. You never know if these energy sources will actually be able to produce electricity when you need it — because you don’t know if the sun will be shining or the wind blowing. Even so, the federal government has subsidized these sources for decades, resulting in higher electricity prices and a less stable grid. . . . President Donald Trump knows what to do: Eliminate green tax credits from the Democrats’ so-called Inflation Reduction Act, including those for wind and solar power. The One Big Beautiful Bill seeks to do that: Along with other proposals, like canceling billions in Biden Green New Deal money and making much-needed investments in the Strategic Petroleum Reserve, it aims to set an aggressive end date for these subsidies and build on the president’s push for affordable, abundant, and secure energy for the nation. . . . As Secretary of Energy — and someone who’s devoted his life to advancing energy innovation to better human lives — I, too, know how these Green New Deal subsidies are fleecing Americans. Wind and solar subsidies have been particularly wasteful and counterproductive. One example: The Renewable Electricity Production Tax Credit was first introduced in 1992, when wind energy was a nascent industry. This tax credit, originally set to phase out in 1999, was sold on a promise of low-cost energy with

Read More »

FERC’s Christie calls for dispatchable resources after grid operators come ‘close to the edge’

The ability of Midcontinent and East Coast grid operators to narrowly handle this week’s extreme heat and humidity without blackouts reflects the urgent need to ensure the United States has adequate power supplies, according to Mark Christie, chairman of the Federal Energy Regulatory Commission. “We’re simply not building generation fast enough, and we’re not keeping generation that we need to keep,” Christie said Thursday during a media briefing after the agency’s open meeting. “Some of our systems really came close to the edge.” The PJM Interconnection, the largest U.S. grid operator, hit a peak load of about 161 GW on Monday, nearly 5% above its 154 GW peak demand forecast for this summer and the highest demand on its system since 2011. The grid operator had about 10 GW to spare at the peak, according to Christie. At that peak, PJM’s fuel mix included gas at about 44%, nuclear at 20%, coal at 19%, solar at 5% and wind at 4%, according to Christie. Also, PJM told Christie that demand response was “essential” at reducing load, he said. PJM used nearly 4,000 MW of demand response to reduce its load, according to FERC Commissioner Judy Chang. “I see load flexibility as a key tool for grid operators to meet the challenges that we face,” Chang said. PJM called on demand response resources on Monday in its mid-Atlantic and Dominion regions, on Tuesday across its footprint and on Wednesday in its eastern zones, according to Dan Lockwood, a PJM spokesman. PJM was within its reserve requirements, but used DR to provide additional resources for the grid, he said in an email. Resource adequacy is the “central issue” facing the U.S., according to Christie, who said blackouts during the extreme heat could have been deadly. “You never know about the next time,

Read More »

New York Gov. Hochul hints at ‘fleet-style approach’ to nuclear deployments

Dive Brief: New York could take a page from Ontario’s playbook and deploy multiple reactors to reach and possibly exceed the 1-GW target Democratic Gov. Kathy Hochul announced on Monday, analysts with Clean Air Task Force said in an interview. Whether the New York Power Authority ultimately selects a large light-water reactor like the Westinghouse AP1000 or multiple units of a small modular design like the GE Hitachi BWRX-300, lessons learned on recent and ongoing nuclear builds could translate to lower final costs, said John Carlson, CATF’s senior Northeast regional policy manager. That could enable a “fleet-style approach” to deployment similar to Ontario Power Generation’s plan to build four 300-MW BWRX-300 reactors in sequence, lowering the final cost per unit, said Victor Ibarra, senior manager for CATF’s advanced nuclear energy program. On Monday, Hochul said the plan would “allow for future collaboration with other states and Ontario.” Dive Insight: Gov. Hochul on Monday directed NYPA and the New York Department of Public Service “to develop at least one new nuclear energy facility with a combined capacity of no less than one gigawatt of electricity, either alone or in partnership with private entities,” in upstate New York. As governor, Hochul has considerable influence over NYPA, the state-owned electric utility. In February, for example, she “demand[ed]” NYPA suspend a proposed rate hike. Hochul’s announcement made no mention of specific reactor types or designs, but the suggestion that multiple plants could be in the offing suggests NYPA could consider small modular designs alongside a large light-water reactor, Ibarra said. “It’s good that they’re taking a minute to explore both options,” Carlson said. “I don’t think they know which one is most beneficial yet.” Hochul said NYPA would immediately begin evaluating “technologies, business models and locations” for the first plant. The preconstruction process will

Read More »

HPE-Juniper deal clears DOJ hurdle, but settlement requires divestitures

In HPE’s press release following the court’s decision, the vendor wrote that “After close, HPE will facilitate limited access to Juniper’s advanced Mist AIOps technology.” In addition, the DOJ stated that the settlement requires HPE to divest its Instant On business and mandates that the merged firm license critical Juniper software to independent competitors. Specifically, HPE must divest its global Instant On campus and branch WLAN business, including all assets, intellectual property, R&D personnel, and customer relationships, to a DOJ-approved buyer within 180 days. Instant On is aimed primarily at the SMB arena and offers a cloud-based package of wired and wireless networking gear that’s designed for so-called out-of-the-box installation and minimal IT involvement, according to HPE. HPE and Juniper focused on the positive in reacting to the settlement. “Our agreement with the DOJ paves the way to close HPE’s acquisition of Juniper Networks and preserves the intended benefits of this deal for our customers and shareholders, while creating greater competition in the global networking market,” HPE CEO Antonio Neri said in a statement. “For the first time, customers will now have a modern network architecture alternative that can best support the demands of AI workloads. The combination of HPE Aruba Networking and Juniper Networks will provide customers with a comprehensive portfolio of secure, AI-native networking solutions, and accelerate HPE’s ability to grow in the AI data center, service provider and cloud segments.” “This marks an exciting step forward in delivering on a critical customer need – a complete portfolio of modern, secure networking solutions to connect their organizations and provide essential foundations for hybrid cloud and AI,” said Juniper Networks CEO Rami Rahim. “We look forward to closing this transaction and turning our shared vision into reality for enterprise, service provider and cloud customers.”

Read More »

Data center costs surge up to 18% as enterprises face two-year capacity drought

“AI workloads, especially training and archival, can absorb 10-20ms latency variance if offset by 30-40% cost savings and assured uptime,” said Gogia. “Des Moines and Richmond offer better interconnection diversity today than some saturated Tier-1 hubs.” Contract flexibility is also crucial. Rather than traditional long-term leases, enterprises are negotiating shorter agreements with renewal options and exploring revenue-sharing arrangements tied to business performance. Maximizing what you have With expansion becoming more costly, enterprises are getting serious about efficiency through aggressive server consolidation, sophisticated virtualization and AI-driven optimization tools that squeeze more performance from existing space. The companies performing best in this constrained market are focusing on optimization rather than expansion. Some embrace hybrid strategies blending existing on-premises infrastructure with strategic cloud partnerships, reducing dependence on traditional colocation while maintaining control over critical workloads. The long wait When might relief arrive? CBRE’s analysis shows primary markets had a record 6,350 MW under construction at year-end 2024, more than double 2023 levels. However, power capacity constraints are forcing aggressive pre-leasing and extending construction timelines to 2027 and beyond. The implications for enterprises are stark: with construction timelines extending years due to power constraints, companies are essentially locked into current infrastructure for at least the next few years. Those adapting their strategies now will be better positioned when capacity eventually returns.

Read More »

Cisco backs quantum networking startup Qunnect

In partnership with Deutsche Telekom’s T-Labs, Qunnect has set up quantum networking testbeds in New York City and Berlin. “Qunnect understands that quantum networking has to work in the real world, not just in pristine lab conditions,” Vijoy Pandey, general manager and senior vice president of Outshift by Cisco, stated in a blog about the investment. “Their room-temperature approach aligns with our quantum data center vision.” Cisco recently announced it is developing a quantum entanglement chip that could ultimately become part of the gear that will populate future quantum data centers. The chip operates at room temperature, uses minimal power, and functions using existing telecom frequencies, according to Pandey.

Read More »

HPE announces GreenLake Intelligence, goes all-in with agentic AI

Like a teammate who never sleeps Agentic AI is coming to Aruba Central as well, with an autonomous supervisory module talking to multiple specialized models to, for example, determine the root cause of an issue and provide recommendations. David Hughes, SVP and chief product officer, HPE Aruba Networking, said, “It’s like having a teammate who can work while you’re asleep, work on problems, and when you arrive in the morning, have those proposed answers there, complete with chain of thought logic explaining how they got to their conclusions.” Several new services for FinOps and sustainability in GreenLake Cloud are also being integrated into GreenLake Intelligence, including a new workload and capacity optimizer, extended consumption analytics to help organizations control costs, and predictive sustainability forecasting and a managed service mode in the HPE Sustainability Insight Center. In addition, updates to the OpsRamp operations copilot, launched in 2024, will enable agentic automation including conversational product help, an agentic command center that enables AI/ML-based alerts, incident management, and root cause analysis across the infrastructure when it is released in the fourth quarter of 2025. It is now a validated observability solution for the Nvidia Enterprise AI Factory. OpsRamp will also be part of the new HPE CloudOps software suite, available in the fourth quarter, which will include HPE Morpheus Enterprise and HPE Zerto. HPE said the new suite will provide automation, orchestration, governance, data mobility, data protection, and cyber resilience for multivendor, multi cloud, multi-workload infrastructures. Matt Kimball, principal analyst for datacenter, compute, and storage at Moor Insights & strategy, sees HPE’s latest announcements aligning nicely with enterprise IT modernization efforts, using AI to optimize performance. “GreenLake Intelligence is really where all of this comes together. I am a huge fan of Morpheus in delivering an agnostic orchestration plane, regardless of operating stack

Read More »

MEF goes beyond metro Ethernet, rebrands as Mplify with expanded scope on NaaS and AI

While MEF is only now rebranding, Vachon said that the scope of the organization had already changed by 2005. Instead of just looking at metro Ethernet, the organization at the time had expanded into carrier Ethernet requirements.  The organization has also had a growing focus on solving the challenge of cross-provider automation, which is where the LSO framework fits in. LSO provides the foundation for an automation framework that allows providers to more efficiently deliver complex services across partner networks, essentially creating a standardized language for service integration.  NaaS leadership and industry blueprint Building on the LSO automation framework, the organization has been working on efforts to help providers with network-as-a-service (NaaS) related guidance and specifications. The organization’s evolution toward NaaS reflects member-driven demands for modern service delivery models. Vachon noted that MEF member organizations were asking for help with NaaS, looking for direction on establishing common definitions and some standard work. The organization responded by developing comprehensive industry guidance. “In 2023 we launched the first blueprint, which is like an industry North Star document. It includes what we think about NaaS and the work we’re doing around it,” Vachon said. The NaaS blueprint encompasses the complete service delivery ecosystem, with APIs including last mile, cloud, data center and security services. (Read more about its vision for NaaS, including easy provisioning and integrated security across a federated network of providers)

Read More »

AMD rolls out first Ultra Ethernet-compliant NIC

The UEC was launched in 2023 under the Linux Foundation. Members include major tech-industry players such as AMD, Intel, Broadcom, Arista, Cisco, Google, Microsoft, Meta, Nvidia, and HPE. The specification includes GPU and accelerator interconnects as well as support for data center fabrics and scalable AI clusters. AMD’s Pensando Pollara 400GbE NICs are designed for massive scale-out environments containing thousands of AI processors. Pollara is based on customizable hardware that supports using a fully programmable Remote Direct Memory Access (RDMA) transport and hardware-based congestion control. Pollara supports GPU-to-GPU communication with intelligent routing technologies to reduce latency, making it very similar to Nvidia’s NVLink c2c. In addition to being UEC-ready, Pollara 400 offers RoCEv2 compatibility and interoperability with other NICs.

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »