Stay Ahead, Stay ONMINE

Notion bets big on integrated LLMs, adds GPT-4.1 and Claude 3.7 to platform

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Productivity platform Notion is betting on large language models (LLMs) powering more of its new enterprise capabilities, including building OpenAI’s GPT-4.1 and Anthropic’s Claude 3.7 into their dashboard. Even as both OpenAI and Anthropic start building […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Productivity platform Notion is betting on large language models (LLMs) powering more of its new enterprise capabilities, including building OpenAI’s GPT-4.1 and Anthropic’s Claude 3.7 into their dashboard.

Even as both OpenAI and Anthropic start building productivity features into their respective chat platforms, bringing these LLMs into a separate service shows how competitive the space is. 

Notion announced its new all-in-one AI toolkit inside the Notion workspace today, including AI meeting notes, enterprise search, research mode and the ability to switch between GPT-4.1 and Anthropic’s Claude 3.7.

One of the new features lets users chat with LLMs inside the Notion workspace and switch between models. Right now, Notion only supports GPT-4.1 and Claude 3.7. The idea is to reduce window and context switching.

The company said early adopters of the new feature include OpenAI, Ramp, Vercel and Harvey. 

Model mixing and fine-tuning

Notion built the features with a mix of LLMs from OpenAI and Claude, and its own models. 

The move away from pure reasoning models makes the model choice interesting for Notion. GPT-4.1 is technically not a reasoning model, while Claude 3.7 is a hybrid model— capable of acting as a regular LLM and a reasoning model. 

Reasoning models are having a moment, though many warn that these models can still sometimes lie. However, while reasoning models like OpenAI’s o3 (and yes, Claude 3.7 Sonnet) take their time to answer, going through different scenarios, they are not the best for quick thinking and data-gathering tasks. Many productivity tasks, like meeting transcriptions or searches for task data, don’t need the power of a reasoning model behind them. 

Sarah Sachs, Notion AI Engineering Lead, told VentureBeat in an email that the company aimed for features that didn’t sacrifice accuracy, safety and privacy, along with responding to queries in the speed enterprises need. 

“In order to achieve a low-latency experience, we fine-tuned the models with internal usage and feedback from trusted testers, in order to make the AI specialized in Notion retrieval tasks,” Sachs said. “This setup helps Notion AI understand business needs, give relevant answers, serve customers with sub-second latency, and keep customer data safe and compliant.”

Sachs said hosting and building with different models allows users to “pick the option that best fits their needs— whether that’s a more conversational tone, better coding capabilities, or faster response times.”

AI meeting notes and more

Notion AI for Work tracks and transcribes meetings for users, especially if they added Notion to their calendar, and it can listen in on calls. 

Users can use Notion for enterprise search by connecting apps like Slack, Microsoft Teams, GitHub, Google Drive, Sharepoint and Gmail. Sachs said Notion AI will search an organization’s internal documents, databases and the connected apps. 

The enterprise search results, along with other uploaded documents or a web search, allow Notion users to access the new Research Mode. It will draft documents directly from Notion while “analyzing all of your sources—plus the web—and think through the best response.”

Notion also added both GPT-4.1 and Claude 3.7 as chat options. OpenAI noted that Notion users can chat with GPT-4.1 on the workspace and create a Notion template directly from the conversation. Sachs said the company is working on adding more models to its chat feature. 

Subscribers to Notion’s Business and Enterprise plans with the Notion AI add-on get immediate access to the new features.  

Compete with the model providers

Even though Notion users can access both Anthropic and OpenAI on the platform, Notion still has to compete with model providers. 

OpenAI’s Deep Research has been hailed as a game-changer for agentic retrieval augmented generation (RAG). Google also has its version of Deep Research. And Anthropic can search the internet for you. 

Not to mention, Notion needs to compete with other platforms that already leverage AI. The meeting space is chock full of companies tracking, transcribing, summarizing and pulling insights from calls with AI.

However, Notion’s big selling point is that it has all these capabilities on one single platform. Enterprises can use all those different services but live outside their chosen productivity platform. Notion said having all of these features in one place, with one all-in-one pricing, will save enterprises from subscribing to different platforms. 

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Network data hygiene: The critical first step to effective AI agents

Many network teams manage some 15 to 30 different dashboards to track data across all the components in an environment, struggling to cobble together relevant information across domains and spending hours troubleshooting a single incident. In short, they are drowning in data. Artificial intelligence tools—and specifically AI agents—promise to ease

Read More »

Key takeaways from IBM Think partner event

The first week of May means flowers from April showers and that it’s time for IBM Think in Boston. The first day of the event has historically been the Partner Plus day, which is devoted to content for IBM partners, which include ISVs, technology partners and resellers. The 2025 keynote

Read More »

LandBridge Posts Higher Revenue

LandBridge Company LLC has reported $44 million in revenue for the first quarter of 2025, up from $36.5 million for the fourth quarter of 2024 and $19 million for the corresponding quarter a year prior. The company attributed the sequential increase to increases in surface use royalties of $6.8 million,

Read More »

Ghana’s Mahama Fine Tuning Regime to Accelerate Crude Extraction

Ghana is encouraging investors to get the country’s crude out of the ground to avoid the assets from getting stranded amid a global decarbonization push. The government is ready to roll out the “red carpet” to any investor prepared to drill and pump crude, President John Mahama said at the Africa CEO Forum in Abidjan, Ivory Coast’s commercial capital, on Tuesday. “Oil is in transition and so anybody who has any assets should be pumping like there’s no tomorrow,” he said. Oil output in Ghana fell for the past five consecutive years, mainly due to a lack of investment in additional exploration and new fields, according to the country’s Public Interest and Accountability Committee. Output reached 48.3 million barrels in 2024, compared with a peak of 71.4 million barrels in 2019, according to the body that was set up to regularly evaluate the efficient use of the nation’s petroleum resources. His call to scale up oil extraction echoes the stance of US President Donald Trump, who urged American producers to pump more oil earlier this year. OPEC has also been increasing output, resulting in Brent prices declining about 13% so far this year to $65.3 per barrel. Mahama won elections in December after pledging to revive the economy recovering from a debt crisis and a high cost of living. “The environment got a bit toxic,” Mahama said. “Government did not seem to prioritize the sector.” WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate or insulting comments will be removed. MORE FROM THIS AUTHOR Bloomberg

Read More »

High voltage sub-sea cable factory planned for Port of Tyne

Plans for a £900m sub-sea cable factory in Tyneside have been announced, with hopes the investment could generate at least 1,500 jobs. LS Eco Advanced Cables (LSEAC) has launched a five-week public consultation on its proposals to develop the cable facility, which will dramatically alter the local landscape with a 200-metre tall tower – more than twice the height of Parliament’s Elizabeth Tower. The public consultation is taking place before the company submits its final planning application to South Tyneside Council later this year. The factory will manufacture high-voltage cables designed to transport energy generated by offshore wind to UK shores. LSEAC says total investment is expected to be around £923m. LSEAC hopes to create 500 direct jobs through the new facility, as well as supporting a further 1,000 in the wider supply chain. The Port of Tyne will work with LSEAC, the North East Combined Authority (NECA), and local schools and colleges to provide training in the skills needed in order to ensure those roles are secured by local people. North East Mayor, Kim McGuinness, who leads NECA said: “We are leading the way to bring this incredible new facility to our region, and with it thousands of new jobs. LS Eco Advanced Cables will truly establish North East England as the home of the green energy revolution while also supporting the nation’s drive for energy security.” Regarding the scale and stature of the plant compared to the local landscape, Ms McGuinness says it will be “a new industrial icon – a symbol of our ambition on a global scale to compare with the Swan Hunter cranes that once overlooked the Tyne.” A new industrial icon Swan Hunter built ships on the banks of the Tyne between 1880 and 2006, although most of its construction operations had ceased in

Read More »

S&P Global Says USA Crude Oil Production Now Expected to Decline in 2026

An analysis piece sent to Rigzone by the S&P Global team on Monday stated that U.S. crude oil production is now expected to decline next year. “Slowing global oil demand amid extreme uncertainty about the future of U.S. trade and a coming supply surplus are expected to hobble U.S. oil production growth later this year and could lead to an annual decline in output in 2026, according to a new analysis by S&P Global Commodity Insights,” the analysis piece – by Jim Burkhard, Vice President and Global Head of Crude Oil Research, Ian Stewart, Associate Director, and the S&P Global Commodity Insights Crude Oil Markets team – stated. The piece noted that the latest update to the S&P Global Commodity Insights Global Crude Oil Markets Short-term Outlook now expects global oil demand growth to average 750,000 barrels per day in 2025. The analysis piece highlighted that this was a downward revision of 500,000 barrels per day from the prior outlook. “The new demand outlook represents a significant shift in momentum following strong oil demand growth in the first quarter of the year when demand grew by an estimated 1.75 million barrels per day year over year,” the piece stated. “In contrast, demand growth for the remaining quarters of [the] year is now expected to average 420,000 barrels per day,” it added. “As a result of the declining demand outlook and expected supply surplus (likely widened by recent OPEC+ decisions to accelerate the pace of production increases), U.S. crude oil production is now expected to decline in 2026 – the first year-on-year decline in U.S. production in roughly a decade, excluding the 2020 Covid-19 pandemic,” it went on to state. The analysis piece noted that total U.S production for 2025 is expected to average 13.46 million barrels per day before falling

Read More »

Gunvor Financed Gabon Deal for Tullow Oil Assets

Commodity trader Gunvor Group has made another significant oil deal in Gabon, its second since the country sought more control of national assets following a coup d’etat in 2023. State-owned Gabon Oil Company financed $220 million of a $300 million deal for Tullow Oil Plc’s assets in the country through an underwritten prepayment facility from Gunvor’s Middle East subsidiary, Tullow said in a filing Tuesday. Gunvor last year provided around $800 million to help Gabon finance another nationalization deal, the acquisition of Carlyle Group’s Assala Energy. A spokesperson for Gunvor declined to comment on the deal. Coming out of the most profitable period in their histories, trading firms are funneling cash into asset deals and prepayments to lock in more lucrative oil supply contracts as more competition from national oil companies and hedge funds lowers trading margins in oil markets. The Tullow assets are set to produce around 10,000 barrels of oil a day this year, the company said, adding that it will use the funds to reduce its debt profile. The company has been setting up deals to sell assets and bring down debt ahead of $1.39 billion in maturities next year.  ©2025 Bloomberg L.P. What do you think? We’d love to hear from you, join the conversation on the Rigzone Energy Network. The Rigzone Energy Network is a new social experience created for you and all energy professionals to Speak Up about our industry, share knowledge, connect with peers and industry insiders and engage in a professional community that will empower your career in energy. MORE FROM THIS AUTHOR Bloomberg

Read More »

Aramco in Talks with Woodside for Louisiana LNG Stake

Woodside Energy Group Ltd. has inked a non-binding collaboration deal with Saudi Arabian Oil Co. (Aramco) including for the Saudi oil giant’s potential offtake from and ownership in Louisiana LNG. The planned cooperation also eyes lower-carbon ammonia, among other global opportunities, the Australian oil and gas explorer and producer said in an online statement Wednesday. “This collaboration aligns with Woodside’s strategic vision to build a diverse and resilient global portfolio”, Woodside chief executive Meg O’Neill said. “It leverages our growing relationship with one of the world’s leading integrated energy and chemicals companies, to explore new opportunities which deliver value for both parties. “It is also another demonstration of the ongoing interest Louisiana LNG is generating among high-quality potential investors, following our recent agreement with Stonepeak to acquire a 40 percent interest in the project’s infrastructure holding company”. The deal was signed in Riyadh at the Saudi-United States Investment Forum, attended by Crown Prince and Prime Minister Mohammed bin Salman and President Donald Trump. On April 29 Woodside announced a positive final investment decision (FID) on the project, which it acquired as Driftwood LNG when it took over Tellurian Inc. for $1.2 billion last year. “The forecast total capital expenditure for the LNG project, pipeline and management reserve is $17.5 billion (100 percent)”, Woodside said. New York City-based Stonepeak Partners LP will gradually contribute $5.7 billion in exchange for a 40 percent interest, under an agreement announced earlier in April. Louisiana LNG has an Energy Department permit to export a cumulative 1.42 trillion cubic feet a year of natural gas equivalent, or 27.6 million metric tons per annum (MMtpa) of liquefied natural gas (LNG) according to Woodside, to FTA and non-FTA countries. The FID is for phase 1, which will build three liquefaction trains with a collective capacity of 16.5 MMtpa.

Read More »

ETZ chooses operator for Aberdeen hydrogen tech testing facility

ETZ Ltd has selected TÜV SÜD to operate its flagship Green Hydrogen Test and Demonstration Facilities (GHTDF) in Aberdeen. In addition, ETZ Ltd has also awarded consultancy Stantec, which has recently established a presence in Aberdeen, with a six-figure contract to provide an engineering design for the project. The GHTDF will offer testing, verification and certification of key components for green hydrogen production at scale, including electrolysers, valves and flow meters, to ensure their safety and long-term reliability. The facility, which is expected to be up and running in 2027, will be based at ETZ Ltd’s hydrogen campus in the south of Aberdeen, part of the company’s Energy Transition Zone project, which will be based on 75-100 acres of land around Aberdeen’s new £420 million south harbour. ETZ Ltd director for hydrogen and CCUS Martin McCormack said: “This facility will be a nationally significant asset for the UK, working as an exemplar for the testing, verification and certification of key industry hydrogen technologies. “In TÜV SÜD we have found the right partner to help deliver this test and demonstration centre, ably supported by Stantec’s deep technical expertise. As we seek to drive forward the hydrogen economy in the region, this is a prime opportunity for the wider hydrogen sector to support the development of this key infrastructure.” Driving hydrogen Ahead of its expected start date, ETZ Ltd, a private sector not-for-profit company backed by industrialist Sir Ian Wood, aims to attract private and public sector funding for the GHTDF. The company noted that industry support will be vital to realising the GHTDF project and ultimately the full benefits of GHTDF to the hydrogen economy. ETZ Ltd is looking to take a final investment decision on the facility later this year and start construction in 2026. In addition, to providing an

Read More »

Shell’s immersive cooling liquids the first to receive official certification from Intel

Along with the certification, Intel is offering a Xeon processor single-phase immersion warranty rider. This indicates Intel’s confidence in the durability and effectiveness of Shell’s fluids. Yates explained that the rider augments Intel’s standard warranty terms and is available to data center operators deploying 4th and 5th generation Xeon processors in Shell immersion fluids. The rider is intended to provide data center operators confidence that their investment is guaranteed when deployed correctly. Shell’s fluids are available globally and can be employed in retrofitted existing infrastructure or used in new builds. Cuts resource use, increases performance Data centers consume anywhere from 10 to 50 times more energy per square foot than traditional office buildings, and they are projected to drive more than 20% of the growth in electricity demand between now and 2030. Largely due to the explosion of AI, data center energy consumption is expected to double from 415 terawatt-hours in 2024 to around 945 TWh by 2030. There are several other technologies used for data center cooling, including air cooling, cold plate (direct-to-chip), and precision cooling (targeted to specific areas), but the use of immersion cooling has been growing, and is expected to account for 36% of data center thermal management revenue by 2028. With this method, servers and networking equipment are placed in cooling fluids that absorb and dissipate heat generated by the electronic equipment. These specialized fluids are thermally conductive but not electrically conductive (dielectric) thus making them safe for submerging electrical equipment.

Read More »

Cisco joins AI infrastructure alliance

“The addition of Cisco reinforces AIP’s commitment to an open-architecture platform and fostering a broad ecosystem that supports a diverse range of partners on a non-exclusive basis, all working together to build a new kind of AI infrastructure,” the group said in a statement.  Separately, Cisco announced AI initiatives centered in the Middle East region. Last week, Cisco CEO Chuck Robbins visited Saudi Arabia, UAE, Qatar, and Bahrain. This week, Jeetu Patel, executive vice president and chief product officer, is in Saudi Arabia, where he is participating in President Trump’s state visit to the region, according to Cisco. Related new projects include:  An initiative with HUMAIN, Saudi Arabia’s new AI enterprise to help build an open, scalable, resilient and cost-efficient AI infrastructure: “This landmark collaboration will set a new standard for how AI infrastructure is designed, secured and delivered – combining Cisco’s global expertise with the Kingdom’s bold AI ambitions. The multi-year initiative aims to position the country as a global leader in digital innovation,” Cisco stated. A collaboration with the UAE-basedG42 to co-develop a secure AI portfolio and AI-native services: Cisco and G42 will work together to assess the potential to co-develop and jointly deploy AI-powered cybersecurity packages, as well as a reference architecture that integrates Cisco’s networking, security, and infrastructure solutions specifically designed for high-performance computing. This collaboration aims to help customers build and secure AI-ready data centers and develop AI workloads effectively, according to the companies. Interest in Qatar’s digital transformation: Qatar’s Ministry of Interior and Cisco signed a letter of intent to collaborate on Qatar’s digital transformation, AI, infrastructure development and cybersecurity.

Read More »

Cato Networks introduces AI-powered policy analysis engine

Cato Networks this week announced a new policy analysis engine for its cloud-based secure access service edge platform that the company says will optimize and improve SASE policies, reduce risk, simplify compliance, and reduce manual maintenance efforts. Cato Autonomous Policies is built into the Cato SASE Cloud Platform and can provide enterprises with AI-driven recommendations to eliminate security exposure, tighten access controls, and improve network performance. The first use case of the policy engine is designed for firewall as a service (FWaaS) environments in which “firewall rule bloat” is present, Cato explained in a statement. The bloat comes from organizations accumulating thousands of rules that were designed to protect the environment, but after becoming outdated or misconfigured, actually lead to increased risk. “Most enterprises rely on a mix of firewalls deployed in data centers, branch offices, and cloud environments. Over time, rule sets grow, become inconsistent, and are filled with redundant, outdated, or conflicting entries,” wrote Demetris Booth, product marketing director at Cato Networks, in a blog post on the product news. “As a result, security policies become hard to manage, even harder to audit, and often misaligned with zero-trust principles. AI-driven firewall policy management is necessary for modern enterprises to streamline and optimize security operations.”

Read More »

Riverbed bolsters network acceleration for AI’s performance bottlenecks

“Enterprises are worried about bad actors capturing encrypted traffic and saving copies for when quantum computing advances can break the encryption, providing the bad actors with free access to data. It’s a real concern,” Frey explains. “Post-quantum cryptography is a way to get ahead of that now.” Riverbed also introduced the SteelHead 90 series of network acceleration appliances, which the company says will provide resilient network performance to customers. The series includes: SteelHead 8090, which delivers up to 60 Gbps of data movement over a WAN. It supports multiple 100 Gigabyte network interfaces to pull data from the LAN. SteelHead 6090, which delivers up to 20 Gbps of data movement over a WAN, targeted for mid-scale data centers. SteelHead 4090 and 2090, which support mid-sized data center and edge use cases, with 500 Mbps and 200 Mbps of accelerated traffic, as well as up to 10 Gbps of total traffic processing for quality of service (QoS) and application classification use cases. Riverbed SteelHead Virtual, is a software-only version designed for virtualization environments and private cloud deployments, which is compatible with VMWare ESXI, KVM, and Microsoft Hyper-V. “For customers that are familiar with Riverbed, this is a big change in performance. We’ve gone from moving one appliance at 30 Gbps to 60 Gbps. We want to make sure that whether it’s new AI projects or existing data projects, we have ubiquitous availability across clouds,” says Chalan Aras, senior vice president and general manager of Acceleration at Riverbed. “We’re making it less expensive to move data—we are about half the price of traditional data movement methods.” With this announcement, Riverbed also unveiled its Flex licensing subscription offering. According to Riverbed, Flex makes it possible for enterprises to transfer licenses from hardware to virtual to cloud devices at no cost. Enterprises can reassign

Read More »

Kyndryl and Microsoft expand partnership to streamline cloud operations

Kyndryl and Microsoft broadened their existing partnership to bring together Microsoft’s adaptive cloud approach and Kyndryl Distributed Cloud services to help customers better develop, manage and secure hybrid cloud operations.  Microsoft says its AI-infused adaptive cloud approach, which leverages Microsoft Azure Arc, Azure Local and Azure Cloud, enables customers to link distributed hybrid, multicloud, edge, and IoT resources under a single, secure application and data platform. The model uses customer data and an AI engine to offer predictive analytics, automated workflows, and threat detection and response to manage the environment, according to Microsoft.  Kyndryl will deliver the adaptive cloud approach to its customers through its Distributed Cloud services, which also use AI to improve automation, optimize workloads, enhance application performance, and reduce operational complexity. Kyndryl Distributed Cloud services create a mesh of interconnected resources and data from the data center to the edge in a multicloud environment, according to Kyndryl. Use cases include data center and edge modernization to support digital twins, AI video, robotic process automation, predictive maintenance, IoT data streams, asset tracking, and anomaly detection. 

Read More »

Tech CEOs warn Senate: Outdated US power grid threatens AI ambitions

The implications are clear: without dramatic improvements to the US energy infrastructure, the nation’s AI ambitions could be significantly constrained by simple physical limitations – the inability to power the massive computing clusters necessary for advanced AI development and deployment. Streamlining permitting processes The tech executives have offered specific recommendations to address these challenges, with several focusing on the need to dramatically accelerate permitting processes for both energy generation and the transmission infrastructure needed to deliver that power to AI facilities, the report added. Intrator specifically called for efforts “to streamline the permitting process to enable the addition of new sources of generation and the transmission infrastructure to deliver it,” noting that current regulatory frameworks were not designed with the urgent timelines of the AI race in mind. This acceleration would help technology companies build and power the massive data centers needed for AI training and inference, which require enormous amounts of electricity delivered reliably and consistently. Beyond the cloud: bringing AI to everyday devices While much of the testimony focused on large-scale infrastructure needs, AMD CEO Lisa Su emphasized that true AI leadership requires “rapidly building data centers at scale and powering them with reliable, affordable, and clean energy sources.” Su also highlighted the importance of democratizing access to AI technologies: “Moving faster also means moving AI beyond the cloud. To ensure every American benefits, AI must be built into the devices we use every day and made as accessible and dependable as electricity.”

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »