Stay Ahead, Stay ONMINE

Anthropic just gave Claude a superpower: real-time web search. Here’s why it changes everything

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Anthropic announced today that its AI assistant Claude can now search and process information from the internet in real time, addressing one of users’ most requested features and closing a critical competitive gap with OpenAI’s ChatGPT. […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Anthropic announced today that its AI assistant Claude can now search and process information from the internet in real time, addressing one of users’ most requested features and closing a critical competitive gap with OpenAI’s ChatGPT.

The new web search capability, available immediately for paid Claude users in the United States, transforms the AI assistant from a tool limited by its training data cutoff to one that can access and synthesize the latest information across the web.

“With web search, Claude has access to the latest events and information, boosting its accuracy on tasks that benefit from the most recent data,” Anthropic said in its announcement. The company emphasized that Claude will provide direct citations to sources, allowing users to fact-check information— a direct response to growing concerns about AI hallucinations and misinformation.

AI arms race intensifies as Anthropic secures billions in funding

This launch comes at an important moment in the rapidly evolving AI sector. Just three weeks ago, Anthropic secured $3.5 billion in Series E funding at a post-money valuation of $61.5 billion, underscoring the high stakes in the AI race. Major backers include Lightspeed Venture Partners, Google (which holds a 14% stake), and Amazon, which has integrated Claude into its Alexa+ service.

The web search rollout also follows Anthropic’s recent release of Claude 3.7 Sonnet, which the company claims has set “a new high-water mark in coding abilities.” This focus on programming proficiency appears strategic, especially in light of CEO Dario Amodei’s recent prediction at a Council on Foreign Relations event that “in three to six months, AI will be writing 90% of the code” that software developers currently produce.

The timing of this feature launch reveals Anthropic’s determination to challenge OpenAI’s dominance in the consumer AI assistant market. While Claude has gained popularity among technical users for its nuanced reasoning and longer context window, the lack of real-time information access has been a significant handicap in head-to-head comparisons with ChatGPT. This update effectively neutralizes that disadvantage.

How Claude’s web search transforms enterprise decision-making

Unlike traditional search engines that return a list of links, Claude processes search results and delivers them in a conversational format. Users simply toggle on web search in their profile settings, and Claude will automatically search the internet when needed to inform its responses.

Anthropic highlighted several business use cases for the web-enabled Claude: sales teams analyzing industry trends, financial analysts assessing current market data, researchers building grant proposals, and shoppers comparing products across multiple sources.

This feature fundamentally changes how enterprise users can interact with AI assistants. Previously, professionals needed to toggle between search engines and AI tools, manually feeding information from one to the other. Claude’s integrated approach streamlines this workflow dramatically, potentially saving hours of research time for knowledge workers.

For financial services firms in particular, the ability to combine historical training data with breaking news creates a powerful analysis tool that could provide genuine competitive advantages. Investment decisions often hinge on connecting disparate pieces of information quickly — exactly the kind of task this integration aims to solve.

Behind the scenes: The technical infrastructure powering Claude’s new capabilities

Behind this seemingly straightforward feature lies considerable technical complexity. Anthropic has likely spent months fine-tuning Claude’s ability to search effectively, understand context, and determine when web search would improve its responses.

The update integrates with other recent technical improvements to the Anthropic API, including cache-aware rate limits, simpler prompt caching, and token-efficient tool use — all aimed at helping developers process more requests while reducing costs. These enhancements, announced earlier this month, can reduce token usage by up to 90% for certain applications.

Anthropic has also upgraded its developer console to enable collaboration among teams working on AI implementations. The revised console allows developers to share prompts, collaborate on refinements, and control extended thinking budgets — features particularly valuable for enterprise customers integrating Claude into their workflows.

The investment in these backend capabilities suggests Anthropic is building for scale, anticipating rapid adoption as more companies integrate AI into their operations. By focusing on developer experience alongside user-facing features, Anthropic is creating an ecosystem rather than just a product — a strategy that has served companies like Microsoft well in enterprise markets.

Voice mode: Anthropic’s next frontier in natural AI interaction

Web search may be just the beginning of Anthropic’s feature expansion. According to a recent report in the Financial Times, the company is developing voice capabilities for Claude, potentially transforming how users interact with the AI assistant.

Mike Krieger, Anthropic’s chief product officer, told the Financial Times that the company is working on experiences that would allow users to speak directly to Claude. “We are doing some work around how Claude for desktop evolves… if it is going to be operating your computer, a more natural user interface might be to [speak to it],” Krieger said.

The company has reportedly held discussions with Amazon and voice-focused AI startup ElevenLabs about potential partnerships, though no deals have been finalized.

Voice interaction would represent a significant leap forward in making AI assistants more accessible and intuitive. The current text-based interaction model creates friction that voice could eliminate, potentially expanding Claude’s appeal beyond tech-savvy early adopters to a much broader user base.

How Anthropic’s safety-first approach shapes regulatory conversations

As Anthropic expands Claude’s capabilities, the company continues to emphasize its commitment to responsible AI development. In response to California Governor Gavin Newsom’s Working Group on AI Frontier Models draft report released earlier this week, Anthropic expressed support for “objective standards and evidence-based policy guidance,” particularly highlighting transparency as “a low-cost, high-impact means of growing the evidence base around a new technology.”

“Many of the report’s recommendations already reflect industry best practices which Anthropic adheres to,” the company stated, noting its Responsible Scaling Policy that outlines how it assesses models for misuse and autonomy risks.

This focus on responsible development represents a core differentiator in Anthropic’s brand positioning since its founding in 2021, when Amodei and six colleagues left OpenAI to create an AI company with greater emphasis on safety.

Anthropic’s approach to regulation appears more collaborative than defensive, positioning the company favorably with policymakers who are increasingly focused on AI oversight. By proactively addressing safety concerns and contributing constructively to regulatory frameworks, Anthropic may be able to shape rules in ways that align with its existing practices while potentially creating compliance hurdles for less cautious competitors.

The future of AI assistants: From chatbots to indispensable digital partners

The addition of web search to Claude represents more than just feature parity with competitors — it signals Anthropic’s ambition to create AI systems that can function as comprehensive digital assistants rather than specialized tools.

This development marks a significant evolution in AI assistants. First-generation large language models were essentially sophisticated autocomplete systems with impressive but limited capabilities. The integration of real-time information access, combined with Claude’s existing reasoning abilities, creates something qualitatively different: a system that can actively help solve complex problems using up-to-date information.

For businesses investing in AI integration, Claude’s new capabilities offer compelling advantages. Cognition, maker of the AI software developer assistant Devin, has already leveraged Anthropic’s prompt caching to provide more context about codebases while reducing costs and latency, according to the company’s leadership.

The real potential of these systems goes far beyond simple information retrieval. By combining current data with deep contextual understanding, AI assistants like Claude could transform knowledge work by handling substantial portions of research, analysis, and content creation — freeing humans to focus on judgment, creativity, and interpersonal aspects of their roles.

What web search means for Claude users today and tomorrow

Web search is available now for all paid Claude users in the United States, with support for free users and international expansion planned “soon,” according to the announcement. Users can activate the feature through their profile settings.

As competition in the AI assistant space intensifies, Anthropic’s deliberate approach to expanding Claude’s capabilities while maintaining its focus on safety and transparency suggests a long-term strategy focused on building user trust alongside technical advancement.

The race between AI companies is increasingly about balancing capability with reliability and trust. Features like web search with source citations serve both goals simultaneously, providing users with more powerful tools while maintaining transparency about information sources.

With Claude now able to tap into the internet’s vast resources while maintaining its characteristic nuanced reasoning, Anthropic has eliminated a key competitive disadvantage. More importantly, the company has taken a significant step toward creating AI systems that don’t just respond to queries but actively help users navigate an increasingly complex information landscape.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Google Cloud partners with mLogica to offer mainframe modernization

Other than the partnership with mLogica, Google Cloud also offers a variety of other mainframe migration tools, including Radis and G4 that can be employed to modernize specific applications. Enterprises can also use a combination of migration tools to modernize their mainframe applications. Some of these tools include the Gemini-powered

Read More »

Uganda-Tanzania Oil Pipeline Secures First Tranche of External Financing

A Uganda-Tanzania petroleum pipeline project majority-owned by TotalEnergies SE has secured the first tranche of external syndicated financing. The group of backers include regional banks African Export Import Bank, Standard Bank of South Africa Ltd., Stanbic Bank Uganda Ltd., KCB Bank Uganda and Islamic Corporation for the Development of the Private Sector (ICD), the project joint venture EACOP Ltd. said in an online statement. EACOP did not disclose any amount. The East African Crude Oil Pipeline will transport up to 246,000 barrels a day from the Lake Albert oilfields in Uganda to the port of Tanga, Tanzania, for export to the global market, according to EACOP. The funding “demonstrates the support of financial institutions on this transformative regional infrastructure”, it said. Construction was more than 50 percent complete at the end of last year, EACOP said, noting over 8,000 Ugandan and Tanzanian citizens are employed for the project. Construction started last year and is expected to take 2 years to complete, according to EACOP. Besides a 1,443-kilometer (896.63 miles), 24-inch buried pipeline, the project will also install 6 pumping stations, 2 pressure reduction stations and a marine export terminal with a 3-megawatt solar plant, according to EACOP. Early last year TotalEnergies, which owns 62 percent of EACOP, said it had commissioned an assessment of its land acquisition process for the pipeline project and an associated oil development project, following allegations by an interfaith organization that the French energy giant failed to protect hundreds of graves. “As the land acquisition process draws to a close, this mission will evaluate the land acquisition procedures implemented, the conditions for consultation, compensation and relocation of the populations concerned, and the grievance handling mechanism”, TotalEnergies said in a press release January 4, 2024. “It will also assess the actions taken by TotalEnergies EP Uganda and

Read More »

EVOL X Scottish Enterprise: It’s high tide time for Scotland’s energy mix

LISTEN: EVOL X Scottish Enterprise: It’s high tide time for Scotland’s energy mix  Scotland is currently leading the world in the development and testing of tidal stream and wave energy technologies. Two of the world’s first tidal stream arrays have been built and operated in Scotland since 2016. Excitement for the sector is also growing since a number of projects have won support through the UK’s contracts for difference (CfD) scheme in recent years. Tune in to hear Energy Voice news editor Erikka Askeland  speak to Johanna Money, energy transition specialist at Scottish Enterprise, about the vast economic and environmental potential for Scotland underneath the water as it develops energy generation technology that is clean and comes from consistent and predictable tides. We also take a trip up to Orkney to speak with Eileen Linklater, corporate affairs director of European Marine Energy Centre (EMEC), where conditions are ideal to prove why the tide will wait for no one.

Read More »

Unlocking Scotland’s tidal energy potential

A new podcast looks at how tidal energy can power Scotland’s future, strengthen its economy, and lead the way in renewable innovation. The talk features Johanna Money, an energy transition specialist at Scottish Enterprise, who outlines the extent of the opportunity and how Scottish Enterprise will support businesses. A new podcast looks at how tidal energy can power Scotland’s future, strengthen its economy, and lead the way in renewable innovation. The talk features Johanna Money, an energy transition specialist at Scottish Enterprise, who outlines the extent of the opportunity and how Scottish Enterprise can  support businesses. Scotland is at the forefront of the tidal energy revolution, with its powerful coastal waters providing an outstanding opportunity to harness this clean, predictable energy. Johanna said: “The marine renewable energy sector   has a huge role to play in Scotland’s journey to net zero. We’re looking at how we can support Scottish companies to  engage with tidal energy and how it can contribute to economic growth in Scotland.” £4.5bn of potential © Supplied by Scottish EnterpriseJohanna Money is a specialist in the energy transition. A recent report commissioned by Scottish Enterprise and authored by the University of Edinburgh’s Policy and Innovation Group, highlights the economic potential of tidal energy. The report estimates that by 2050, Scotland could be home to 4.3 gigawatts of tidal energy, and from those and a further 1.9 GW of other UK projects generate a GVA benefit to Scotland of £4.5bn, as well as creating almost 6,000 jobs from this domestic market. One of the key factors setting tidal energy apart from other renewables is its predictability. “Unlike wind or solar, tidal energy is entirely predictable,” said Johanna. “We know exactly when the tide will go in and out for centuries to come. That kind of reliability provides significant energy

Read More »

Repsol, NEO Announce New Joint Venture

In a statement posted on its website on Thursday, Repsol Resources UK Limited announced a “strategic consolidation of Repsol Resources UK with NEO UK to form NEO NEXT”. “Repsol is pleased to announce a landmark agreement to consolidate its UK North Sea upstream business with NEO Energy Group Limited, a leading independent UK oil and gas operator, through a share for share combination,” Repsol noted in the statement. “The resulting joint venture will position itself as a robust and diversified UK North Sea-focused oil and gas company that is poised to become one of the largest independent producers in the UK Continental Shelf (UKCS),” it added. Repsol highlighted in the statement that this strategic combination will “significantly enhance the operational scale, efficiency, and growth prospects of the combined entity, while also reinforcing Repsol E&P’s long-term commitment to maximizing the value of its UK assets”. The new joint venture has projected 2025 production of approximately 130,000 barrels of oil equivalent per day, according to the statement, which said the combined company’s enhanced and competitive positioning is expected to drive substantial growth and long-term value creation for shareholders. In the statement, Repsol said the new entity will operate “a highly diversified portfolio, including 11 production hubs and substantial undeveloped reserves” and noted that the combined group is targeting synergies exceeding $1 billion. It added that the leadership of new joint venture will be guided by a joint executive team, representing both Repsol UK and NEO UK. Repsol also outlined in the statement that Repsol E&P will retain a decommissioning funding commitment “up to a nominal amount of $1.8 billion, representing approximately a 40 percent of the decom liabilities related to its legacy assets”. It also said Repsol E&P will continue to provide decommissioning security for existing Repsol E&P legacy assets. On deal

Read More »

Nigerian State Oil Giant Seeks Advisers for Delayed IPO

Nigeria’s state oil company is courting international and local markets for advisers as it prepares for a long-delayed initial public offering. The Nigerian National Petroleum Company Ltd. is seeking investor relations specialists, IPO advisers and investment bank partners to help manage the share sale, Chief Finance and Investor Relations Officer Olugbenga Oluwaniyi said, according to a statement issued in Abuja, the nation’s capital. He did not give a time line for the IPO, saying that the company was in the “final stage” of obtaining a listing. NNPC had repeatedly delayed a share sale after it was first proposed nearly a decade ago. An IPO was last mooted four years ago, when it posted a profit for the first time in 44 years. Then-President Muhammad Buhari signed legislation in 2021 that allows the company to sell shares to the public. The company would draw on the experience of Saudi Aramco, the Saudi Arabian oil giant that raised $25.6 billion in the largest IPO in history in 2019, Mele Kyari, the group chief executive officer, said in 2021. What do you think? We’d love to hear from you, join the conversation on the Rigzone Energy Network. The Rigzone Energy Network is a new social experience created for you and all energy professionals to Speak Up about our industry, share knowledge, connect with peers and industry insiders and engage in a professional community that will empower your career in energy. MORE FROM THIS AUTHOR Bloomberg

Read More »

BP, Iraq Finalize Deal on Oil Development in Kurds-Claimed Province

Baghdad has fully ratified a contract with BP PLC for oil and gas redevelopments and potential exploration in Kirkuk province, the company said. The agreement, laid out last year, provides for re-entry and rehabilitation in several “giant oil fields”, according to the British energy giant. The assets eyed for production are the Kirkuk oilfield’s Avanah and Baba domes and three adjacent fields: Bai Hassan, Jambur and Khabbaz. All of these are operated by state-owned North Oil Co. (NOC). Besides BP and NOC, the tripartite deal also includes the Iraqi state’s North Gas Co. (NGC). The agreement is for an initial phase with a target production of over 3 billion barrels of oil equivalent (boe). “The wider resource opportunity across the contract and surrounding area is believed to include up to 20 billion barrels of oil equivalent”, BP said in an online statement. “This is an enormous opportunity as we grow bp’s oil and gas business and fully aligned with our strategy of strengthening our upstream portfolio”, commented chief executive Murray Auchincloss. BP will now form a new unincorporated company to take over operatorship from NOC. The operator will be composed mostly of personnel from NOC and NGC, with secondees from BP. “Subsequent to this agreement, bp expects to form a standalone incorporated joint venture to hold its interests in the operator”, BP earlier said February 25. “Under the terms of the agreement, bp will work with NOC, NGC and the new operator to stabilize and grow production”, it said. “Work will include a drilling program, the rehabilitation of existing wells and facilities, and the construction of new infrastructure, including gas expansion projects. “Under the agreement, bp’s remuneration will be linked to incremental production volumes, price and costs. bp will be able to book a share of production and reserves proportionate to

Read More »

Former Arista COO launches NextHop AI for customized networking infrastructure

Sadana argued that unlike traditional networking where an IT person can just plug a cable into a port and it works, AI networking requires intricate, custom solutions. The core challenge is creating highly optimized, efficient networking infrastructure that can support massive AI compute clusters with minimal inefficiencies. How NextHop is looking to change the game for hyperscale networking NextHop AI is working directly alongside its hyperscaler customers to develop and build customized networking solutions. “We are here to build the most efficient AI networking solutions that are out there,” Sadana said. More specifically, Sadana said that NextHop is looking to help hyperscalers in several ways including: Compressing product development cycles: “Companies that are doing things on their own can compress their product development cycle by six to 12 months when they partner with us,” he said. Exploring multiple technological alternatives: Sadana noted that hyperscalers might try and build on their own and will often only be able to explore one or two alternative approaches. With NextHop, Sadana said his company will enable them to explore four to six different alternatives. Achieving incremental efficiency gains: At the massive cloud scale that hyperscalers operate, even an incremental one percent improvement can have an oversized outcome. “You have to make AI clusters as efficient as possible for the world to use all the AI applications at the right cost structure, at the right economics, for this to be successful,” Sadana said. “So we are participating by making that infrastructure layer a lot more efficient for cloud customers, or the hyperscalers, which, in turn, of course, gives the benefits to all of these software companies trying to run AI applications in these cloud companies.” Technical innovations: Beyond traditional networking In terms of what the company is actually building now, NextHop is developing specialized network switches

Read More »

Microsoft abandons data center projects as OpenAI considers its own, hinting at a market shift

A potential ‘oversupply position’ In a new research note, TD Cowan analysts reportedly said that Microsoft has walked away from new data center projects in the US and Europe, purportedly due to an oversupply of compute clusters that power AI. This follows reports from TD Cowen in February that Microsoft had “cancelled leases in the US totaling a couple of hundred megawatts” of data center capacity. The researchers noted that the company’s pullback was a sign of it “potentially being in an oversupply position,” with demand forecasts lowered. OpenAI, for its part, has reportedly discussed purchasing billions of dollars’ worth of data storage hardware and software to increase its computing power and decrease its reliance on hyperscalers. This fits with its planned Stargate Project, a $500 billion, US President Donald Trump-endorsed initiative to build out its AI infrastructure in the US over the next four years. Based on the easing of exclusivity between the two companies, analysts say these moves aren’t surprising. “When looking at storage in the cloud — especially as it relates to use in AI — it is incredibly expensive,” said Matt Kimball, VP and principal analyst for data center compute and storage at Moor Insights & Strategy. “Those expenses climb even higher as the volume of storage and movement of data grows,” he pointed out. “It is only smart for any business to perform a cost analysis of whether storage is better managed in the cloud or on-prem, and moving forward in a direction that delivers the best performance, best security, and best operational efficiency at the lowest cost.”

Read More »

PEAK:AIO adds power, density to AI storage server

There is also the fact that many people working with AI are not IT professionals, such as professors, biochemists, scientists, doctors, clinicians, and they don’t have a traditional enterprise department or a data center. “It’s run by people that wouldn’t really know, nor want to know, what storage is,” he said. While the new AI Data Server is a Dell design, PEAK:AIO has worked with Lenovo, Supermicro, and HPE as well as Dell over the past four years, offering to convert their off the shelf storage servers into hyper fast, very AI-specific, cheap, specific storage servers that work with all the protocols at Nvidia, like NVLink, along with NFS and NVMe over Fabric. It also greatly increased storage capacity by going with 61TB drives from Solidigm. SSDs from the major server vendors typically maxed out at 15TB, according to the vendor. PEAK:AIO competes with VAST, WekaIO, NetApp, Pure Storage and many others in the growing AI workload storage arena. PEAK:AIO’s AI Data Server is available now.

Read More »

SoftBank to buy Ampere for $6.5B, fueling Arm-based server market competition

SoftBank’s announcement suggests Ampere will collaborate with other SBG companies, potentially creating a powerful ecosystem of Arm-based computing solutions. This collaboration could extend to SoftBank’s numerous portfolio companies, including Korean/Japanese web giant LY Corp, ByteDance (TikTok’s parent company), and various AI startups. If SoftBank successfully steers its portfolio companies toward Ampere processors, it could accelerate the shift away from x86 architecture in data centers worldwide. Questions remain about Arm’s server strategy The acquisition, however, raises questions about how SoftBank will balance its investments in both Arm and Ampere, given their potentially competing server CPU strategies. Arm’s recent move to design and sell its own server processors to Meta signaled a major strategic shift that already put it in direct competition with its own customers, including Qualcomm and Nvidia. “In technology licensing where an entity is both provider and competitor, boundaries are typically well-defined without special preferences beyond potential first-mover advantages,” Kawoosa explained. “Arm will likely continue making independent licensing decisions that serve its broader interests rather than favoring Ampere, as the company can’t risk alienating its established high-volume customers.” Industry analysts speculate that SoftBank might position Arm to focus on custom designs for hyperscale customers while allowing Ampere to dominate the market for more standardized server processors. Alternatively, the two companies could be merged or realigned to present a unified strategy against incumbents Intel and AMD. “While Arm currently dominates processor architecture, particularly for energy-efficient designs, the landscape isn’t static,” Kawoosa added. “The semiconductor industry is approaching a potential inflection point, and we may witness fundamental disruptions in the next 3-5 years — similar to how OpenAI transformed the AI landscape. SoftBank appears to be maximizing its Arm investments while preparing for this coming paradigm shift in processor architecture.”

Read More »

Nvidia, xAI and two energy giants join genAI infrastructure initiative

The new AIP members will “further strengthen the partnership’s technology leadership as the platform seeks to invest in new and expanded AI infrastructure. Nvidia will also continue in its role as a technical advisor to AIP, leveraging its expertise in accelerated computing and AI factories to inform the deployment of next-generation AI data center infrastructure,” the group’s statement said. “Additionally, GE Vernova and NextEra Energy have agreed to collaborate with AIP to accelerate the scaling of critical and diverse energy solutions for AI data centers. GE Vernova will also work with AIP and its partners on supply chain planning and in delivering innovative and high efficiency energy solutions.” The group claimed, without offering any specifics, that it “has attracted significant capital and partner interest since its inception in September 2024, highlighting the growing demand for AI-ready data centers and power solutions.” The statement said the group will try to raise “$30 billion in capital from investors, asset owners, and corporations, which in turn will mobilize up to $100 billion in total investment potential when including debt financing.” Forrester’s Nguyen also noted that the influence of two of the new members — xAI, owned by Elon Musk, along with Nvidia — could easily help with fundraising. Musk “with his connections, he does not make small quiet moves,” Nguyen said. “As for Nvidia, they are the face of AI. Everything they do attracts attention.” Info-Tech’s Bickley said that the astronomical dollars involved in genAI investments is mind-boggling. And yet even more investment is needed — a lot more.

Read More »

IBM broadens access to Nvidia technology for enterprise AI

The IBM Storage Scale platform will support CAS and now will respond to queries using the extracted and augmented data, speeding up the communications between GPUs and storage using Nvidia BlueField-3 DPUs and Spectrum-X networking, IBM stated. The multimodal document data extraction workflow will also support Nvidia NeMo Retriever microservices. CAS will be embedded in the next update of IBM Fusion, which is planned for the second quarter of this year. Fusion simplifies the deployment and management of AI applications and works with Storage Scale, which will handle high-performance storage support for AI workloads, according to IBM. IBM Cloud instances with Nvidia GPUs In addition to the software news, IBM said its cloud customers can now use Nvidia H200 instances in the IBM Cloud environment. With increased memory bandwidth (1.4x higher than its predecessor) and capacity, the H200 Tensor Core can handle larger datasets, accelerating the training of large AI models and executing complex simulations, with high energy efficiency and low total cost of ownership, according to IBM. In addition, customers can use the power of the H200 to process large volumes of data in real time, enabling more accurate predictive analytics and data-driven decision-making, IBM stated. IBM Consulting capabilities with Nvidia Lastly, IBM Consulting is adding Nvidia Blueprint to its recently introduced AI Integration Service, which offers customers support for developing, building and running AI environments. Nvidia Blueprints offer a suite pre-validated, optimized, and documented reference architectures designed to simplify and accelerate the deployment of complex AI and data center infrastructure, according to Nvidia.  The IBM AI Integration service already supports a number of third-party systems, including Oracle, Salesforce, SAP and ServiceNow environments.

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »