Stay Ahead, Stay ONMINE

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year.

1. Agents: the next generation of automation

AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs.

At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved.

“Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail.

Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to pick out the best output to make a decision on. 

Another part of the secret sauce? Retrieval-augmented generation (RAG), which allows agents to store and reuse knowledge efficiently, is getting better. Imagine a travel agent bot that not only plans trips but books flights and hotels in real time based on updated preferences and budgets.

Takeaway: Businesses need to identify use cases where agents can provide high ROI — be it in customer service, sales, or internal workflows. Tool use and advanced reasoning capabilities will define the winners in this space.

2. Evals: the foundation of reliable AI

Evaluations, or “evals,” are the backbone of any robust AI deployment. This is the process of choosing which LLM — among the hundreds now available — to use for your task. This is important for accuracy, but also for aligning AI outputs with enterprise goals. A good eval ensures that a chatbot understands tone, a recommendation system provides relevant options, and a predictive model avoids costly errors.

For example, a company’s eval for a customer-support chatbot might include metrics for average resolution time, accuracy of responses, and customer satisfaction scores.

A lot of companies have been investing a lot of time into processing inputs and outputs so that they conform to a company’s expectations and workflows, but this can take a lot of time and resources. As models themselves get better, many companies are saving effort by relying more on the models themselves to do the work, so picking the right one gets more important.

And this process is forcing clear communication and better decisions. When you “get a lot more conscious of how to evaluate the output of something and what it is that you actually want, not only does that make you better with LLMs and AI, it actually makes you better with humans,” said Witteveen.  “When you can clearly articulate to a human: This is what I want, here’s how I want it to look like, here’s what I’m going to expect in it. When you get really specific about that, humans suddenly perform a lot better.” 

Witteveen noted that company managers and other developers are telling him: “Oh, you know, I’ve gotten much better at giving directions to my team just from getting good at prompt engineering or just getting good at, you know, looking at writing the right evals for models.”

By writing clear evals, businesses force themselves to clarify objectives — a win for both humans and machines.

Takeaway: Crafting high-quality evals is essential. Start with clear benchmarks: response accuracy, resolution time, and alignment with business objectives. This ensures that your AI not only performs but aligns with your brand’s values.

3. Cost efficiency: scaling AI without breaking the bank

AI is getting cheaper, but strategic deployment remains key. Improvements at every level of the LLM chain are bringing dramatic cost reductions. Intense competition among LLM providers, and from open-source rivals, is leading to regular price cuts.

Meanwhile, post-training software techniques are making LLMs more efficient.

Competition from new hardware vendors such as Groq’s LPUs, and improvements by the legacy GPU provider Nvidia, are dramatically reducing inference costs, making AI accessible for more use cases.

The real breakthroughs come from optimizing the way models are put to work in applications, which is the time of inference, rather than the time of training, when models are first built using data. Other techniques like model distillation, along with hardware innovations, mean companies can achieve more with less. It’s no longer about whether you can afford AI — you can do most projects much less expensively this year than even six months ago — but how you scale it.

Takeaway: Conduct a cost-efficiency analysis for your AI projects. Compare hardware options and explore techniques like model distillation to cut costs without compromising performance.

4. Memory personalization: tailoring AI to your users

Personalization is no longer optional — it’s expected. In 2025, memory-enabled AI systems are making this a reality. By remembering user preferences and past interactions, AI can deliver more tailored and effective experiences.

Memory personalization isn’t widely or openly discussed because users often feel uneasy about AI applications storing personal information to enhance service. There are privacy concerns, and the ick factor when a model spits out answers that show it knows a great deal about you — for example, how many kids you have, what you do for a living, and what your personal tastes are. OpenAI, for one, safeguards information about ChatGPT users in its system memory — which can be turned off and deleted, though it is on by default.

While businesses using OpenAI and other models that are doing this can not get the same information, what they can do is create their own memory systems using RAG, ensuring data is both secure and impactful. However, enterprises must tread carefully, balancing personalization with privacy.

Takeaway: Develop a clear strategy for memory personalization. Opt-in systems and transparent policies can build trust while delivering value.

5. Inference and test-time compute: the new efficiency and reasoning frontier

Inference is where AI meets the real world. In 2025, the focus is on making this process faster, cheaper and more powerful. Chain-of-thought reasoning — where models break down tasks into logical steps — is revolutionizing how enterprises approach complex problems. Tasks requiring deeper reasoning, like strategy planning, can now be tackled effectively by AI.

For instance, OpenAI’s o3-mini model is expected to be released later this month, followed by the full o3 model at a later date. They introduce advanced reasoning capabilities that decompose complex problems into manageable chunks, thereby reducing AI hallucinations and improving decision-making accuracy. These reasoning improvements work in areas like math, coding, and science applications where increased thought can help — though in other areas, like synthesizing language, advancements may be limited. 

However, these improvements will also come with increased computational demands, and so higher operational costs. The o3-mini is meant to provide a compromise offering to contain costs while keeping performance high.

Takeaway: Identify workflows that can benefit from advanced inference techniques. Implementing your own company’s special chain-of-thought reasoning steps, and selecting optimized models, can give you an edge here.

Conclusion: Turning insights into action

AI in 2025 isn’t just about adopting new tools; it’s about making strategic choices. Whether it’s deploying agents, refining evals, or scaling cost-efficiently, the path to success lies in thoughtful implementation. Enterprises should embrace these trends with a clear, focused strategy.

For more detail on these trends, check out the full video podcast between Sam Witteveen and myself here:

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

SonicWall rolls out eight new firewalls, expands cyber warranty

NSa 4800: 24x1GbE, 8x10G SFP+, 20 Gbps firewall throughput, 13 Gbps threat prevention NSa 5800: 24x1GbE, 8x10G SFP+, 30 Gbps firewall throughput, 24 Gbps threat prevention SonicWall Multi-gigabit connectivity addresses real market demand The Generation 8 portfolio introduces multi-gigabit connectivity across both product lines. Even lower-end desktop models now support

Read More »

ConocoPhillips lets well stimulation services contract for North Sea assets

@import url(‘https://fonts.googleapis.com/css2?family=Inter:[email protected]&display=swap’); a { color: #c19a06; } .ebm-page__main h1, .ebm-page__main h2, .ebm-page__main h3, .ebm-page__main h4, .ebm-page__main h5, .ebm-page__main h6 { font-family: Inter; } body { line-height: 150%; letter-spacing: 0.025em; font-family: Inter; } button, .ebm-button-wrapper { font-family: Inter; } .label-style { text-transform: uppercase; color: var(–color-grey); font-weight: 600; font-size: 0.75rem; } .caption-style

Read More »

Cisco strengthens AI networking story

“Overall, AI demand in the enterprise will grow over time. But enterprise customers need to see the value, see the ROI. Also, they have to have a well-defined use case,” Wollenweber said, noting that the 12-month innovation cycles of GPU vendors can be problematic if customers choose the wrong platform.

Read More »

DEF CON research takes aim at ZTNA, calls it a bust

Major vendor vulnerabilities span authentication and design flaws The research exposed critical vulnerabilities across Check Point, Zscaler and Netskope that fell into three primary categories: authentication bypasses, credential storage failures and cross-tenant exploitation. Authentication bypass vulnerabilities Zscaler’s SAML implementation contained the most severe authentication flaw. The researchers discovered that the

Read More »

Energy Secretary Chris Wright Visits Ames National Laboratory, Advancing U.S. Leadership in Critical Materials, Science, and Innovation

WASHINGTON— U.S. Secretary of Energy Chris Wright joined U.S. Senator Joni Ernst (R-IA), U.S. Representative Bob Latta (R-OH-05), U.S. Representative Bruce Westerman (R-AR-04), U.S. Representative Mariannette Miller-Meeks (R-IA-01), and U.S. Representative Ashley Hinson (R-IA-02) at the Ames National Laboratory to see firsthand the lab’s cutting-edge research and its pivotal role in addressing national energy and materials engineering challenges.  Secretary Wright’s visit included briefings and discussions with scientists and lab experts in key mission areas, including rare earth minerals, critical materials, AI, and next-generation energy technologies. Wright toured the Critical Materials Innovation Hub, a Department of Energy (DOE) Innovation Hub led by Ames, that is spearheading efforts to restore American energy dominance, secure domestic supply chains, reduce dependence on foreign actors, and strengthen the scientific capabilities essential to U.S. competitiveness. “Ames National Laboratory is leading the way in groundbreaking science and innovation that will strengthen America’s energy future,” said Secretary Wright. “By investing in critical minerals, advanced manufacturing, and AI, we are building more secure domestic supply chains—bolstering America’s national security and increasing access to affordable, reliable energy for all Americans. These bold steps position the U.S. not just to compete, but to lead, for generations to come.” “Secretary Wright is the fourth cabinet official to visit Iowa this week and it is an absolute breath of fresh air how accessible the Trump administration is compared to the previous four years,” said U.S. Senator Joni Ernst. “America’s national labs are not just critical to technological advancement but also national security. The Ames National Lab is a hub of innovation and will help ensure that our nation has the energy and critical materials to power the AI revolution.” “As Chairman of the Energy Subcommittee on the House Energy and Commerce Committee, I appreciated the opportunity to visit Ames National Laboratory with U.S.

Read More »

Energy Department Extends Emergency Order to Strengthen Grid Security in Puerto Rico

The continuation of emergency protocols ensures work can continue to address Puerto Rico’s long-standing grid reliability issues. WASHINGTON—The Department of Energy (DOE) today announced the extension of the emergency orders that are strengthening Puerto Rico’s electric grid. The orders, issued in May, have empowered Puerto Rico’s government to conduct commonsense fixes to restore the fragile grid and put Puerto Rico on track for sustained energy resilience. Extending the orders will ensure this critical work can continue throughout the peak summer demand season.  “A reliable and secure power grid is essential for modern life, and the residents of Puerto Rico deserve solutions now. Thanks to President Trump’s leadership, we are able to take action, moving from years of instability toward measurable, lasting progress,” said U.S. Secretary of Energy Chris Wright. “By extending these orders, DOE is ensuring critical work continues, urgent energy reliability needs are addressed, and the grid is more prepared to withstand the most demanding stretch of hurricane season for the 3.2 million Americans who call Puerto Rico home.”  “I thank Secretary Wright and strongly support the Department of Energy’s extension of the 202(c) emergency orders for Puerto Rico, which have provided needed flexibilities to maintain sufficient power generation capacity and conduct vegetation control activities along critical transmission lines. The Trump Administration’s close collaboration and commitment to stabilize and rebuild our power grid has been unprecedented and is already yielding results,” said Puerto RicoGovernor Jenniffer González-Colón.  “I look forward to building on this momentum and continue partnering with President Trump and Secretary Wright as we strengthen Puerto Rico’s electrical system and ensure an affordable, reliable, and secure supply of energy for the island’s 3.2 million Americans.”   Decades of deferred maintenance, insufficient investment, the bankruptcy of the system owner, and devasting hurricanes and earthquakes have significantly deteriorated the Puerto Rican electric grid.

Read More »

Exxon Aims to Build on Guyana Success in Trinidad and Tobago

Exxon Mobil Corp. signed a deal to explore for oil and gas off the coast of Trinidad and Tobago, a decade after making the world’s largest discovery in a generation in nearby Guyana.  The Texas oil giant was awarded a 7,165 square kilometer block named TTUD-1 about 190 miles (306 kilometers) off the coast of the Caribbean nation on Tuesday. Exxon has agreed to acquire 3D seismic data over a portion of the block as part of the production sharing contract, Prime Minister Kamla Persad-Bissessar said. The “mandatory” first phase is expected to cost the company $42.5 million, she said.  Exxon’s projected spend could rise to more than $20 billion if the company finds oil or gas in the ultra deep water block, according to Trinidad and Tobago Energy Minister Roodal Moonilal.  The oil major discovered more than 11 billion barrels of crude in Guyana after drilling the first exploration well in 2015, at a time when many of its peers balked at the prospect.  “We’re going to bring Exxon Mobil’s leading technology capability directly from Guyana,” said John Ardill, Exxon’s vice president for exploration.  The aim is to make Trinidad and Tobago the “energy capital” of the Caribbean once again, Persad-Bissessar said. Her government would accelerate the “responsible” exploration of unexplored land and marine acreages, continuously improve acreage award processes and undertake a comprehensive review of the oil and gas taxation regime to ensure her country remains an internationally competitive hydrocarbon province. To reduce risk, Trinidad and Tobago merged seven blocks into one during negotiations with Exxon. “Consolidation of the blocks removes risk in both exploration and development of the block,” the Energy Chamber of Trinidad and Tobago said. WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are

Read More »

Oil Climbs From Two Month Lows

West Texas Intermediate rose from a two-month low to settle above $63 a barrel in thin summer trading. Investors are watching for a signs that Friday’s summit between Donald Trump and Vladimir Putin will result in an easing or tightening of Washington’s sanctions on the OPEC+ member. Trump warned he would impose “very severe consequences” if Putin didn’t agree to a ceasefire, following a call with European leaders. The markets are largely in “wait-and-see mode,” said Rebecca Babin, a senior energy trader at CIBC Private Wealth Group. “Consensus is that we won’t see a definitive ceasefire or aggressive sanctions from Trump.” On the supply side, the picture is increasing bearish. Oil has lost more than 10% this year as the Organization of the Petroleum Exporting Countries and its partners complete the reversal of output curbs started in 2023, though price moves have been limited in recent days as trading activity eases over the summer months in the northern hemisphere. Expectations for a record supply glut next year are also weighing on the market. Supplies are already “primed to back up in the Atlantic Basin,” RBC analysts including Brian Leisen wrote in a note, adding that “crude oil balances for the rest of the year are weaker” than a year prior. Oil Prices WTI for September delivery rose 2.09% to settle at $63.96 a barrel in New York. Brent for October settlement rose 1.84% to settle at $66.84 a barrel. What do you think? We’d love to hear from you, join the conversation on the Rigzone Energy Network. The Rigzone Energy Network is a new social experience created for you and all energy professionals to Speak Up about our industry, share knowledge, connect with peers and industry insiders and engage in a professional community that will empower your career in energy.

Read More »

Nuveen raises $1.3B for energy and power infrastructure fund

Dive Brief: Nuveen, the investment manager for retirement services company TIAA, said Wednesday it completed a $1.3 billion funding round for its second energy and power infrastructure credit fund to support a growing global power demand due to artificial intelligence, digitalization and electrification.  The firm’s EPIC II fund will provide companies with credit solutions aimed at “supporting the build-out of secure and reliable energy and power generation while also focusing on credit opportunities involving sustainable infrastructure,” Wednesday’s release said. The fund has a $2.5 billion target and will take an all-of-the-above strategy to its energy investments, investing in renewables and energy storage technologies, along with liquefied natural gas and other fossil fuels. Dive Insight: Nuveen manages more than $1.3 trillion in assets, including more than $35 billion in infrastructure assets as of March 31, according to the release. Those assets include solar and battery storage in the United States. The EPIC II fund — which shares the same investment strategy and name as its predecessor EPIC I — will provide project and corporate financing to companies for equipment and growth capital, along with financing acquisitions, capital restructuring and other structured credit options. The senior managing director and portfolio manager of Nuveen’s energy infrastructure credit platform, Don Dimitrievich, said the fund is “focused on deploying capital into resilient companies and projects across the energy and power ecosystem” to capitalize on a “historic market opportunity.”  “Investors are increasingly interested in strategies that capitalize on their conviction in the growing global energy demand brought on by digitalization, electrification and reindustrialization while also seeking downside risk mitigation to guard against macro volatility, and inflationary and geopolitical risk,” Dimitrievich said in the release. The initial funding round announced Wednesday was “anchored” by TIAA and an unnamed “leading Canadian pension fund manager,” Nuveen said. Investors other

Read More »

Top India Oil Explorer’s Profit Falls

State-run Oil and Natural Gas Corp.’s quarterly profit declined on lower crude prices and stagnant production from its aging fields.  The New Delhi-based explorer’s net income fell 10% on year in the three months ended June 30 to 80.24 billion rupees ($915 million), according to a stock exchange filing. That was almost in-line with 80.74 billion rupees estimated by analysts, according to a Bloomberg survey.  The tepid earnings comes as the South Asian nation’s largest oil and gas producer is looking to cut dependence on exploration by diversifying into refining and liquefied natural gas. It is investing 2 trillion rupees on new energy offerings and emission control in its goal to become an integrated energy company.  ONGC’s weak June quarter performance contrasts with those of global peers like Exxon Mobil Corp. and Chevron Corp., which made up for weak crude prices with record production.  The company’s oil output increased just 1%, and gas production was flat as most of its older fields are witnessing a natural decline in performance. To boost production, it will focus on recent discoveries and enhance recovery from aging fields, Chairman Arun Kumar Singh said in ONGC’s annual report published last week. The company is pursuing deepwater collaboration with global giants like BP Plc, ExxonMobil and TotalEnergies SE to mitigate exploration risks in difficult regions, he said.  ONGC, which accounts for two-thirds of India’s oil and more than half the gas output, saw its quarterly revenue fall 9.3% to 320 billion rupees. Its crude oil earnings per barrel plunged 20.4% to $66.13 a barrel, while that from gas rose 2.2% to $6.64 per million British thermal units.  WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate

Read More »

Cisco’s 9% security growth is misleadingly low

AI infrastructure ahead of plan and growing Cisco reported more than $800 million in AI infrastructure orders from webscale customers in the fourth quarter, bringing the total for 2025 to more than $2 billion, which is over 2x what the company originally projected. This is a mix of its own Nexus switches, optics, AI PODs, UCS servers and Silicon One. Success here is critical for Cisco, as at one time, the company had next to no business with the hyperscalers. The development of Silicon One was pivotal in Cisco’s success with this audience, as it’s given them market leading price performance. Cisco has also cultivated a partnership with Nvidia and is the only company to have its silicon integrated into the GPU maker’s Spectrum-X product. There is another wave of business coming for Cisco in this area selling AI infrastructure to non-hyperscalers. Robbins talked about this: “The Cisco Secure AI Factory with Nvidia provides a blueprint for building AI-ready data centers for enterprises, sovereign cloud providers and newly emerging neocloud providers. We expect the sovereign AI opportunity to build momentum in the second half of fiscal year ’26.” AI will drive campus upgrades Most of the focus of network growth in AI has been in the data center, as that’s where the growth has been. However, the traffic agentic AI creates will drive campus upgrades as well. On the call, Cisco showed a chart of traffic generated by chatbots pre and post agentic, and it shows Cisco is expecting agentic to drive a consistent level of traffic that most networks are not able to handle. Robbin explained: “Network traffic will not only increase beyond the peaks of current chatbot interaction but will remain consistently high with agents in constant interaction.” The impact of this is twofold. The bump in traffic

Read More »

Nvidia targets data center with new servers, AI software

More Nvidia news from SIGGRAPH The RTX Pro server wasn’t the only news at the show. Nvidia also announced two new models of its Nemotron model family – the Nemotron Nano 2 and Llama Nemotron Super 1.5 – with advanced reasoning capabilities for building smarter AI agents. Nemotron is a family of enterprise-ready, open large language models (LLM) designed to enhance agentic AI for tasks requiring sophisticated reasoning, instruction following, coding, tool use, and multimodal (text+vision) understanding. These models deliver high accuracy for their relative size in areas such as scientific reasoning, coding, tool use, instruction following and chat, according to Nvidia. They are designed to imbue AI agents with deeper cognitive abilities and help AI systems explore options, weigh decisions and deliver results within defined constraints. Nvidia claims Nemotron Nano 2 achieves up to six times higher token generation throughput compared to other models its size. Llama Nemotron Super 1.5 offers top-tier performance and leads in reasoning accuracy, making it suitable for handling complex enterprise tasks. Also, Nvidia is empowering robotics and machines to “see” and react to what they see with new AI models that can ingest visual information and think about said information. The vendor just announced Cosmos Reason, a new open, customizable 7 billion-parameter reasoning Vision Language Models, or VLMs. VLMs allow robots and vision agents to think about what they see, just like a human. Up to now, robots have had the ability to “see,” but their reaction to what they saw was extremely limited. A VLM provides robotics with the ability to think about their actions.

Read More »

Cisco Q4 results: AI infrastructure orders surpass goal

Adding a little more color to the enterprise results, Robbins called out some of Cisco’s network modernization offerings for customers. “We’ve got [our] routing refresh, we’ve got a lot of new technology in our data center networking business … we have Wi-Fi 7 gear, which grew triple digits year over year. We’re in year eight of the [Catalyst 9000] offering, and honestly, if you look at products that were pre-[Catalyst 9000] that are still installed in our customer base, there’s tens of billions of dollars of install base there that we can go after,” Robbins said. Demand for AI infrastructure will extend to enterprises, Robbins said: “If you think about the AI revolution…We tend to see these things begin first in the cloud providers, which we’re clearly seeing the AI play in the cloud providers. Then we see it shift into the enterprise. We see a shift in the back end, in this case, from the back end to the front end. We believe that will occur as enterprises start using more of these services, and then enterprises will also build out inferencing… and we’re even seeing the telco business actually pick up as they’re telling us they’re increasing their network capacity and they’re modernizing their infrastructure in preparation for AI. So, we think AI is going to drive network modernization across all of these segments.” Taking aim at security Overall security order growth in Q4 was up double digits, Robbins said. “We have 80 new Hypershield customers, largely connected to this new smart switch [the N9300 Smart Switch]. So that strategy is working. And I would say that we had 480-plus new SSE customers during the quarter. So that’s our secure services edge [which] is really getting good traction. Robbins singled out the demand for integrated networking and security

Read More »

Uptime Institute’s Jay Dietrich on Why Net Zero Isn’t Enough for Sustainable Data Centers

In the latest episode of the Data Center Frontier Show podcast, Editor-in-Chief Matt Vincent sits down with Jay Dietrich, Research Director of Sustainability at Uptime Institute, to examine what real sustainability looks like inside the data center, and why popular narratives around net zero, offsets, and carbon neutrality often obscure more than they reveal. Over the course of our conversation, Dietrich walks listeners through Uptime’s expanding role in guiding data center operators toward measurable sustainability outcomes; not just certifications, but operational performance improvements at the facility level. “Window Dressing” vs. Real Progress Dietrich is candid about the challenges operators face in navigating the current landscape of sustainability reporting. Despite high-level claims of carbon neutrality, many facilities still operate inefficiently, relying heavily on carbon offsets or energy attribute certificates to hit corporate goals. “An EU survey found that 80% of data centers report carbon-free operations based on market calculations, while their national grids run at only 55% renewable,” Dietrich says. “The only thing that truly matters is the performance of the actual facility.” To close this gap, Uptime offers a Sustainability Gap Analysis and a Sustainable Operations Certification, helping data center operators minimize energy and water use, improve cooling efficiency, and increase the useful work delivered per megawatt hour. Redefining the Sustainable Data Center One of the discussion’s core messages: a net zero data center is not necessarily a sustainable one. Dietrich stresses the need to shift focus from corporate carbon accounting toward IT utilization, emphasizing metrics like: Work delivered per unit of energy consumed. Work delivered per metric ton of CO₂ emitted (location-based). Actual IT infrastructure utilization rates. Underutilized IT infrastructure — still common across the industry — is one of the biggest sustainability blind spots. “Running IT at 10% utilization wastes capacity, space, and energy,” says Dietrich. “Increasing that

Read More »

Data Center Jobs: Engineering, Construction, Commissioning, Sales, Field Service and Facility Tech Jobs Available in Major Data Center Hotspots

Each month Data Center Frontier, in partnership with Pkaza, posts some of the hottest data center career opportunities in the market. Here’s a look at some of the latest data center jobs posted on the Data Center Frontier jobs board, powered by Pkaza Critical Facilities Recruiting. Peter Kazella of Pkaza Critical Facilities Recruiting provides tips for building a healthy applicant pool. Switchgear Field Service Technician – Critical Facilities Nationwide Travel  This position is also available in any major data center region: Ashburn, VA; Charlotte, NC; Atlanta, GA; Denver, CO; Portland, OR; Seattle, WA; Las Vegas, NV; or Phoenix, AZ. Multiple opportunities for both senior and mid-level switchgear field service technicians. These openings are with a nationwide market leader of power distribution solutions, specializing in switchgear and controls for mission-critical environments. This company provides customized solutions for enterprise, colocation, and hyperscale data centers, ensuring reliability and uptime through controls integration, power distribution solutions, and switchgear installations. Their services include installations, retrofits, upgrades, turnkey electrical solutions, and preventive & corrective maintenance of UPS, switchgear, generators, and PLC systems. This is an excellent career-growth opportunity to work on exciting projects with leading-edge technology and competitive compensation. Electrical Commissioning Engineer New Albany, OH This traveling position is also available in: Richmond, VA; Ashburn, VA; Charlotte, NC; Atlanta, GA; Hampton, GA; Fayetteville, GA; Minneapolis, MN; Phoenix, AZ; Dallas, TX; or Chicago, IL. *** ALSO looking for a LEAD EE and ME CxA agents and CxA PMs *** Our client is an engineering design and commissioning company that has a national footprint and specializes in MEP critical facilities design. They provide design, commissioning, consulting and management expertise in the critical facilities space. They have a mindset to provide reliability, energy efficiency, sustainable design and LEED expertise when providing these consulting services for enterprise, colocation and hyperscale companies. This career-growth

Read More »

DCF Trends Summit 2025: Power Chat

Data Center Frontier Editor in Chief Matt Vincent and Contributing Editor Bill Kleyman (CEO/Apolo) recently had another video chat to discuss the Data Center Frontier Trends Summit 2025 and, in particular, the event’s focus on data center power. The second annual Trends Summit is scheduled for August 26-28 in Reston, Virginia. Register Now This second QuickChat in our series—following the opening discussion that set the stage for the conference—now zeroes in on data center power challenges. The conversation explores how AI’s exponential growth is straining power infrastructure, turning energy into both a critical market-entry barrier and a defining theme of the event. The Power Challenge: Quantifying the Scale Kleyman outlined the sheer scale of the power challenge facing the industry. He cited a Goldman Sachs report projecting that U.S. data center power consumption, currently around 2-3% of total consumption, could more than double to 8-9% by 2028. He also highlighted a forecast that the data center industry will require upwards of 50 gigawatts (GW) of power by 2035, a figure he put into perspective by noting that 1 GW can power a city of a million people. The discussion mentions a few large-scale projects that illustrate this trend: Kevin O’Leary‘s Wonder Valley project in Alberta, Canada, aiming for 8 GW. Tract’s new multi-gigawatt campus project in Texas amid plans for a staggering 25 GW national build-out. Kleyman also referenced the NVL576 rack unveiled at Nvidia GTC 2025, which is capable of supporting 600 kilowatts (kW) per rack, signaling the official arrival of the “megawatt era class of data center racks.” Grid-Optional Solutions and the Summit Panels The discussion transitioned to how the industry is moving from being solely grid-reliant to “grid-optional.” A significant trend highlighted is the inclusion of behind-the-meter power generation in new builds, with 30% of new U.S. data

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »