Stay Ahead, Stay ONMINE

Independent power providers hit back at utility critics over PJM price surge

A year ago, capacity prices spiked in the PJM Interconnection. Last month they hit another record high of nearly $330/MW-day. Companies including Exelon, FirstEnergy and PPL Corp., which have utilities in states that bar them from owning generation, have been pressing for state legislation that would lift that restriction. They contend that the jump in PJM capacity […]

A year ago, capacity prices spiked in the PJM Interconnection. Last month they hit another record high of nearly $330/MW-day.

Companies including Exelon, FirstEnergy and PPL Corp., which have utilities in states that bar them from owning generation, have been pressing for state legislation that would lift that restriction. They contend that the jump in PJM capacity prices is increasing customer bills but failing to spur independent power producers to build power plants.

To get a nonutility perspective, on Monday Utility Dive talked with Todd Snitchler, president and CEO of the Electric Power Supply Association, a trade group for independent power producers, about how power providers are moving to bring electricity supplies to PJM, what’s behind rising electric bills and the challenges all companies face in building power plants.

EPSA members own and operate about 175,000 MW of capacity in U.S. regions with access to competitive wholesale electricity markets. Members include Invenergy, LS Power, NRG Energy, Talen Energy, Tenaska and Vistra.

This interview has been edited for length and clarity.

UTILITY DIVE: Have we seen a response by IPPs to increased capacity prices in PJM?

TODD SNITCHLER: I think you are seeing a response from the IPP or competitive generator sector in response to price signals, but not exclusively that. 

You saw a significant response to the Reliability Resource Initiative process at PJM, which tried to accelerate projects to the front of the line that were ready to go or could be constructed and operational by 2030. Ninety-four projects were submitted and 51 selected. It was close to 10,000 MW that is supposed to come online by 2030.

And that was just a part of how independent producers tried to respond, before the second price signal was even sent in the July auction, which I think we would all agree continued to send the signal for new investment.

One of the challenges that we’re seeing is this compressed auction schedule. There hasn’t been an auction that’s been on time in several years now. So you’ve got these six-month increments, which really make it difficult for generators or investors to make a decision that says, “Okay, let’s move this project from development into ‘we’re going to move forward’ in that short of a time horizon.”

Normally, these auctions are spaced a year apart — it gives people an opportunity to respond.

Even so, you’re seeing people make affirmative decisions about investments, delaying retirement, projects that were not moving forward are now back on the drawing board. You see things like the Crane Clean Energy Center, formerly Three Mile Island … you are seeing that come back. So I think there are all kinds of indicators that show that the market is responding to the signals that are being sent.

Would it be helpful if we could move faster? Sure, it would be. Is that always possible? Honestly, no. We’ve got issues outside of queue reform, which everyone likes to blame, and which I think is kind of behind us, if you look at the statistics

But there are supply chain issues that affect everybody, whether you’re a vertically integrated utility, or you’re a competitive power generator, or you’re a rural co-op or public power. We’re all subject to the same issues with regard to supply chain.

And there’s a workforce issue that we’re going to have to address because, assuming we all agree … that we’re going to need to build a whole lot of new megawatts in the coming near-term time horizon, everybody’s competing for the workforce that’s skilled and able to construct those facilities.

I hear some of our utility critics say, “Well, if you just let us do it, we’d have this problem solved.” Well, I’d like to know how you’re able to skip ahead and have your equipment, whether it’s switchgears or transformers or turbines. We’re all competing for the same equipment on the same time horizon, so it’s a little bit difficult to understand how they could do it so much faster, so much cheaper, when that’s historically never been the case.

If somebody were to order a turbine today, how long would it take to get it?

It varies, but conventional wisdom right now is it’s probably five years, maybe six. We’ve seen some discussion about how we are going to address additional manufacturing capability, and I think we’ll see that. But you don’t build a new factory overnight, and you certainly don’t turn out these pieces of equipment in a week or two.

What about getting fuel supply for gas-fired power plants? I’ve heard pipelines are very tight. Storage is tight.

It varies by region. So if you’re New England, you’ve got a different set of issues than if you’re in central Pennsylvania or Ohio, where you’re sitting on top of the gas, where it’s being produced, and it’s much easier to move around.

We are going to need to see some additional pipeline capacity that’s going to take gas from where it’s being produced into other parts of the country where they are bottlenecked, whether that’s the Southeast or the Northeast.

What about the cost of building a new gas-fired power plant? I’m seeing roughly $2,200 to $2,500 per kilowatt.

That range sounds pretty close to what we have heard, which is much more than it was three or five years ago. From our members’ perspective, that’s cost that they bear, that’s not anything that’s put on the backs of ratepayers. In a vertically integrated environment in contrast, that $2.2 billion to $2.5 billion for 1,000 megawatts has to be paid by someone, and that’s the captive utility customer with a non-bypassable charge on the bill. And so that’s a direct pass-through to customers, plus the rate of return that the utility gets, where, if our members build it, they’ve got to make sure that they build it as inexpensively but as reliably as they can so they can operate, and that has the effect of driving down costs for consumers.

How are your members navigating this new import tariff landscape?

We have some members that are in the process of trying to get equipment into the United States and not sure what the price will be when it hits the shore based on the tariff rate with the countries that it may be coming from. For one member, the price variation was potentially 40% higher. So if you’ve got a $100 million piece of equipment, and the price is either $100 million or $140 million, that’s meaningful dollars. That is an area that our members are hoping that we can get some greater degree of certainty because if we’re going to achieve the administration’s objective of beating China at AI, we’re going to need more energy. More energy that is going to be produced with the same turbines that potentially are exposed to some of the tariff-related issues. There is some disruptive effect to that, but I think the administration is aware of it.

How is EPSA thinking about the issue of affordability?

We are one of the best ways for people to mitigate those costs by avoiding non-bypassable charges and letting us be the ones who put the risk on shareholders and investors for the generation portion of your bill.

Transmission and distribution charges are up substantially over the last 10 years, 15 years, and those costs are real and have translated into increased costs on consumers’ bills. So I find it a little disingenuous for some of the utilities that have had the transmission and distribution charges go up substantially over that period of time to say that this one capacity auction and now a second capacity auction is what’s driving up customer bills, and that’s why they should be allowed back in the [generation] game, because if you look at it in raw data numbers, the facts just don’t support their thesis.

In fact, the generation portion is flat or declining, and it has been that way as a result of both fuel switching, improved efficiency and new resources that came onto the system. But because of the most recent auction results, I think it’s a useful tool for those people who are trying to push responsibility for price increases off of their portion of the bill and onto someone else. 

Do you have concerns about politicians feeling the pressure around affordability and asking themselves if restructuring is getting us more needed generation?

We’re always mindful of the need for people to do something. Sometimes that means doing something that’s not necessarily good policy. We’re trying to educate policymakers and legislators about what markets are actually doing, what they have actually delivered.

I’m still hard-pressed to find an example where a vertically-integrated utility has delivered a new generating project on time and under budget and had any accountability in the way that wholesale power generators do.

Some people have been pointing to the consolidation in the power sector as a sign that the markets aren’t working. In PJM, people are preferring to just buy existing gas-fired power plants. What’s behind the consolidation? 

Our members all have different perspectives that they are operating under. Some are typically more operator-driven and others are more development-driven. Those that are in the development space are working on a number of projects in PJM as we speak, so I would urge a little caution from those who are suggesting that there’s no new development in gas.

Some companies are trying to acquire resources, to build their portfolio, because those entities have a different investment thesis, which is, “We’re operators, not developers.”

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Intel details new efficient Xeon processor line

The new chips will be able to support up to 12-channel DDR5 memory with speeds of up to 8000 MT/s, a substantial increase over the 8 channels of 6400MT/s in the prior generation. In addition to that, the platform will support up to 6 UPI 2.0 links with up to

Read More »

Eni Progresses Permit Process for Its 2nd Biorefinery in Sicily

Eni SpA said Thursday it had launched the authorization process for an Environmental Impact Assessment to build a biorefinery and a chemical recycling plant in Sicily, after receiving approval from Italy’s Environment and Energy Security Ministry. Designed to produce up to 500,000 metric tons per annum (MMtpa) of hydrogenated vegetable oil (HVO) diesel and sustainable aviation fuel (SAF), the Priolo project will rise on the site of an Eni ethylene plant set to be decommissioned, the state-backed energy major said in a press release. It will be Eni’s second biorefinery in Sicily, after the Gela plant, which began biofuel production 2019 after being transformed from a petrochemical site, according to Eni. The Priolo biorefinery will be fed mainly by vegetable waste and oils and animal fats, Eni said. Besides the Ecofining™ plant, the project will include a biogenic feedstock pre-treatment unit and a hydrogen production plant, it said. Meanwhile the project’s chemical recycling unit will have a processing capacity of 40,000 MMtpa and produce around 32,000 MMtpa of pyrolysis oil, Eni said. The recycling plant will use the proprietary Hoop technology of Eni’s chemicals business Versalis SpA, based on the pyrolysis of mixed plastic packaging waste. The Priolo project will be the first deployment of the technology at a commercial scale, after the decommissioning of a demonstration plant at the Versalis plant in Mantua in June, Eni said. “This technology, which complements mechanical recycling, transforms mixed plastic waste into raw material for the production of new plastic products suitable for all applications, including food contact and pharmaceutical packaging”, Eni said. Developed from a joint project with Italian engineering company Servizi di Ricerche e Sviluppo SRL, the technology has high material recovery yields and feedstock flexibility, according to Eni. Last month Eni said it had received approval from the ministry to convert

Read More »

Companies Paying Record Sums to Develop Geothermal Energy

For the first time in years, the US federal government has leased every parcel of public land it has opened for geothermal development — and at record prices.  Average prices for public land leases for geothermal energy development soared by 282% this year to $127 per acre, driven by growing energy demand from data centers and President Donald Trump’s administration embracing the energy source. The record prices are major step up from $33 per-acre average last year, according to data from the US Bureau of Land Management, the agency that conducts the auctions. Geothermal companies tap the Earth’s heat to generate power. While it currently generates a small share of all US electricity, a wave of new technologies is opening the door for geothermal to play a bigger role, as is policy support.  “With the current administration, I think the focus is very much on energy security, energy reliability and energy independence,” said Koenraad Beckers, geothermal energy lead at ResFrac, a hydraulic fracturing simulation software designer. “Geothermal checks all those boxes.” The Trump administration has stymied wind projects and scaled back solar incentives, but geothermal is one type of renewable energy that’s seen some support. In February, Energy Secretary Chris Wright directed his agency to “prioritize affordable, reliable, and secure energy technologies,” including geothermal. The tax law Trump signed in July also preserved geothermal projects’ access to tax credits while weakening incentives for wind and solar. Artificial intelligence’s voracious appetite for energy has also driven interest in geothermal to new highs. Alphabet Inc.’s Google and Meta Platforms Inc. are among the data center operators who have inked deals with geothermal companies to provide much-needed power. Advances in geothermal drilling techniques are also helping give the industry new life. Companies are deploying methods traditionally used to extract oil and gas, including fracking, to use the Earth’s heat. New technology is making the process cheaper and easier to deploy, while also expanding the areas suitable

Read More »

WTI Falls Below $59 on Tariff Threats

Oil sunk to the lowest since May amid a US threat to slap new tariffs on China and as tensions ease in the Middle East, while global supply surplus concerns underpinned bearish gloom. West Texas Intermediate crude prices slumped 4.2% to settle just below $59 a barrel after US President Donald Trump said he saw “no reason” to meet Chinese President Xi Jinping and threatened a “massive increase” of tariffs on goods from China. The statements revived concerns that a tariff war between the world’s two largest economies would hurt oil consumption. They also blunted appetite for risk assets, hurting equities. “If Trump were to follow through on these latest threats, it would result in negative economic consequences and a hit to crude oil and refined product demand,” said John Kilduff, founding partner of Again Capital LLC. Traders had largely written off the possibility of extreme US tariffs against China following an amicable exchange between Trump and his Chinese counterpart, after which Trump said he would meet Xi on the sidelines of the Asia-Pacific Economic Cooperation summit later this month. The about-face in sentiment comes as China slapped port fees on American ships in retaliation for similar US measures, a move that threatened to bolster freight rates. Commodity trading advisers, which can accelerate price momentum, liquidated long positions to sit at 91% short in WTI on Friday, compared with 55% short on Oct. 9, according to data from Bridgeton Research Group. Brent was holding at a similar level, the group added. “Crude is facing a triple whammy today — renewed tariff tensions weighing on the demand outlook, a broader selloff in risk assets keeping dip buyers on the sidelines, and systematic strategies likely adding to short positions,” said Rebecca Babin, a senior energy trader at CIBC Private Wealth Group. “Without

Read More »

Kyiv Power Cut as Russia Steps Up Strikes

Russia hit Kyiv with a massive drone and missile attack that left parts of the capital without power and disrupted water supply as Moscow targets energy infrastructure ahead of winter. Power was out in several city districts and at least 12 civilians were injured, Kyiv Mayor Vitali Klitschko said in a Telegram post on Friday. He added that firefighters extinguished a blaze in a multistory building in the city center which was hit by drone debris. The attack also disrupted water supply to people’s homes, Klitschko said. The power loss mainly affected the east of the city and a central district while the blackout disrupted the operations of Kyiv’s metro, the city administration said on Telegram. Trains were unable to operate on a line running from downtown to the left bank of the River Dnipro. “It was a cynical and calculated attack, with more than 450 drones and over thirty missiles targeting everything that sustains normal life, everything the Russians want to deprive us of,” Ukraine’s President Volodymyr Zelenskiy said in a post on X.  Power outages were reported in a number of cities from Donetsk in the east, Odesa in the south, Chernihiv in the north, Zelenskiy said. A child was killed in the city of Zaporizhzhia and 20 people were injured elsewhere across the country, according to the president. Russia has significantly increased its attacks on Ukraine’s power infrastructure as the cold season approaches. That’s made disruptions to electricity a regular problem in regions bordering Russia, but it’s been a while since the capital suffered a blackout similar to that experienced in the winter season of 2022-2023. Ukraine’s air-defense has intercepted 420 drones and missiles of 497 that were launched, it said on Telegram. Those downed included Russia’s hypersonic missile known as Kinzhal. Missiles and drones hit 19

Read More »

The week in 5 numbers: Utilities make it rain, wildfire costs rage on

The amount Solar For All funding plaintiffs say the federal government is illegally withholding in a lawsuit filed this week in the U.S. District Court for the District of Rhode Island. The plaintiffs argue that the executive branch did not have the statutory authority to rescind the funding, even after Congress repealed large parts of the Inflation Reduction Act, as Congress “did not repeal the Solar for All program retroactively.”

Read More »

Colombian Industry Braces for Gas Rationing

Colombian companies are bracing for gas rationing as the nation’s sole import terminal closes for scheduled maintenance, casting a spotlight on the Andean country’s widening supply deficit.   In a bid to avoid blackouts on the populous Atlantic coast, the Energy Ministry has said it will prioritize supply to local gas-fired power plants, households and small businesses during the Oct. 10-14 outage of the SPEC liquefied natural gas terminal outside Cartagena.  While the LNG import facility pauses operations, around one-third of industrial gas demand will be unmet, according to industry trade group Asoenergía. Although the maintenance is brief, some factories have furnaces that require several days to completely shut off and days more to restart.  The scarcity is pitting Colombia’s biggest industrial and commercial energy users against power generators.  The government prefers to allocate gas supply cuts to industries even though power generators should have timely backups in place that they are paid to maintain, said Sandra Fonseca, head of Asoenergía. “This hurts industrial output,” she said in a written response to questions, adding that the companies won’t be compensated for the shortfall.  The trade group representing power generators, known as Andeg, declined to comment.  Colombia began importing LNG in 2016 to occasionally fuel thermal power plants as recurring droughts restricted the country’s mainstay hydroelectric units.  Then at the end of 2024, Colombia resorted to LNG to meet demand from households and businesses as domestic gas production declined.  The Andean nation, which is now facing a structural gas shortfall of around 17% of total demand, will see that gap widen to as much as 20% next year, according to trade group Naturgas. If by 2029 no new domestic sources of gas are added, the shortfall could reach as high as 56%, the group says. While Colombia’s oil and gas reserves

Read More »

Cisco seriously amps-up Silicon One chip, router for AI data center connectivity

Some say deep buffers shouldn’t be used to handle this type of traffic; the contention is that these buffers fill and drain, creating jitter in the workloads, and that slows things down, Chopra told Network World. “But the real source of that challenge is not the buffers. It’s a poor congestion management scheme and poor load balancing with AI workloads, which are completely deterministic and predictable. You can actually proactively figure out how to place flows across the network and avoid the congestion,” he said. The 8223’s deep-buffer design provides ample memory to temporarily store packets during congestion or traffic bursts, an essential feature for AI networks where inter-GPU communication can create unpredictable, high-volume data flows, according to Gurudatt Shenoy, vice president of Cisco Provider Connectivity. “Combined with its high-radix architecture, the 8223 allows more devices to connect directly, reducing latency, saving rack space, and further lowering power consumption. The result is a flatter, more efficient network topology supporting high-bandwidth, low-latency communication that is critical for AI workloads,” Shenoy wrote in a blog post. NOS options Notably, the first operating systems that the 8223 supports are the Linux Foundation’s Software for Open Networking in the Cloud (SONiC) and Facebook open switching system (FBOSS) – not Cisco’s own IOS XR.  IXR will be supported, too, but at a later date, according to Cisco.  SONiC decouples network software from the underlying hardware and lets it run on hundreds of switches and ASICs from multiple vendors while supporting a full suite of network features such as Border Gateway Protocol (BGP), remote direct memory access (RDMA), QoS, and Ethernet/IP. One of the keys to SONiC is its switch-abstraction interface, which defines an API to provide a vendor-independent way of controlling forwarding elements such as a switching ASIC, an NPU, or a software switch in a uniform

Read More »

Utilities Race to Meet Surging Data Center Demand With New Power Models

Over the last 18 months or so, the energy generation industry and its public utilities have been significantly impacted by the AI data center boom. It has been demonstrated across North America that the increase in demand for power, as driven by the demand for hyperscale and AI data centers, greatly exceeds the ability of the industry to actually generate and deliver power to meet the demand. We have covered many of the efforts being made to control the availability of power. In response, utilities and regulators have begun rethinking how to manage power availability through means such as: temporary moratoriums on new data center interconnections; the creation of new rate classes; cogeneration and load-sharing agreements; renewable integration; and power-driven site selection strategies.  But the bottom line is that in many locations utilities will need to change the way they work and how and where they spend their CAPEX budgets. The industry has already realized that their demand forecast models are hugely out of date, and that has had a ripple effect on much of the planning done by public utilities to meet the next generation of power demand. Most utilities now acknowledge that their demand forecasting models have fallen behind reality, triggering revisions to Integrated Resource Plans (IRPs) and transmission buildouts nationwide. This mismatch between forecast and actual demand is forcing a fundamental rethink of capital expenditure priorities and long-term grid planning. Spend More, Build Faster Utilities are sharply increasing CAPEX and rebalancing their resource portfolios—not just for decarbonization, but to keep pace with multi-hundred-megawatt data center interconnects. This trend is spreading across the industry, not confined to a few isolated utilities. Notable examples include: Duke Energy raised its five-year CAPEX plan to $83 billion (a 13.7% increase) and plans to add roughly 5 GW of natural gas capacity

Read More »

Duos Pairs Mobile Power and Modular Edge Data Centers for Rapid Texas Rollout

Duos Technology Group has launched the fifth of its AI edge data centers, part of a plan to deploy 15 units by the end of 2025. The projects are executed through Duos Edge AI, a subsidiary focused on modular, rapidly installed edge data centers (EDCs) in underserved markets, beginning with school districts and regional carrier hubs across Texas. The newest site is being deployed on-premises with the Dumas Independent School District in Dumas, Texas. High-Density Edge Design Duos’ EDCs emphasize very high rack densities (100 kW+ per rack), SOC 2 Type II compliance, N+1 power with dual generators, and a 90-day build/turn-up cycle. Each site is positioned approximately 12 miles from end users, cutting latency for real-time workloads. To meet the power demands of these edge deployments, Duos formed Duos Energy and partnered with Fortress/APR Energy to deliver behind-the-meter mobile gas turbines. This approach allows compute to go live in 90 days without waiting years for utility interconnection upgrades. The goal is straightforward: move power and compute close to demand, with rapid deployment. Duos’ modular pods are designed for exurban and rural locations as localized compute hubs for carriers, schools, healthcare systems, and municipal users. The rugged design pairs high-density racks with the short deployment cycle and proximity targeting, enabling a wide range of applications. With Dumas ISD now live, Duos has five sites in Texas, including Amarillo/Region 16, Victoria/Region 3, Dumas ISD, and multiple Corpus Christi locations. Mobile Power vs. Modular Compute While Duos doesn’t consistently describe its data center units as “mobile,” they are modular and containerized, engineered for rapid, site-agnostic deployment. The “mobile” label more explicitly applies to Duos’ power strategy—a turbine fleet that can be fielded or re-fielded to match demand. From an operator’s perspective, the combined proposition functions like a mobile platform: pre-integrated compute pods

Read More »

Report: AMD could be Intel’s next foundry customer

[ Related: More Intel news and insights ] AMD has lagged behind Nvidia in the AI business but has done well in the federal supercomputing business, holding numerous top spots with supercomputers like El Capitan and Frontier. Manufacturing its chips in the United States would be a good way to get the Trump administration off its back given its push for domestic manufacturing of semiconductors. The Trump administration is pushing for 50% of chips sold in America to be manufactured domestically, and tariffs on chips that are not. It also faces outbound restrictions. Earlier this year, AMD faced export restrictions GPUs meant for China as part of U.S. export controls against China’s AI business. “I believe this is a smart move by AMD to secure capacity in the local market without fighting against Nvidia and Apple and their deeper pockets for the limited capacity at TSMC,” said Alvi Nguyen, senior analyst with Forrester Research.” With the US investment in Intel, followed by Nvidia, this is can be seen as diversifying their supply chain and providing cheaper, locally sourced parts.” For Intel, this will continue a streak of good news it has enjoyed recently. “Having customers take up capacity at their foundries will go a long way in legitimizing their semiconductor processes and hopefully create the snowball effect of getting even more US-based customers,” said Nguyen. In recent weeks, Intel has partnered with Nvidia to jointly make PC and data center chips. Nvidia also took a $5B stake in Intel. Earlier the Trump administration made a $11.1B, or 10%, stake in Intel.

Read More »

AI Infra Summit 2025 Insights: AI Factories at the Core of the Fifth Industrial Revolution

NVIDIA’s AI Factory: Engineering the Future of Compute NVIDIA’s keynote illuminated the invisible but indispensable heart of the AI revolution—the AI factory. This factory blends hardware and software innovation to achieve performance breakthroughs that transcend traditional limits. Technologies such as disaggregated rack-scale GPUs and the novel 4-bit floating point numerical representation move beyond incremental improvements; they redefine what is achievable in energy efficiency and cost-effectiveness. The software ecosystem NVIDIA fosters, including open-source frameworks like Dynamo, enables unprecedented flexibility in managing inference workloads across thousands of GPUs. This adaptability is crucial given the diverse, dynamic demands of modern AI, where workloads can fluctuate dramatically in scale and complexity. The continuous leap in benchmark performance, often quadrupling hardware capabilities through software alone, continues to reinforce their accelerated innovation cycle. NVIDIA’s framing of AI factories as both technology platforms and business enablers highlights an important shift. The value computed is not merely in processing raw data but in generating economic streams through optimizing speed, reducing costs, and creating new AI services. This paradigm is central to understanding how the new industrial revolution will operate through highly efficient AI factories uniting production and innovation. AWS and the Cloud’s Role in Democratizing AI Power Amazon Web Services (AWS) represents a key pillar in making AI capabilities accessible across the innovation spectrum. AWS’ focus on security and fault tolerance reflects maturity in cloud design, ensuring trust and resilience are priorities alongside raw compute power. The evolution towards AI agents capable of specification-driven operations signifies a move beyond traditional computing paradigms towards contextual, autonomous AI services embedded deeply in workflows. Their launch of EC2 P6-B200 instances with next-generation Blackwell GPUs and specialized Trainium chips represents a continual drive to optimize AI training and inference at scale and out-of-box improvements in performance of 85% reduction in training time;

Read More »

Data Center Jobs: Engineering, Construction, Commissioning, Sales, Field Service and Facility Tech Jobs Available in Major Data Center Hotspots

Each month Data Center Frontier, in partnership with Pkaza, posts some of the hottest data center career opportunities in the market. Here’s a look at some of the latest data center jobs posted on the Data Center Frontier jobs board, powered by Pkaza Critical Facilities Recruiting. Looking for Data Center Candidates? Check out Pkaza’s Active Candidate / Featured Candidate Hotlist Business Development Manager – Mechanical Data Center Solutions Remote / traveler This position can be located anywhere in the U.S. as long as candidate can travel and will need to be located near a major airport in the U.S. Key for this role is data center mechanical experience selling engineered products and solutions with a rolodex of either colo or hyperscale contacts.  This opportunity is working directly with a successful global OEM looking to expand market share in the critical facilities industry. They help data centers reduce energy and operating costs by providing mechanical solutions that modernize their infrastructure. By ensuring high-reliability mission-critical facilities for some of the world’s largest hyperscale, colocation, and enterprise customers, this company offers a career-growth-minded opportunity with exciting projects, cutting-edge technology, competitive salaries, and benefits. Engineering Design Director – Data CenterDenver, CO / Dallas, TX / Remote This position is a remote position in the Midwest supporting HPC / AI colo design projects being built in TX. Previous A/E experience with data center hyperscale design a must! We are seeking an experienced director of data center engineering design who will lead the development and oversight of critical power and mechanical infrastructure design for new data center builds and existing data center facilities. The ideal candidate will bring strong technical acumen, leadership skills, and extensive experience in mission-critical environments. You will oversee all aspects of engineering design, guide project teams, interface with stakeholders, and ensure best practices are upheld in quality,

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »