Stay Ahead, Stay ONMINE

Writer releases Palmyra X5, delivers near GPT-4 performance at 75% lower cost

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Writer, the enterprise generative AI company valued at $1.9 billion, today released Palmyra X5, a new large language model (LLM) featuring an expansive 1-million-token context window that promises to accelerate the adoption of autonomous AI agents […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Writer, the enterprise generative AI company valued at $1.9 billion, today released Palmyra X5, a new large language model (LLM) featuring an expansive 1-million-token context window that promises to accelerate the adoption of autonomous AI agents in corporate environments.

The San Francisco-based company, which counts Accenture, Marriott, Uber, and Vanguard among its hundreds of enterprise customers, has positioned the model as a cost-efficient alternative to offerings from industry giants like OpenAI and Anthropic, with pricing set at $0.60 per million input tokens and $6 per million output tokens.

“This model really unlocks the agentic world,” said Matan-Paul Shetrit, Director of Product at Writer, in an interview with VentureBeat. “It’s faster and cheaper than equivalent large context window models out there like GPT-4.1, and when you combine it with the large context window and the model’s ability to do tool or function calling, it allows you to start really doing things like multi-step agentic flows.”

A comparison of AI model efficiency showing Writer’s Palmyra X5 achieving nearly 20% accuracy on OpenAI’s MRCR benchmark at approximately $0.60 per million tokens, positioning it favorably against more expensive models like GPT-4.1 and GPT-4o (right) that cost over $2.00 per million tokens. (Credit: Writer)

AI economics breakthrough: How Writer trained a powerhouse model for just $1 million

Unlike many competitors, Writer trained Palmyra X5 with synthetic data for approximately $1 million in GPU costs — a fraction of what other leading models require. This cost efficiency represents a significant departure from the prevailing industry approach of spending tens or hundreds of millions on model development.

“Our belief is that tokens in general are becoming cheaper and cheaper, and the compute is becoming cheaper and cheaper,” Shetrit explained. “We’re here to solve real problems, rather than nickel and diming our customers on the pricing.”

The company’s cost advantage stems from proprietary techniques developed over several years. In 2023, Writer published research on “becoming self-instruct,” which introduced early stopping criteria for minimal instruct tuning. According to Shetrit, this allows Writer to “cut costs significantly” during the training process.

“Unlike other foundational shops, our view is that we need to be effective. We need to be efficient here,” Shetrit said. “We need to provide the fastest, cheapest models to our customers, because ROI really matters in these cases.”

Million-token marvel: The technical architecture powering Palmyra X5’s speed and accuracy

Palmyra X5 can process a full million-token prompt in approximately 22 seconds and execute multi-turn function calls in around 300 milliseconds — performance metrics that Writer claims enable “agent behaviors that were previously cost- or time-prohibitive.”

The model’s architecture incorporates two key technical innovations: a hybrid attention mechanism and a mixture of experts approach. “The hybrid attention mechanism…introduces attention mechanism that inside the model allows it to focus on the relevant parts of the inputs when generating each output,” Shetrit said. This approach accelerates response generation while maintaining accuracy across the extensive context window.

Palmyra X5’s hybrid attention architecture processes massive inputs through specialized decoder blocks, enabling efficient handling of million-token contexts. (Credit: Writer)

On benchmark tests, Palmyra X5 achieved notable results relative to its cost. On OpenAI’s MRCR 8-needle test — which challenges models to find eight identical requests hidden in a massive conversation — Palmyra X5 scored 19.1%, compared to 20.25% for GPT-4.1 and 17.63% for GPT-4o. It also places eighth in coding on the BigCodeBench benchmark with a score of 48.7.

These benchmarks demonstrate that while Palmyra X5 may not lead every performance category, it delivers near-flagship capabilities at significantly lower costs — a trade-off that Writer believes will resonate with enterprise customers focused on ROI.

From chatbots to business automation: How AI agents are transforming enterprise workflows

The release of Palmyra X5 comes shortly after Writer unveiled AI HQ earlier this month — a centralized platform for enterprises to build, deploy, and supervise AI agents. This dual product strategy positions Writer to capitalize on growing enterprise demand for AI that can execute complex business processes autonomously.

“In the age of agents, models offering less than 1 million tokens of context will quickly become irrelevant for business-critical use cases,” said Writer CTO and co-founder Waseem AlShikh in a statement.

Shetrit elaborated on this point: “For a long time, there’s been a large gap between the promise of AI agents and what they could actually deliver. But at Writer, we’re now seeing real-world agent implementations with major enterprise customers. And when I say real customers, it’s not like a travel agent use case. I’m talking about Global 2000 companies, solving the gnarliest problems in their business.”

Early adopters are deploying Palmyra X5 for various enterprise workflows, including financial reporting, RFP responses, support documentation, and customer feedback analysis.

One particularly compelling use case involves multi-step agentic workflows, where an AI agent can flag outdated content, generate suggested revisions, share them for human approval, and automatically push approved updates to a content management system.

This shift from simple text generation to process automation represents a fundamental evolution in how enterprises deploy AI — moving from augmenting human work to automating entire business functions.

Writer’s Palmyra X5 offers an 8x increase in context window size over its predecessor, allowing it to process the equivalent of 1,500 pages at once. (Credit: Writer)

Cloud expansion strategy: AWS partnership brings Writer’s AI to millions of enterprise developers

Alongside the model release, Writer announced that both Palmyra X5 and its predecessor, Palmyra X4, are now available in Amazon Bedrock, Amazon Web Services’ fully managed service for accessing foundation models. AWS becomes the first cloud provider to deliver fully managed models from Writer, significantly expanding the company’s potential reach.

“Seamless access to Writer’s Palmyra X5 will enable developers and enterprises to build and scale AI agents and transform how they reason over vast amounts of enterprise data—leveraging the security, scalability, and performance of AWS,” said Atul Deo, Director of Amazon Bedrock at AWS, in the announcement.

The AWS integration addresses a critical barrier to enterprise AI adoption: the technical complexity of deploying and managing models at scale. By making Palmyra X5 available through Bedrock’s simplified API, Writer can potentially reach millions of developers who lack the specialized expertise to work with foundation models directly.

Self-learning AI: Writer’s vision for models that improve without human intervention

Writer has staked a bold claim regarding context windows, announcing that 1 million tokens will be the minimum size for all future models it releases. This commitment reflects the company’s view that large context is essential for enterprise-grade AI agents that interact with multiple systems and data sources.

Looking ahead, Shetrit identified self-evolving models as the next major advancement in enterprise AI. “The reality is today, agents do not perform at the level we want and need them to perform,” he said. “What I think is realistic is as users come to AI HQ, they start doing this process mapping…and then you layer on top of that, or within it, the self-evolving models that learn from how you do things in your company.”

These self-evolving capabilities would fundamentally change how AI systems improve over time. Rather than requiring periodic retraining or fine-tuning by AI specialists, the models would learn continuously from their interactions, gradually improving their performance for specific enterprise use cases.

“This idea that one agent can rule them all is not realistic,” Shetrit noted when discussing the varied needs of different business teams. “Even two different product teams, they have so many such different ways of doing work, the PMs themselves.”

Enterprise AI’s new math: How Writer’s $1.9B strategy challenges OpenAI and Anthropic

Writer’s approach contrasts sharply with that of OpenAI and Anthropic, which have raised billions in funding but focus more on general-purpose AI development. Writer has instead concentrated on building enterprise-specific models with cost profiles that enable widespread deployment.

This strategy has attracted significant investor interest, with the company raising $200 million in Series C funding last November at a $1.9 billion valuation. The round was co-led by Premji Invest, Radical Ventures, and ICONIQ Growth, with participation from strategic investors including Salesforce Ventures, Adobe Ventures, and IBM Ventures.

According to Forbes, Writer has a remarkable 160% net retention rate, indicating that customers typically expand their contracts by 60% after initial adoption. The company reportedly has over $50 million in signed contracts and projects this will double to $100 million this year.

For enterprises evaluating generative AI investments, Writer’s Palmyra X5 presents a compelling value proposition: powerful capabilities at a fraction of the cost of competing solutions. As the AI agent ecosystem matures, the company’s bet on cost-efficient, enterprise-focused models could position it advantageously against better-funded competitors that may not be as attuned to business ROI requirements.

“Our goal is to drive widespread agent adoption across our customer base as quickly as possible,” Shetrit emphasized. “The economics are straightforward—if we price our solution too high, enterprises will simply compare the cost of an AI agent versus a human worker and may not see sufficient value. To accelerate adoption, we need to deliver both superior speed and significantly lower costs. That’s the only way to achieve large-scale deployment of these agents within major enterprises.”

In an industry often captivated by technical capabilities and theoretical performance ceilings, Writer’s pragmatic focus on cost efficiency might ultimately prove more revolutionary than another decimal point of benchmark improvement. As enterprises grow increasingly sophisticated in measuring AI’s business impact, the question may shift from “How powerful is your model?” to “How affordable is your intelligence?” — and Writer is betting its future that economics, not just capabilities, will determine AI’s enterprise winners.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

IBM aims for autonomous security operations

“IBM’s proactive threat hunting augments traditional security solutions to uncover anomalous activity and IBM’s proactive threat hunters work with organizations to help identify their crown jewel assets and critical concerns. This input enables the threat hunting team to create fully tailored threat hunt reports and customized detections,” IDC stated. “AI/ML

Read More »

Cisco automates AI-driven security across enterprise networks

“The result is automated detection and response for the most common attacks,” Shipley wrote in a blog post about the new XDR capabilities. “Machine learning, machine reasoning, and LLMs combine to trigger multiple AI agents acting on different parts of the investigation lifecycle. Each investigation has a clear verdict. This

Read More »

ISACA Barcelona president warns of quantum illiteracy

Gallego says the challenge is not theoretical but practical, adding, “We are already seeing clear warning signs.” He warns of “the so-called ‘harvest now, decrypt later’ attacks, which consist of intercepting encrypted data today to decrypt it in the future with quantum technology. This is not science fiction, but a

Read More »

Crude Dips to Two-Week Low Amid Trade War Fears

Oil declined as traders parsed economic data from the US and awaited further developments regarding President Donald Trump’s trade war, including China’s efforts to support its tariff-hit economy. West Texas Intermediate fell by 1.5% to settle near $62 a barrel, the lowest close in nearly two weeks. US equities also weakened amid a selloff in big tech and after a report found that Texas manufacturing activity had reached the lowest levels since May 2020. Treasury Secretary Scott Bessent told CNBC that while the US government is in contact with China, the onus is on Beijing to begin de-escalating the trade war with the US. In China — the top crude importer — officials vowed to provide more support for exporters affected by Trump’s tariffs, while denying any trade talks with Washington. Authorities are confident they can reach an expansion target of about 5% in 2025, Zhao Chenxin, vice chairman of the National Development and Reform Commission, said in a briefing. US crude is headed for a monthly slump of more than 13%, the biggest since 2021, after touching a four-year low. Futures have been burdened by concerns that the US-led trade war will stifle economic activity and hurt energy demand. At the same time, the OPEC+ cartel has compounded bearish sentiment by ramping up idled production. The group will hold discussions on May 5 to weigh output plans for June. The market’s next cues could come from a raft of major economic data this week, including US growth and payrolls figures. Investors will also get a chance to hear views on the global crude market outlook this week as oil supermajors BP Plc, Shell Plc, Chevron Corp. and Exxon Mobil Corp. report earnings. “Near-term crude looks decent despite bearish sentiment” and aided by low onshore crude inventories, said Aldo Spanjer,

Read More »

United States Signs Agreement to Advance American Civil Nuclear Deal in Poland

WARSAW, POLAND—U.S. Secretary of Energy Chris Wright today joined the Westinghouse/Bechtel Consortium (WBC), Polskie Elektrownie Jądrowe (PEJ) and Polish Prime Minister Donald Tusk for the signing of the Engineering Development Agreement (EDA) for Poland’s first AP-1000 nuclear power plant. This agreement is a follow-on to an initial contract and is part of a larger nuclear energy security deal worth tens of billions of dollars to build large-scale civil nuclear reactors in Poland. This initiative originally started during the first Trump Administration and could ultimately accommodate a total of six AP-1000s across two sites.  “The strong partnership between our two countries is united by a long history of steadfast friendship and shared values, and I am proud to be here to mark this significant milestone for the United States and Poland,” said Energy Secretary Chris Wright. “This first of a kind intergovernmental agreement, which began during President Trump’s first term, will provide Poland enormous levels of energy security and create tens of thousands of Polish and American jobs. It will truly be a joint endeavor that will include not just the construction of a very large power plant that will power the Polish economy for decades to come, but also marks the start of a long-term nuclear cooperation between the United States and Poland that will result in building future reactors.”   The EDA is an important step for the project in Choczewo, Pomeranian Voivodeship and paves the way for the next steps of the project, which include design, sitework, regulatory and procurement activities. This Westinghouse AP-1000, already operating in the U.S., will be constructed by Bechtel as a key partner in the proposed U.S. delivery team. Project construction for the three-unit site in northeastern Poland will generate almost 40,000 well-paying U.S. manufacturing, engineering and related jobs that support the project. Actual construction is

Read More »

Energy Secretary Chris Wright Delivers Keynote Remarks at the Three Seas Business Forum in Warsaw, Poland

WARSAW, POLAND— U.S. Energy Secretary Chris Wright delivered keynote remarks today at the Inaugural Session of the Three Seas Business Forum.  Secretary Wright’s full remarks from the Three Seas Business Forum are below: It is a great honor to stand here before you all at the 2025 Three Seas. A truly visionary idea from 10 years ago to unite the proud Central European nations in building infrastructure and investment in pursuit of opportunity and prosperity. Eight years ago, President Trump addressed the Three Seas right here in Poland and I will quote his words: “We support your drive for greater prosperity and security. We applaud your initiative to expand infrastructure. And we welcome this historic opportunity to deepen our economic partnership with your region.”  I can’t top those words, but I can reiterate them today. The United States stands here in partnership with all of you. We seek to work with you all for much betterment via energy, economic and strategic cooperation.  President Trump’s agenda is simple: Prosperity at home and peace abroad. He was elected by the American people to bring back commonsense to Washington and focus on bettering the lives of our citizens and our allies. I am in a room full of allies. Thank you all for that.  I thank Poland for hosting this fabulous conference and for inviting me to attend. I thank Poland and its people for its steadfast alliance with the United States that began with our Revolutionary War and continues today, as evidenced by our growing cooperation in LNG and our large-scale partnership in nuclear energy that was highlighted earlier today with a signing ceremony and press conference.  This nuclear partnership is strategic and long-lasting. It will grow and scale as we jointly pursue expansions of nuclear deployment in Poland and other countries.

Read More »

Shropshire Council approves 10 MW solar farm

Shropshire Council has granted planning permission to a major 10 MW solar farm and battery project. G Power Solar is developing the 24.31 hectare project, located at the Beeches in Alkington near Whitchurch. The battery facility would have capacity to store up to 12.5 MWh of energy. According to a development management report, the project could generate enough energy to power up to 3,000 homes in Shropshire, or roughly 2.1% of the region’s houses. According to the project’s application, the proposed development could save 2,150 tonnes of CO2 per annum – equivalent to at least 86,000 tonnes over the project’s lifetime of around 40 years. While the grid connection location has yet to be confirmed, the application states that there is a potential point of connection located 1 km to the east. The application site lies within designated open countryside and is bounded by hedgerows and trees. The majority of the site is classed as Grade 3 Agricultural Land (ALC), with a small section of the north-eastern part classed as Grade 2. In addition, the land around the solar panels will be available for sheep grazing, with only around 20% of the land occupied by the proposed equipment. The solar panels would be divided between five different field parcels, while the battery energy storage system (BESS) would be located in the north-eastern corner of the site. The project’s substation will also be included onsite. G Power Solar’s Shropshire project is part of a major solar push by the Labour government. Soon after coming to power last year, UK energy secretary Ed Miliband vowed to triple the UK’s solar capacity by 2030. Recent research found that solar panel installations in England rose by 10% in the first six months since Labour took office. This accounted for nearly 75,000 solar panels installed

Read More »

National Grid, Con Edison, other utilities urge FERC to consider gas pipeline reliability measures

A group of electric and gas utilities on Friday urged the Federal Energy Regulatory Commission to launch an inquiry to consider options for improving gas pipeline reliability. Increased demand and extreme weather in the last decade have stressed the interstate gas transportation and storage system and reduced pipeline companies’ operational flexibility, according to a filing at FERC by four National Grid utilities — Boston Gas, Brooklyn Union Gas, KeySpan Gas East and Niagara Mohawk Power — along with Consolidated Edison Co. of New York, Old Dominion Electric Cooperative, Orange & Rockland Utilities and Washington Gas Light. “The challenges facing interstate pipelines, and other parts of the natural gas value chain, continue to intensify warranting greater regulatory involvement and standardization of practices,” the utilities said. Under its Natural Gas Act authority, FERC could consider the following, the utilities suggest: Requiring interstate pipelines to submit annual reports on reliability metrics. Requiring pipeline owners to consider whether facilities they plan to replace or retire may still offer reliability benefits to customers as redundant facilities. Revisiting its policy regarding pipeline force majeure and reservation charge crediting provisions to rebalance risk-sharing between pipelines and their customers and create incentives for pipeline reliability. Additional enhanced communication protocols for extreme weather events, greater standardization of pipeline scheduling and confirmation practices, and other mechanisms to make sure customers that pay for firm service are receiving full contractual value in exchange for their financial commitments. FERC and state utility regulators are set to meet on Wednesday to discuss gas-electric coordination and gas storage. Speakers at the Federal and State Current Issues Collaborative meeting include Jim Robb, North American Electric Reliability Corp president and CEO; Gordon van Welie, ISO New England president and CEO; and Paul Cicio, Industrial Energy Consumers of America president and CEO.

Read More »

Trump’s battery tariffs threaten utility-scale storage and US grid reliability

Brandon N. Owens is the founder of Clean Power Shift, an independent journal and research platform. President Trump’s April 2 “Liberation Day” tariffs slap a 64.5% composite duty on Chinese lithium-ion batteries today and hard-wire an 82% bite for 2026, a one-two punch that could upend the economics of every grid-scale storage project still penciled into developers’ spreadsheets. Solar inputs get singed too — Chinese polysilicon, wafers, and cells now carry a straight 60% levy — but the real story is batteries, where the new math pushes imported packs toward price points the U.S. market has not seen since before the storage boom. Utility-scale battery systems are no longer peripheral components of the power grid — they are essential infrastructure. These systems balance intermittent wind and solar generation, defer costly transmission upgrades, and provide flexible capacity during peak load events and extreme weather. In 2024, the U.S. added 8.7 GW of battery storage capacity — a 66% year-over-year increase, according to the U.S. Energy Information Administration. In 2025, developers are scheduled to install over 18 GW, with the majority of battery units still sourced internationally. As of last year, nearly 70% of imported grid battery systems came from China, according to BloombergNEF. The newly announced tariffs will make these imports significantly more expensive, placing immediate upward pressure on project costs and increasing the risk of deployment delays. Several utilities and independent power producers may need to revise procurement strategies, renegotiate project terms, or in some cases, reassess near-term feasibility. For utilities facing ambitious clean energy mandates, resource adequacy challenges and extreme weather risks, battery systems have become indispensable. They play a pivotal role in evening ramp support, frequency regulation, contingency reserves and load shifting. The new tariffs will not only affect procurement costs but could also force utilities to turn

Read More »

Deep Data Center: Neoclouds as the ‘Picks and Shovels’ of the AI Gold Rush

In 1849, the discovery of gold in California ignited a frenzy, drawing prospectors from around the world in pursuit of quick fortune. While few struck it rich digging and sifting dirt, a different class of entrepreneurs quietly prospered: those who supplied the miners with the tools of the trade. From picks and shovels to tents and provisions, these providers became indispensable to the gold rush, profiting handsomely regardless of who found gold. Today, a new gold rush is underway, in pursuit of artificial intelligence. And just like the days of yore, the real fortunes may lie not in the gold itself, but in the infrastructure and equipment that enable its extraction. This is where neocloud players and chipmakers are positioned, representing themselves as the fundamental enablers of the AI revolution. Neoclouds: The Essential Tools and Implements of AI Innovation The AI boom has sparked a frenzy of innovation, investment, and competition. From generative AI applications like ChatGPT to autonomous systems and personalized recommendations, AI is rapidly transforming industries. Yet, behind every groundbreaking AI model lies an unsung hero: the infrastructure powering it. Enter neocloud providers—the specialized cloud platforms delivering the GPU horsepower that fuels AI’s meteoric rise. Let’s examine how neoclouds represent the “picks and shovels” of the AI gold rush, used for extracting the essential backbone of AI innovation. Neoclouds are emerging as indispensable players in the AI ecosystem, offering tailored solutions for compute-intensive workloads such as training large language models (LLMs) and performing high-speed inference. Unlike traditional hyperscalers (e.g., AWS, Azure, Google Cloud), which cater to a broad range of use cases, neoclouds focus exclusively on optimizing infrastructure for AI and machine learning applications. This specialization allows them to deliver superior performance at a lower cost, making them the go-to choice for startups, enterprises, and research institutions alike.

Read More »

Soluna Computing: Innovating Renewable Computing for Sustainable Data Centers

Dorothy 1A & 1B (Texas): These twin 25 MW facilities are powered by wind and serve Bitcoin hosting and mining workloads. Together, they consumed over 112,000 MWh of curtailed energy in 2024, demonstrating the impact of Soluna’s model. Dorothy 2 (Texas): Currently under construction and scheduled for energization in Q4 2025, this 48 MW site will increase Soluna’s hosting and mining capacity by 64%. Sophie (Kentucky): A 25 MW grid- and hydro-powered hosting center with a strong cost profile and consistent output. Project Grace (Texas): A 2 MW AI pilot project in development, part of Soluna’s transition into HPC and machine learning. Project Kati (Texas): With 166 MW split between Bitcoin and AI hosting, this project recently exited the Electric Reliability Council of Texas, Inc. planning phase and is expected to energize between 2025 and 2027. Project Rosa (Texas): A 187 MW flagship project co-located with wind assets, aimed at both Bitcoin and AI workloads. Land and power agreements were secured by the company in early 2025. These developments are part of the company’s broader effort to tackle both energy waste and infrastructure bottlenecks. Soluna’s behind-the-meter design enables flexibility to draw from the grid or directly from renewable sources, maximizing energy value while minimizing emissions. Competition is Fierce and a Narrower Focus Better Serves the Business In 2024, Soluna tested the waters of providing AI services via a  GPU-as-a-Service through a partnership with HPE, branded as Project Ada. The pilot aimed to rent out cloud GPUs for AI developers and LLM training. However, due to oversupply in the GPU market, delayed product rollouts (like NVIDIA’s H200), and poor demand economics, Soluna terminated the contract in March 2025. The cancellation of the contract with HPE frees up resources for Soluna to focus on what it believes the company does best: designing

Read More »

Quiet Genius at the Neutral Line: How Onics Filters Are Reshaping the Future of Data Center Power Efficiency

Why Harmonics Matter In a typical data center, nonlinear loads—like servers, UPS systems, and switch-mode power supplies—introduce harmonic distortion into the electrical system. These harmonics travel along the neutral and ground conductors, where they can increase current flow, cause overheating in transformers, and shorten the lifespan of critical power infrastructure. More subtly, they waste power through reactive losses that don’t show up on a basic utility bill, but do show up in heat, inefficiency, and increased infrastructure stress. Traditional mitigation approaches—like active harmonic filters or isolation transformers—are complex, expensive, and often require custom integration and ongoing maintenance. That’s where Onics’ solution stands out. It’s engineered as a shunt-style, low-pass filter: a passive device that sits in parallel with the circuit, quietly siphoning off problematic harmonics without interrupting operations.  The result? Lower apparent power demand, reduced electrical losses, and a quieter, more stable current environment—especially on the neutral line, where cumulative harmonic effects often peak. Behind the Numbers: Real-World Impact While the Onics filters offer a passive complement to traditional mitigation strategies, they aren’t intended to replace active harmonic filters or isolation transformers in systems that require them—they work best as a low-complexity enhancement to existing power quality designs. LoPilato says Onics has deployed its filters in mission-critical environments ranging from enterprise edge to large colos, and the data is consistent. In one example, a 6 MW data center saw a verified 9.2% reduction in energy consumption after deploying Onics filters at key electrical junctures. Another facility clocked in at 17.8% savings across its lighting and support loads, thanks in part to improved power factor and reduced transformer strain. The filters work by targeting high-frequency distortion—typically above the 3rd harmonic and up through the 35th. By passively attenuating this range, the system reduces reactive current on the neutral and helps stabilize

Read More »

New IEA Report Contrasts Energy Bottlenecks with Opportunities for AI and Data Center Growth

Artificial intelligence has, without question, crossed the threshold—from a speculative academic pursuit into the defining infrastructure of 21st-century commerce, governance, and innovation. What began in the realm of research labs and open-source models is now embedded in the capital stack of every major hyperscaler, semiconductor roadmap, and national industrial strategy. But as AI scales, so does its energy footprint. From Nvidia-powered GPU clusters to exascale training farms, the conversation across boardrooms and site selection teams has fundamentally shifted. It’s no longer just about compute density, thermal loads, or software frameworks. It’s about power—how to find it, finance it, future-proof it, and increasingly, how to generate it onsite. That refrain—“It’s all about power now”—has moved from a whisper to a full-throated consensus across the data center industry. The latest report from the International Energy Agency (IEA) gives this refrain global context and hard numbers, affirming what developers, utilities, and infrastructure operators have already sensed on the ground: the AI revolution will be throttled or propelled by the availability of scalable, sustainable, and dispatchable electricity. Why Energy Is the Real Bottleneck to Intelligence at Scale The major new IEA report puts it plainly: The transformative promise of AI will be throttled—or unleashed—by the world’s ability to deliver scalable, reliable, and sustainable electricity. The stakes are enormous. Countries that can supply the power AI craves will shape the future. Those that can’t may find themselves sidelined. Importantly, while AI poses clear challenges, the report emphasizes how it also offers solutions: from optimizing energy grids and reducing emissions in industrial sectors to enhancing energy security by supporting infrastructure defenses against cyberattacks. The report calls for immediate investments in both energy generation and grid capabilities, as well as stronger collaboration between the tech and energy sectors to avoid critical bottlenecks. The IEA advises that, for countries

Read More »

Colorado Eyes the AI Data Center Boom with Bold Incentive Push

Even as states work on legislation to limit data center development, it is clear that some locations are looking to get a bigger piece of the huge data center spending that the AI wave has created. It appears that politicians in Colorado took a look around and thought to themselves “Why is all that data center building going to Texas and Arizona? What’s wrong with the Rocky Mountain State?” Taking a page from the proven playbook that has gotten data centers built all over the country, Colorado is trying to jump on the financial incentives for data center development bandwagon. SB 24-085: A Statewide Strategy to Attract Data Center Investment Looking to significantly boost its appeal as a data center hub, Colorado is now considering Senate Bill 24-085, currently making its way through the state legislature. Sponsored by Senators Priola and Buckner and Representatives Parenti and Weinberg, this legislation promises substantial economic incentives in the form of state sales and use tax rebates for new data centers established within the state from fiscal year 2026 through 2033. Colorado hopes to position itself strategically to compete with neighboring states in attracting lucrative tech investments and high-skilled jobs. According to DataCenterMap.com, there are currently 53 data centers in the state, almost all located in the Denver area, but they are predominantly smaller facilities. In today’s era of massive AI-driven hyperscale expansion, Colorado is rarely mentioned in the same breath as major AI data center markets.  Some local communities have passed their own incentive packages, but SB 24-085 aims to offer a unified, statewide framework that can also help mitigate growing NIMBY (Not In My Backyard) sentiment around new developments. The Details: How SB 24-085 Works The bill, titled “Concerning a rebate of the state sales and use tax paid on new digital infrastructure

Read More »

Wonder Valley and the Great AI Pivot: Kevin O’Leary’s Bold Data Center Play

Data Center World 2025 drew record-breaking attendance, underscoring the AI-fueled urgency transforming infrastructure investment. But no session captivated the crowd quite like Kevin O’Leary’s electrifying keynote on Wonder Valley—his audacious plan to build the world’s largest AI compute data center campus. In a sweeping narrative that ranged from pandemic pivots to stranded gas and Branson-brand inspiration, O’Leary laid out a real estate and infrastructure strategy built for the AI era. A Pandemic-Era Pivot Becomes a Case Study in Digital Resilience O’Leary opened with a Shark Tank success story that doubled as a business parable. In 2019, a woman-led startup called Blueland raised $50 million to eliminate plastic cleaning bottles by shipping concentrated cleaning tablets in reusable kits. When COVID-19 shut down retail in 2020, her inventory was stuck in limbo—until she made an urgent call to O’Leary. What followed was a high-stakes, last-minute pivot: a union-approved commercial shoot in Brooklyn the night SAG-AFTRA shut down television production. The direct response ad campaign that resulted would not only liquidate the stranded inventory at full margin, but deliver something more valuable—data. By targeting locked-down consumers through local remnant TV ad slots and optimizing by conversion, Blueland saw unheard-of response rates as high as 17%. The campaign turned into a data goldmine: buyer locations, tablet usage patterns, household sizes, and contact details. Follow-up SMS campaigns would drive 30% reorders. “It built such a franchise in those 36 months,” O’Leary said, “with no retail. Now every retailer wants in.” The lesson? Build your infrastructure to control your data, and you build a business that scales even in chaos. This anecdote set the tone for the keynote: in a volatile world, infrastructure resilience and data control are the new core competencies. The Data Center Power Crisis: “There Is Not a Gig on the Grid” O’Leary

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »