Stay Ahead, Stay ONMINE

DeepSeek’s success shows why motivation is key to AI innovation

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More January 2025 shook the AI landscape. The seemingly unstoppable OpenAI and the powerful American tech giants were shocked by what we can certainly call an underdog in the area of large language models (LLMs). DeepSeek, a […]

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


January 2025 shook the AI landscape. The seemingly unstoppable OpenAI and the powerful American tech giants were shocked by what we can certainly call an underdog in the area of large language models (LLMs). DeepSeek, a Chinese firm not on anyone’s radar, suddenly challenged OpenAI. It is not that DeepSeek-R1 was better than the top models from American giants; it was slightly behind in terms of the benchmarks, but it suddenly made everyone think about the efficiency in terms of hardware and energy usage.

Given the unavailability of the best high-end hardware, it seems that DeepSeek was motivated to innovate in the area of efficiency, which was a lesser concern for larger players. OpenAI has claimed they have evidence suggesting DeepSeek may have used their model for training, but we have no concrete proof to support this. So, whether it is true or it’s OpenAI simply trying to appease their investors is a topic of debate. However, DeepSeek has published their work, and people have verified that the results are reproducible at least on a much smaller scale.

But how could DeepSeek attain such cost-savings while American companies could not? The short answer is simple: They had more motivation. The long answer requires a little bit more of a technical explanation.

DeepSeek used KV-cache optimization

One important cost-saving for GPU memory was optimization of the Key-Value cache used in every attention layer in an LLM.

LLMs are made up of transformer blocks, each of which comprises an attention layer followed by a regular vanilla feed-forward network. The feed-forward network conceptually models arbitrary relationships, but in practice, it is difficult for it to always determine patterns in the data. The attention layer solves this problem for language modeling.

The model processes texts using tokens, but for simplicity, we will refer to them as words. In an LLM, each word gets assigned a vector in a high dimension (say, a thousand dimensions). Conceptually, each dimension represents a concept, like being hot or cold, being green, being soft, being a noun. A word’s vector representation is its meaning and values according to each dimension.

However, our language allows other words to modify the meaning of each word. For example, an apple has a meaning. But we can have a green apple as a modified version. A more extreme example of modification would be that an apple in an iPhone context differs from an apple in a meadow context. How do we let our system modify the vector meaning of a word based on another word? This is where attention comes in.

The attention model assigns two other vectors to each word: a key and a query. The query represents the qualities of a word’s meaning that can be modified, and the key represents the type of modifications it can provide to other words. For example, the word ‘green’ can provide information about color and green-ness. So, the key of the word ‘green’ will have a high value on the ‘green-ness’ dimension. On the other hand, the word ‘apple’ can be green or not, so the query vector of ‘apple’ would also have a high value for the green-ness dimension. If we take the dot product of the key of ‘green’ with the query of ‘apple,’ the product should be relatively large compared to the product of the key of ‘table’ and the query of ‘apple.’ The attention layer then adds a small fraction of the value of the word ‘green’ to the value of the word ‘apple’. This way, the value of the word ‘apple’ is modified to be a little greener.

When the LLM generates text, it does so one word after another. When it generates a word, all the previously generated words become part of its context. However, the keys and values of those words are already computed. When another word is added to the context, its value needs to be updated based on its query and the keys and values of all the previous words. That’s why all those values are stored in the GPU memory. This is the KV cache.

DeepSeek determined that the key and the value of a word are related. So, the meaning of the word green and its ability to affect greenness are obviously very closely related. So, it is possible to compress both as a single (and maybe smaller) vector and decompress while processing very easily. DeepSeek has found that it does affect their performance on benchmarks, but it saves a lot of GPU memory.

DeepSeek applied MoE

The nature of a neural network is that the entire network needs to be evaluated (or computed) for every query. However, not all of this is useful computation. Knowledge of the world sits in the weights or parameters of a network. Knowledge about the Eiffel Tower is not used to answer questions about the history of South American tribes. Knowing that an apple is a fruit is not useful while answering questions about the general theory of relativity. However, when the network is computed, all parts of the network are processed regardless. This incurs huge computation costs during text generation that should ideally be avoided. This is where the idea of the mixture-of-experts (MoE) comes in.

In an MoE model, the neural network is divided into multiple smaller networks called experts. Note that the ‘expert’ in the subject matter is not explicitly defined; the network figures it out during training. However, the networks assign some relevance score to each query and only activate the parts with higher matching scores. This provides huge cost savings in computation. Note that some questions need expertise in multiple areas to be answered properly, and the performance of such queries will be degraded. However, because the areas are figured out from the data, the number of such questions is minimised.

The importance of reinforcement learning

An LLM is taught to think through a chain-of-thought model, with the model fine-tuned to imitate thinking before delivering the answer. The model is asked to verbalize its thought (generate the thought before generating the answer). The model is then evaluated both on the thought and the answer, and trained with reinforcement learning (rewarded for a correct match and penalized for an incorrect match with the training data).

This requires expensive training data with the thought token. DeepSeek only asked the system to generate the thoughts between the tags and and to generate the answers between the tags and . The model is rewarded or penalized purely based on the form (the use of the tags) and the match of the answers. This required much less expensive training data. During the early phase of RL, the model tried generated very little thought, which resulted in incorrect answers. Eventually, the model learned to generate both long and coherent thoughts, which is what DeepSeek calls the ‘a-ha’ moment. After this point, the quality of the answers improved quite a lot.

DeepSeek employs several additional optimization tricks. However, they are highly technical, so I will not delve into them here.

Final thoughts about DeepSeek and the larger market

In any technology research, we first need to see what is possible before improving efficiency. This is a natural progression. DeepSeek’s contribution to the LLM landscape is phenomenal. The academic contribution cannot be ignored, whether or not they are trained using OpenAI output. It can also transform the way startups operate. But there is no reason for OpenAI or the other American giants to despair. This is how research works — one group benefits from the research of the other groups. DeepSeek certainly benefited from the earlier research performed by Google, OpenAI and numerous other researchers.

However, the idea that OpenAI will dominate the LLM world indefinitely is now very unlikely. No amount of regulatory lobbying or finger-pointing will preserve their monopoly. The technology is already in the hands of many and out in the open, making its progress unstoppable. Although this may be a little bit of a headache for the investors of OpenAI, it’s ultimately a win for the rest of us. While the future belongs to many, we will always be thankful to early contributors like Google and OpenAI.

Debasish Ray Chawdhuri is senior principal engineer at Talentica Software.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Lenovo targets AI workloads with massive storage update

These systems are aimed at enterprises that want to use both AI and virtualized systems, since AI hardware is not virtualized but bare-metal. For example, the all-new Lenovo ThinkAgile Converged Solution for VMware, which combines the enterprise-class features of the ThinkAgile VX Series with the data-management capabilities of the ThinkSystem

Read More »

AI plus digital twins could be the pairing enterprises need

3. Link AI with transactional workflows The third point of cooperation is the linkage of AI with transactional workflows. Companies already have applications that take orders, ship goods, move component parts around, and so forth. It’s these applications that currently drive the commercial side of a business, but taking an

Read More »

GeoPark Exceeds Q1 Production Guidance

GeoPark Limited, a Latin American energy company, has said its pro-forma average daily production in the first quarter reached 36,279 barrels of oil equivalent (boe), surpassing its initial guidance for the year of 35,000 boe. The company said production from organic operations was 8 percent lower than the fourth quarter of 2024 primarily due to an operational suspension on January 8, 2025, of the Platanillo Block and the divestiture of Llanos 32. Production in Vaca Muerta reached a record of 17,358 barrels of oil equivalent per day (boepd) gross during February 2025, the company said, adding that it drilled nine wells during the quarter under review, four of which are in Vaca Muerta. In the quarter, Vaca Muerta averaged 15,533 boepd. Key developments during the quarter included bringing the PAD-9 area online, which consists of three wells each with a lateral length of 3,200 meters (10,500 feet), and completing the drilling of the PAD-12 area, featuring four wells each extending 3,000 meters (9,842 feet) laterally, both within the Mata Mora Norte Block. GeoPark said it initiated the construction of the Central Processing Facility (CPF-2), aiming for a total treatment capacity of 40,000 bopd by May 2026. Additionally, the Duplicar Plus Pipeline commenced operations, securing a substantial transportation capacity of 19,000 gross bopd. GeoPark also reported record quarterly production in Llanos, achieving an average of 1,894 boepd (3,787 boepd gross) in the first quarter of 2025, fueled by strong performance from recently connected wells in the Llanos 123 Block. The Currucutu-1 exploration well, completed in April 2025, encountered 79 feet of net pay and is producing 1,360 bopd gross with minimal water. Additionally, the Bisbita Oeste-1 well is contributing 585 bopd gross, and a January workover on the Zorzal Este-2 appraisal well in the Llanos 87 Block is now averaging 285-300

Read More »

DV Group Embarks on Big Push Into Commodities Beyond Energy

DV Group is expanding its nascent commodities-trading business, with plans to hire as many as 15 traders by the end of the year, particularly in the increasingly volatile softs business.  The Chicago-based proprietary trading firm — one of the biggest market makers across the oil industry — will add staff for base and precious metals as well as agriculture and soft commodities. The firm previously specialized in crude, refined products, natural gas and related markets across the US, Europe and Asia. To emphasize the broader focus, the trading operation will change its name from DV Energy to DV Commodities. The new businesses will contribute half of the firm’s revenue and account for two-thirds of its risk exposure, said Sean Lambert, partner at DV Trading and global head of DV Commodities.  “The rebrand is meant to reflect the markets that we’re heavily involved in,” Lambert, who founded DV Energy in 2013, said in an interview. “It’s not just energy anymore.” Commodities beyond energy have been a focus for hedge funds and merchant traders in recent years as wider price swings create opportunities for bumper profits. At the same time, haven assets such as precious metals have become a popular bet as US economic dominance comes into question. Base metals — used for electric-vehicle batteries, power cables and renewable-energy facilities — are also expected to be a high-growth area in the years ahead.  Bloomberg reported earlier that Vitol Group and Gunvor Group both took significant long positions in LME aluminum contracts, while Mercuria Energy Group Ltd. snapped up copper to leverage a widening price differential on the heels of US tariffs. DV plans to expand into physical base metals, but that’s not an “immediate priority,” Lambert said.  Soft commodities are also having a moment. Last year, Andurand Capital Management raised eyebrows after exiting all positions in oil futures in

Read More »

FirstEnergy faces minimal direct tariff impact, but ‘uncertainty’ weighs on customers: CEO

FirstEnergy faces little direct exposure to the Trump administration’s tariffs on foreign imports, said Brian Tierney, FirstEnergy chair, president and CEO, during a Thursday earnings call. The utility company’s tariff exposure represents less than 0.2% on its $28 billion, five-year capital investment program, according to Tierney. “Proactive management of our supply chain since COVID has resulted in a diversified supplier base with little exposure to single-source suppliers,” he said, noting that most of the company’s operations and maintenance expense is labor. However, Tierney said he agreed with comments made by Beth Hammack, president and CEO of the Federal Reserve Bank of Cleveland, that tariffs add economic uncertainty — making it difficult for people to make investment decisions. “We don’t anticipate a significant impact from an income standpoint in the near-term associated with the near-term uncertainty that we’re dealing with,” Tierney said in response to an analyst question about a possible economic slowdown. “But the quicker there’s certainty from an investment cycle standpoint, I think the quicker we’ll be able to see people be able to make those investment decisions and get on with investing in their business, whatever the answer is from a tariff standpoint.” A report released Tuesday by the Cleveland Fed found that 64% of respondents to a February survey expected that their business would be affected by import tariffs. “Sizable majorities of respondents expected tariffs to increase both input costs and selling prices while decreasing demand for their products and services,” Cleveland Fed analysts said in the report. FirstEnergy’s industrial sales fell about 6% to 12.8 million MWh in the first quarter, down from 13.6 million MWh in the third quarter last year, according to the company’s earnings presentation. The dip is mainly from steel manufacturers slowing production related to automotive demand, said Jon Taylor, senior vice

Read More »

PG&E foresees ‘bright future’ with lower prices, higher demand

Dive Brief: Pacific Gas & Electric expects to file a rate case with its lowest requested rate increase in a decade as the company continues to make strides toward cutting costs and improving its credit rating, CEO Patti Poppe told investors during a Thursday earnings call. Increased demand from prospective data center customers — which grew sharply — should help the company reduce residential rates, and PG&E does not expect to be heavily impacted by tariffs, according to Poppe. Company leaders held up a recent rating upgrade by Moody’s as proof of their progress, but cautioned that further improvements in the company’s credit score are unlikely pending legislation to reform California’s overwhelmed wildfire insurance fund. Dive Insight: With growing demand and possible legislative reforms in the works, Poppe expressed optimism about PG&E’s future even though the San Francisco-based company saw its first-quarter earnings decline slightly. Customers may remain skeptical of the company’s ability to deliver, but PG&E is on a path toward lower costs and more modest electric bills, she said. “We’re interrupting a pattern here in California for affordability for customers,” she said. “We know our customers don’t feel that yet, so there is a doubt that we can deliver on this but we are able to deliver, and we’ve delivered this year.” The company’s upcoming general rate case, which PG&E plans to initiate next month, will request the lowest increase in electric rates in a decade, Poppe said — and it doesn’t include the effects of other potential cost-saving developments, including the improved credit rating by Moody’s and growing electrical demand. PG&E’s regulatory framework, Poppe said, should allow the company to cut overall customer bills by 1% to 2% for every gigawatt of new demand from data centers. PG&E’s data center pipeline grew from 5.5 GW at the beginning

Read More »

WTI Ends Higher But Logs Weekly Loss

Oil edged higher in a day of listless trading as investors parsed conflicting messaging on the progress of trade talks between the US and China. West Texas Intermediate futures rose to settle near $63 a barrel, but still notched their third weekly loss in the past four. Chinese authorities are weighing removing additional levies on a number of products including ethane, according to people familiar with the matter, as economic costs mount for certain industries. Shares in China’s top buyers of the fuel from the US jumped. Still, an agreement on trade between the US and China appears far off. President Donald Trump said Thursday that his administration was talking with China about trade, despite Beijing earlier denying the existence of negotiations and demanding that unilateral tariffs be revoked. The president later said that he wouldn’t drop tariffs on China unless “something substantial” is offered in return. Oil has dropped sharply this month on concerns that Trump’s sweeping tariffs and retaliatory measures from trading partners including China will cripple economic activity and throttle energy demand. In an effort to reassure US oil firms, Energy Secretary Chris Wright said that the trade turmoil will be fleeting and that the administration fully supports more crude output. “Our president is very clear and he wants lower energy prices,” Wright said during an interview with Bloomberg Television at an energy conference in Oklahoma City. Oil prices at $50 per barrel “in today’s world probably is not sustainable for our producers in this country.” The president “wants American industry and American consumers to thrive,” he added. The OPEC+ alliance has added to bearish headwinds by ramping up idled oil production, stoking fears of an oversupply. The group will meet on May 5 to discuss its output plans for June. Still, some metrics are pointing to

Read More »

Trade War Is Diverting USA Petroleum Gas Cargoes Away From China

Multiple carriers of petroleum-based gases traveling from the US to China have begun diverting to other countries due to the intensifying trade war between the world’s two largest economies.  Four cargoes of propane have shifted their routes from China to alternate destinations over the past week, bound for countries including Japan and South Korea, according to a report from analytics firm Vortexa. At least one cargo of ethane — which is used in plastics production — has been scrapped entirely, according to a person familiar with the matter. The diversions show the disruption to supply chains caused by the trade fight between the US and China, historically a major buyer of US ethane and petroleum gases. President Donald Trump has levied 145% tariffs on most US imports from China, and the US Trade Representative more recently imposed steep fees on Chinese-linked vessels seeking to access American ports. Eight Very Large Gas Carriers carrying US LPG were still on course for China as of this week, while the four diversions have all been recorded since April 17, according to Vortexa. Diverted vessels include the Zakher, Maple Gas, BW Gemini and Eiger Explorer, all departing from the US Gulf Coast.  The G. Arete, a propane carrier, diverted to South Korea from China, while a chemical tanker named STI Notting Hill is also rerouting to South Korea, Vortexa said. The US exported about 310,000 barrels of propane to China per day in 2024, double the volume from a year earlier, according to East Daley Analytics. Spot ethane shipments may continue to be affected by the trade war, while committed cargoes are harder to unwind, the person said. Asia-bound flows of ethylene — used in plastics and industrial solvents — have already slowed because of seasonal factors but may be further reduced by the tariffs, the

Read More »

Deep Data Center: Neoclouds as the ‘Picks and Shovels’ of the AI Gold Rush

In 1849, the discovery of gold in California ignited a frenzy, drawing prospectors from around the world in pursuit of quick fortune. While few struck it rich digging and sifting dirt, a different class of entrepreneurs quietly prospered: those who supplied the miners with the tools of the trade. From picks and shovels to tents and provisions, these providers became indispensable to the gold rush, profiting handsomely regardless of who found gold. Today, a new gold rush is underway, in pursuit of artificial intelligence. And just like the days of yore, the real fortunes may lie not in the gold itself, but in the infrastructure and equipment that enable its extraction. This is where neocloud players and chipmakers are positioned, representing themselves as the fundamental enablers of the AI revolution. Neoclouds: The Essential Tools and Implements of AI Innovation The AI boom has sparked a frenzy of innovation, investment, and competition. From generative AI applications like ChatGPT to autonomous systems and personalized recommendations, AI is rapidly transforming industries. Yet, behind every groundbreaking AI model lies an unsung hero: the infrastructure powering it. Enter neocloud providers—the specialized cloud platforms delivering the GPU horsepower that fuels AI’s meteoric rise. Let’s examine how neoclouds represent the “picks and shovels” of the AI gold rush, used for extracting the essential backbone of AI innovation. Neoclouds are emerging as indispensable players in the AI ecosystem, offering tailored solutions for compute-intensive workloads such as training large language models (LLMs) and performing high-speed inference. Unlike traditional hyperscalers (e.g., AWS, Azure, Google Cloud), which cater to a broad range of use cases, neoclouds focus exclusively on optimizing infrastructure for AI and machine learning applications. This specialization allows them to deliver superior performance at a lower cost, making them the go-to choice for startups, enterprises, and research institutions alike.

Read More »

Soluna Computing: Innovating Renewable Computing for Sustainable Data Centers

Dorothy 1A & 1B (Texas): These twin 25 MW facilities are powered by wind and serve Bitcoin hosting and mining workloads. Together, they consumed over 112,000 MWh of curtailed energy in 2024, demonstrating the impact of Soluna’s model. Dorothy 2 (Texas): Currently under construction and scheduled for energization in Q4 2025, this 48 MW site will increase Soluna’s hosting and mining capacity by 64%. Sophie (Kentucky): A 25 MW grid- and hydro-powered hosting center with a strong cost profile and consistent output. Project Grace (Texas): A 2 MW AI pilot project in development, part of Soluna’s transition into HPC and machine learning. Project Kati (Texas): With 166 MW split between Bitcoin and AI hosting, this project recently exited the Electric Reliability Council of Texas, Inc. planning phase and is expected to energize between 2025 and 2027. Project Rosa (Texas): A 187 MW flagship project co-located with wind assets, aimed at both Bitcoin and AI workloads. Land and power agreements were secured by the company in early 2025. These developments are part of the company’s broader effort to tackle both energy waste and infrastructure bottlenecks. Soluna’s behind-the-meter design enables flexibility to draw from the grid or directly from renewable sources, maximizing energy value while minimizing emissions. Competition is Fierce and a Narrower Focus Better Serves the Business In 2024, Soluna tested the waters of providing AI services via a  GPU-as-a-Service through a partnership with HPE, branded as Project Ada. The pilot aimed to rent out cloud GPUs for AI developers and LLM training. However, due to oversupply in the GPU market, delayed product rollouts (like NVIDIA’s H200), and poor demand economics, Soluna terminated the contract in March 2025. The cancellation of the contract with HPE frees up resources for Soluna to focus on what it believes the company does best: designing

Read More »

Quiet Genius at the Neutral Line: How Onics Filters Are Reshaping the Future of Data Center Power Efficiency

Why Harmonics Matter In a typical data center, nonlinear loads—like servers, UPS systems, and switch-mode power supplies—introduce harmonic distortion into the electrical system. These harmonics travel along the neutral and ground conductors, where they can increase current flow, cause overheating in transformers, and shorten the lifespan of critical power infrastructure. More subtly, they waste power through reactive losses that don’t show up on a basic utility bill, but do show up in heat, inefficiency, and increased infrastructure stress. Traditional mitigation approaches—like active harmonic filters or isolation transformers—are complex, expensive, and often require custom integration and ongoing maintenance. That’s where Onics’ solution stands out. It’s engineered as a shunt-style, low-pass filter: a passive device that sits in parallel with the circuit, quietly siphoning off problematic harmonics without interrupting operations.  The result? Lower apparent power demand, reduced electrical losses, and a quieter, more stable current environment—especially on the neutral line, where cumulative harmonic effects often peak. Behind the Numbers: Real-World Impact While the Onics filters offer a passive complement to traditional mitigation strategies, they aren’t intended to replace active harmonic filters or isolation transformers in systems that require them—they work best as a low-complexity enhancement to existing power quality designs. LoPilato says Onics has deployed its filters in mission-critical environments ranging from enterprise edge to large colos, and the data is consistent. In one example, a 6 MW data center saw a verified 9.2% reduction in energy consumption after deploying Onics filters at key electrical junctures. Another facility clocked in at 17.8% savings across its lighting and support loads, thanks in part to improved power factor and reduced transformer strain. The filters work by targeting high-frequency distortion—typically above the 3rd harmonic and up through the 35th. By passively attenuating this range, the system reduces reactive current on the neutral and helps stabilize

Read More »

New IEA Report Contrasts Energy Bottlenecks with Opportunities for AI and Data Center Growth

Artificial intelligence has, without question, crossed the threshold—from a speculative academic pursuit into the defining infrastructure of 21st-century commerce, governance, and innovation. What began in the realm of research labs and open-source models is now embedded in the capital stack of every major hyperscaler, semiconductor roadmap, and national industrial strategy. But as AI scales, so does its energy footprint. From Nvidia-powered GPU clusters to exascale training farms, the conversation across boardrooms and site selection teams has fundamentally shifted. It’s no longer just about compute density, thermal loads, or software frameworks. It’s about power—how to find it, finance it, future-proof it, and increasingly, how to generate it onsite. That refrain—“It’s all about power now”—has moved from a whisper to a full-throated consensus across the data center industry. The latest report from the International Energy Agency (IEA) gives this refrain global context and hard numbers, affirming what developers, utilities, and infrastructure operators have already sensed on the ground: the AI revolution will be throttled or propelled by the availability of scalable, sustainable, and dispatchable electricity. Why Energy Is the Real Bottleneck to Intelligence at Scale The major new IEA report puts it plainly: The transformative promise of AI will be throttled—or unleashed—by the world’s ability to deliver scalable, reliable, and sustainable electricity. The stakes are enormous. Countries that can supply the power AI craves will shape the future. Those that can’t may find themselves sidelined. Importantly, while AI poses clear challenges, the report emphasizes how it also offers solutions: from optimizing energy grids and reducing emissions in industrial sectors to enhancing energy security by supporting infrastructure defenses against cyberattacks. The report calls for immediate investments in both energy generation and grid capabilities, as well as stronger collaboration between the tech and energy sectors to avoid critical bottlenecks. The IEA advises that, for countries

Read More »

Colorado Eyes the AI Data Center Boom with Bold Incentive Push

Even as states work on legislation to limit data center development, it is clear that some locations are looking to get a bigger piece of the huge data center spending that the AI wave has created. It appears that politicians in Colorado took a look around and thought to themselves “Why is all that data center building going to Texas and Arizona? What’s wrong with the Rocky Mountain State?” Taking a page from the proven playbook that has gotten data centers built all over the country, Colorado is trying to jump on the financial incentives for data center development bandwagon. SB 24-085: A Statewide Strategy to Attract Data Center Investment Looking to significantly boost its appeal as a data center hub, Colorado is now considering Senate Bill 24-085, currently making its way through the state legislature. Sponsored by Senators Priola and Buckner and Representatives Parenti and Weinberg, this legislation promises substantial economic incentives in the form of state sales and use tax rebates for new data centers established within the state from fiscal year 2026 through 2033. Colorado hopes to position itself strategically to compete with neighboring states in attracting lucrative tech investments and high-skilled jobs. According to DataCenterMap.com, there are currently 53 data centers in the state, almost all located in the Denver area, but they are predominantly smaller facilities. In today’s era of massive AI-driven hyperscale expansion, Colorado is rarely mentioned in the same breath as major AI data center markets.  Some local communities have passed their own incentive packages, but SB 24-085 aims to offer a unified, statewide framework that can also help mitigate growing NIMBY (Not In My Backyard) sentiment around new developments. The Details: How SB 24-085 Works The bill, titled “Concerning a rebate of the state sales and use tax paid on new digital infrastructure

Read More »

Wonder Valley and the Great AI Pivot: Kevin O’Leary’s Bold Data Center Play

Data Center World 2025 drew record-breaking attendance, underscoring the AI-fueled urgency transforming infrastructure investment. But no session captivated the crowd quite like Kevin O’Leary’s electrifying keynote on Wonder Valley—his audacious plan to build the world’s largest AI compute data center campus. In a sweeping narrative that ranged from pandemic pivots to stranded gas and Branson-brand inspiration, O’Leary laid out a real estate and infrastructure strategy built for the AI era. A Pandemic-Era Pivot Becomes a Case Study in Digital Resilience O’Leary opened with a Shark Tank success story that doubled as a business parable. In 2019, a woman-led startup called Blueland raised $50 million to eliminate plastic cleaning bottles by shipping concentrated cleaning tablets in reusable kits. When COVID-19 shut down retail in 2020, her inventory was stuck in limbo—until she made an urgent call to O’Leary. What followed was a high-stakes, last-minute pivot: a union-approved commercial shoot in Brooklyn the night SAG-AFTRA shut down television production. The direct response ad campaign that resulted would not only liquidate the stranded inventory at full margin, but deliver something more valuable—data. By targeting locked-down consumers through local remnant TV ad slots and optimizing by conversion, Blueland saw unheard-of response rates as high as 17%. The campaign turned into a data goldmine: buyer locations, tablet usage patterns, household sizes, and contact details. Follow-up SMS campaigns would drive 30% reorders. “It built such a franchise in those 36 months,” O’Leary said, “with no retail. Now every retailer wants in.” The lesson? Build your infrastructure to control your data, and you build a business that scales even in chaos. This anecdote set the tone for the keynote: in a volatile world, infrastructure resilience and data control are the new core competencies. The Data Center Power Crisis: “There Is Not a Gig on the Grid” O’Leary

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »