Your Gateway to Power, Energy, Datacenters, Bitcoin and AI

Dive into the latest industry updates, our exclusive Paperboy Newsletter, and curated insights designed to keep you informed. Stay ahead with minimal time spent.

Discover What Matters Most to You

Explore ONMINE’s curated content, from our Paperboy Newsletter to industry-specific insights tailored for energy, Bitcoin mining, and AI professionals.

AI

Lorem Ipsum is simply dummy text of the printing and typesetting industry.

Bitcoin:

Lorem Ipsum is simply dummy text of the printing and typesetting industry.

Datacenter:

Lorem Ipsum is simply dummy text of the printing and typesetting industry.

Energy:

Lorem Ipsum is simply dummy text of the printing and typesetting industry.

Shape
Discover What Matter Most to You

Featured Articles

Avista, PG&E, Ameren AI demonstrations show big potential – but are other utilities ready?

Utilities and system operators are discovering new ways for artificial intelligence and machine learning to help meet reliability threats in the face of growing loads, utilities and analysts say. There has been an “explosion into public consciousness of generative AI models,” according to a 2024 Electric Power Research Institute, or EPRI, paper. The explosion has resulted in huge 2025 AI financial commitments like the $500 billion U.S. Stargate Project and the $206 billion European Union fund. And utilities are beginning to realize the new possibilities. “Utility executives who were skeptical of AI even five years ago are now using cloud computing, drones, and AI in innovative projects,” said Electric Power Research Institute Executive Director, AI and Quantum, Jeremy Renshaw. “Utilities rapid adoption may make what is impossible today standard operating practice in a few years.” Concerns remain that artificial intelligence and machine learning, or AI/ML, algorithms, could bypass human decision-making and cause the reliability failures they are intended to avoid. “But any company that has not taken its internal knowledge base into a generative AI model that can be queried as needed is not leveraging the data it has long paid to store,” said NVIDIA Senior Managing Director Marc Spieler. For now, humans will remain in the loop and AI/ML algorithms will allow better decision-making by making more, and more relevant, data available faster, he added. In real world demonstrations, utilities and software providers are using AI/ML algorithms to improve tasks as varied as nuclear power plant design and electric vehicle, or EV, charging. But utilities and regulators must face the conundrum of making proprietary data more accessible for the new digital intelligence to increase reliability and reduce customer costs while also protecting it.    The old renewed The power system has already put AI/ML algorithms to work in cybersecurity applications

Read More »

Infinite Realms turns fantasy books into living, breathing game worlds with help of AI

Infinite Realms wants to turn beloved fantasy books with big followings into living, breathing game worlds.

The company, which was born from the game studio startup Unleashed Games, wants to license fantasy books from bestselling authors and then turn their creations into games, said Irena Pereira, CEO of Infinite Worlds. It’s not unlike part of the plot of Electronic Arts’ new game, Split Fiction.

Pereira said she came upon the plan with chief marketing officer Vanessa Camones while talking with a seasoned venture capitalist. Unleashed will continue to build a World of Warcraft-like adventure fantasy game called Haven. But Infinite Realms bring together the worlds of fantasy authors, the creativity of small game developers (or even players), and the speedy development of AI tools, Pereira said.

Infinite Realms started out as the back end for Unleashed, but now it is being spun off on its own.

“Infinite Realms is a backend AI-driven engine that can intake book manuscripts and turn them into living, breathing worlds that you can play,” Pereira said. “We’ll be able to license out these intellectual properties to any game studio for them to make their own games based on these IPs. It’s essentially a AI-driven licensing engine for IPs.”

Addressing the industry’s biggest creativity problems

Irena Pereira demos Haven for Rob Foote at GDC 2024.

Pereira said the company is addressing some of the industry’s big problems. Making games is too expensive, original IP is risky, and gamers are getting tired of sequels. Platform fees are taking the profits out of the business. The result is layoffs among game developers and unhappy players.

“The way to solve this problem is to literally hack distribution, by finding new ways to get to players in terms of connecting them with their favorite worlds. These might might not have the economics that are considered worthy of investment by an EA or a Microsoft because the revenues are too small, but they’re the right size for us to get access to the IP that have large built-in audiences,” Pereira said.

She added, “We want to connect fans with their favorite authors.”

And she said that some of the authors are her personal friends. They have sold as many as 40 million books, their IPs have won awards and they’ve been on the New York Times Bestseller lists. Some fans have been obsessed with these IPs for decades and consider them to be core to their own personalities.

“The people who love these books are mega fans and would jump at the opportunity to play any of these stories,” Pereira said. “So we’ve built an engine that can take these books and turn it into a game experience, and then we create this wonderful virtuous cycle where these book lovers go into our game, and then we use that to drive a bigger audiences, which turns back and drives more book sales to properties that we know resonate but might have been sitting on a shelf collecting dust for the last 20 years because they’ve been lost to time.”

Reigniting forgotten worlds

Infinite Realms is combining fantasy book sales and AI and UGC.

The company knows that those communities and the fandom still exists and it’s possible to reignite this in a new generation using games. Using AI, the company can shorten the game development time and lower the costs by leveraging large language models (LLMs) that are custom tailored to each world.

Infinite Realms can take the author’s work and put it into a custom LLM that is partially owned by the author and by the company. That LLM can be licensed out not only to other game studios but to players who want to make their own custom experiences.

It’s also interesting to test how small and efficient an LLM can be and still have intelligence. The LLM has a bunch of lore in it, but it also needs to have a base level of intelligence, or enough data to create a subconsious awareness so to speak so that it knows how to have a conversation about the lore. The LLM can have conversations with the fans, and the fans can feed more data and create more lore for the LLM.

“The possibilities are endless, and the same workflow and partnerships that we developed with Unleashed Games for creating worlds pretty much on the fly can allow us to build games super fast, in as little as six months, because we already have the gameplay sorted out,” Pereira said.

She said that in the past, people would buy books and maybe those books would be adapted into movies and television. Game of Thrones and Wheel of Time are some great examples.

“But with Infinite Realms, we’re building AI powered worlds that you can step inside and interact with some of these characters that you fell in love with when you were 15 years old,” Pereira said. “And by doing that, we create what we’re calling the Netflix of living worlds.”

I noted that the Wheel of Time’s owners have put all 14 books in the series into an LLM that they can make available for user-generated content and players. It can have encyclopedic answers for the fans questions, but it can also serve as the canon police for anyone creating a new experience with the lore.

Things that players create with the tools can be as simple as an ambient video or screensaver on a TV. Or it could be used to create a full game — the full range of potential experiences.

“We can see how this scales, as there are so many other IPs, and you can see us becoming a digital bookshelf,” she said. “You could go from one world to the other on the fly, and we open that up to players to be able to collect these books. So we, in turn, become a digital publisher, where we take these properties that have had them in print, and we’re essentially using them as the start of our transmedia strategy, and then turning them into playable experiences.”

Being respectful of IP

Infinite Realms wants to create AI LLMs around fantasy lore.

All of it will be done with the authors’ approval, and the LLMs themselves can govern what the players can or can’t do. Of course, J.R.R. Tolkien’s The Lord of the Rings is the biggest fantasy franchise, but there are others like Terry Brooks’ The Sword of Shannara, which has reached 40 million fans, down to smaller ones that have sold a few million. The latter are easier and less expensive to work with.

“We essentially become a digital publisher,” Pereira said. “We can deepen our relationships and use the data” to make better decisions on marketing and choosing new IPs.

She added, “This is a great cycle to where we could use our platform to help revive the book publishing industry.”

Pereira is raising a funding round and hopes to be able to accomplish that by getting traction with some of the fantasy authors.

Unleashed Games will likely seek its own money for Haven and Infinite Realms will grow its own business. The companies can use the same technology but still be positioned separately. Infinite Realms has 18 people and it has a partner among AI developers that is also helping.

To judge the market, Infinite Realms is creating ways to test the market for IPs by doing tests with fans.

“I’ve worked with IP holders, and that’s like the No. 1 thing that I’ve been hearing from a lot of IP holders is that they’re trying to find game studios to develop games for their IPs, but they’re unwilling to provide funding for it,” Pereira said.

At the same time, Pereira said, “We’re trying to find a way to re-architect how we think about AI so that it’s respectful of copyright and is constructed with the intention of protecting people’s work.”

Read More »

The Download: gene de-extinction, and Ukraine’s Starlink connection

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology. The short, strange history of gene de-extinction This week saw the release of some fascinating news about some very furry rodents—so-called “woolly mice”—created as part of an experiment to explore how we might one day resurrect the woolly mammoth. The idea of bringing back extinct species has gained traction thanks to advances in sequencing of ancient DNA. This ancient genetic data is deepening our understanding of the past—for instance, by shedding light on interactions among prehistoric humans. But researchers are becoming more ambitious. Rather than just reading ancient DNA, they want to use it—by inserting it into living organisms.
Because this idea is so new and attracting so much attention, I decided it would be useful to create a record of previous attempts to add extinct DNA to living organisms. And since the technology doesn’t have a name, let’s give it one: “chronogenics.” Read the full story. —Antonio Regalado
This article first appeared in The Checkup, MIT Technology Review’s weekly biotech newsletter. To receive it in your inbox every Thursday, and read articles like this first, sign up here.  If you’re interested in de-extinction, why not check out: + How much would you pay to see a woolly mammoth? We spoke to Sara Ord, director of species restoration at Colossal, the world’s first “de-extinction” company, about its big ambitions.+ Colossal is also a de-extinction company, which is trying to resurrect the dodo. Read the full story.+ DNA that was frozen for 2 million years has been sequenced. The ancient DNA fragments come from a Greenland ecosystem where mastodons roamed among flowering plants. It may hold clues to how to survive a warming climate. The must-reads I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology. 1 Ukraine is worried the US could sever its vital Starlink connectionIts satellite internet is vital to Ukraine’s drone operations. (WP $)+ Thankfully, there are alternative providers. (Wired $)+ Ukraine is due to start a fresh round of war-ending negotiations next week. (FT $)+ Meet the radio-obsessed civilian shaping Ukraine’s drone defense. (MIT Technology Review) 2 Israel’s military has trained a powerful AI model on intercepted Palestinian dataThe ChatGPT-like tool can answer queries about the people it’s monitoring. (The Guardian)

3 Donald Trump has suspended tariffs on Canada and MexicoUntil April 2, at least. (Reuters)+ It’s the second time Trump has rolled back import taxes in as many days. (BBC)+ How Trump’s tariffs could drive up the cost of batteries, EVs, and more. (MIT Technology Review) 4 Can someone check on NASA’s Athena lunar lander?While we know it reached the moon, it appears to have toppled over. (NYT $)+ If it remains in an incorrect position, it may be unable to complete its mission. (CNN)+ Its engineers aren’t sure exactly where it is on the moon, either. (NBC News) 5 Shutting down 2G is easier said than doneMillions of vulnerable people around the world still rely on it to communicate. (Rest of World) 6 The hunt for the world’s oldest functional computer codeSpoiler: it may no longer be on Earth. (New Scientist $) 7 Robots are set to compete with humans in a Beijing half marathon🦿My money’s on the flesh and blood competitors. (Insider $)+ Researchers taught robots to run. Now they’re teaching them to walk. (MIT Technology Review) 8 Where did it all go wrong for Skype?It was the world leading video-calling app—until it wasn’t. (The Verge)  9 Dating is out, matchmaking is inWhy swipe when a platform can do the hard work for you? (Wired $)+ Forget dating apps: Here’s how the net’s newest matchmakers help you find love. (MIT Technology Review) 10 Apps are back, baby! 📱It’s like the original smartphone app boom all over again. (Bloomberg $)
Quote of the day
“You can only get so much juice out of every lemon.” —Carl-Benedikt Frey, a professor of AI and work at Oxford University’s Internet Institute, explains why pushing AI as a means of merely increasing productivity won’t always work, the Financial Times reports. The big story The cost of building the perfect wave June 2024
For nearly as long as surfing has existed, surfers have been obsessed with the search for the perfect wave. While this hunt has taken surfers from tropical coastlines to icebergs, these days that search may take place closer to home. That is, at least, the vision presented by developers and boosters in the growing industry of surf pools, spurred by advances in wave-­generating technology that have finally created artificial waves surfers actually want to ride. But there’s a problem: some of these pools are in drought-ridden areas, and face fierce local opposition. At the core of these fights is a question that’s also at the heart of the sport: What is the cost of finding, or now creating, the perfect wave—and who will have to bear it? Read the full story. —Eileen Guo

Read More »

Norway Opens Application for One CO2 Storage Exploration Area

Norway’s Energy Ministry has designated another area of the North Sea for application for licenses to explore the potential of carbon dioxide (CO2) storage. The acreage comprises defined blocks on the Norwegian side of the sea, upstream regulator the Norwegian Offshore Directorate said in an online statement. This is the eighth time acreage is being offered for CO2 storage exploration or exploitation on the Norwegian continental shelf, it noted. The application window for the latest acreage offer closes April 23. “In line with the regulations on transportation and storage of CO2 into subsea reservoirs on the continental shelf, the ministry normally expects to award an exploration license prior to awarding an exploitation license in a relevant area”, the Energy Ministry said separately. Norway has so far awarded 13 CO2 storage licenses: 12 for exploration and one for exploitation. Energy Minister Terje Aasland commented, “The purpose of allocating land is to be able to offer stakeholders in Europe large-scale CO2 storage on commercial terms”. Licensing for CO2 storage is part of Norwegian regulations passed December 2014 to support CO2 storage to mitigate climate change.  “Norway has great potential for storage on the continental shelf”, the ministry added. The Norwegian continental shelf holds a theoretical CO2 storage capacity of 80 billion metric tons, representing about 1,600 years of Norwegian CO2 emissions at current levels, according to a statement by the ministry April 30, 2024. In the latest awards two consortiums with Norway’s majority state-owned Equinor ASA won two exploration licenses in the North Sea. Equinor and London-based Harbour Energy PLC together won a permit straddling blocks 15/8, 15/9, 15/11 and 15/12. The permit, EXL012, lasts four years with three phases. Harbour Energy Norge AS holds a 60 percent stake as operator while Equinor Low Carbon Solution AS has 40 percent, according to a work

Read More »

MP for Truro and Falmouth calls for Cornwall offshore wind strategy

A Labour politician in Cornwall has called for the region to ramp up its domestic offshore wind supply chain. Jayne Kirkham, member of parliament for Truro and Falmouth, said: “At a recent Celtic Sea Power event, I saw just how many brilliant companies are doing amazing things here.” She made the comments months after The Crown Estate entered the second stage of leasing acreage in the Celtic Seas last autumn. “Cornwall has a long history of industrial innovation,” Kirkham said while meeting with marine construction firm MintMech in Penryn. “We’ve got the heritage and the expertise, now we need a strategy that ensures Cornwall maximises the benefits of offshore wind.” The Crown Estate entered the latest phase in its fifth offshore wind leasing round to establish floating offshore wind farms in the Celtic Sea, off the south-west of England and South Wales coast, in August. The second phase of the leasing round was launched, in which bidders must lay out plans to deliver new wind farms and explain how they will benefit local communities. The round has the potential to source up to 4.5GW of new wind capacity and spur investment in the local supply chain. Kirkham expressed hope that Cornish companies will soon be busy on UK projects. She said there are ongoing conversations with the National Energy System Operator (NESO) about ensuring potential wind energy hubs are well connected to the grid. The minister also referenced The Crown Estate’s £50 million Supply Chain Development Fund, which was launched to ensure the UK is prepared to meet offshore wind demands. The first £5m from the fund was awarded in 2024. Kirkham met with directors of Penryn-based marine construction firm MintMech in Jubilee Wharf to discuss the role Cornwall can play in the expansion of the UK’s offshore wind industry.

Read More »

Payroll in USA Oil and Gas Totals $168 Billion in 2024

Payroll in the U.S. oil and gas industry totaled $168 billion in 2024. That’s what the Texas Independent Producers & Royalty Owners Association (TIPRO) said in its latest State of Energy report, which was released this week, highlighting that this figure was “an increase of nearly $5 billion compared to the previous year”. Texas had the highest oil and gas payroll in the country in 2024, according to the report, which pointed out that this figure stood at $62 billion. The report outlined that California was “a distant second” with an oil and gas payroll figure of $15 billion, and that Louisiana was third, with an oil and gas payroll figure of $10 billion. Gasoline Stations with Convenience Stores had the highest U.S. oil and gas payroll by industry figure last year, at $26.8 billion, the report showed. Support Activities for Oil and Gas Operations had the second highest U.S. oil and gas payroll by industry figure in 2024, at $23.9 billion, and Crude Petroleum Extraction had the third highest, at $19.1 billion, the report outlined. The number of U.S. oil and gas businesses totaled 165,110, subject to revisions, TIPRO’s latest report stated. It highlighted that direct oil and natural gas Gross Regional Product exceeded $1 trillion last year and said the U.S. oil and natural gas industry purchased goods and services from over 900 different U.S. industry sectors in the amount of $865 billion in 2024. According to the report, Texas had the highest number of oil and gas businesses in the nation last year, with 23,549. This was followed by California, with 9,486 oil and gas businesses, Florida, with 7,695 oil and gas businesses, Georgia, with 6,453 oil and gas businesses, and New York, with 5,768 oil and gas businesses, the report outlined. The report noted that, in

Read More »

Avista, PG&E, Ameren AI demonstrations show big potential – but are other utilities ready?

Utilities and system operators are discovering new ways for artificial intelligence and machine learning to help meet reliability threats in the face of growing loads, utilities and analysts say. There has been an “explosion into public consciousness of generative AI models,” according to a 2024 Electric Power Research Institute, or EPRI, paper. The explosion has resulted in huge 2025 AI financial commitments like the $500 billion U.S. Stargate Project and the $206 billion European Union fund. And utilities are beginning to realize the new possibilities. “Utility executives who were skeptical of AI even five years ago are now using cloud computing, drones, and AI in innovative projects,” said Electric Power Research Institute Executive Director, AI and Quantum, Jeremy Renshaw. “Utilities rapid adoption may make what is impossible today standard operating practice in a few years.” Concerns remain that artificial intelligence and machine learning, or AI/ML, algorithms, could bypass human decision-making and cause the reliability failures they are intended to avoid. “But any company that has not taken its internal knowledge base into a generative AI model that can be queried as needed is not leveraging the data it has long paid to store,” said NVIDIA Senior Managing Director Marc Spieler. For now, humans will remain in the loop and AI/ML algorithms will allow better decision-making by making more, and more relevant, data available faster, he added. In real world demonstrations, utilities and software providers are using AI/ML algorithms to improve tasks as varied as nuclear power plant design and electric vehicle, or EV, charging. But utilities and regulators must face the conundrum of making proprietary data more accessible for the new digital intelligence to increase reliability and reduce customer costs while also protecting it.    The old renewed The power system has already put AI/ML algorithms to work in cybersecurity applications

Read More »

Infinite Realms turns fantasy books into living, breathing game worlds with help of AI

Infinite Realms wants to turn beloved fantasy books with big followings into living, breathing game worlds.

The company, which was born from the game studio startup Unleashed Games, wants to license fantasy books from bestselling authors and then turn their creations into games, said Irena Pereira, CEO of Infinite Worlds. It’s not unlike part of the plot of Electronic Arts’ new game, Split Fiction.

Pereira said she came upon the plan with chief marketing officer Vanessa Camones while talking with a seasoned venture capitalist. Unleashed will continue to build a World of Warcraft-like adventure fantasy game called Haven. But Infinite Realms bring together the worlds of fantasy authors, the creativity of small game developers (or even players), and the speedy development of AI tools, Pereira said.

Infinite Realms started out as the back end for Unleashed, but now it is being spun off on its own.

“Infinite Realms is a backend AI-driven engine that can intake book manuscripts and turn them into living, breathing worlds that you can play,” Pereira said. “We’ll be able to license out these intellectual properties to any game studio for them to make their own games based on these IPs. It’s essentially a AI-driven licensing engine for IPs.”

Addressing the industry’s biggest creativity problems

Irena Pereira demos Haven for Rob Foote at GDC 2024.

Pereira said the company is addressing some of the industry’s big problems. Making games is too expensive, original IP is risky, and gamers are getting tired of sequels. Platform fees are taking the profits out of the business. The result is layoffs among game developers and unhappy players.

“The way to solve this problem is to literally hack distribution, by finding new ways to get to players in terms of connecting them with their favorite worlds. These might might not have the economics that are considered worthy of investment by an EA or a Microsoft because the revenues are too small, but they’re the right size for us to get access to the IP that have large built-in audiences,” Pereira said.

She added, “We want to connect fans with their favorite authors.”

And she said that some of the authors are her personal friends. They have sold as many as 40 million books, their IPs have won awards and they’ve been on the New York Times Bestseller lists. Some fans have been obsessed with these IPs for decades and consider them to be core to their own personalities.

“The people who love these books are mega fans and would jump at the opportunity to play any of these stories,” Pereira said. “So we’ve built an engine that can take these books and turn it into a game experience, and then we create this wonderful virtuous cycle where these book lovers go into our game, and then we use that to drive a bigger audiences, which turns back and drives more book sales to properties that we know resonate but might have been sitting on a shelf collecting dust for the last 20 years because they’ve been lost to time.”

Reigniting forgotten worlds

Infinite Realms is combining fantasy book sales and AI and UGC.

The company knows that those communities and the fandom still exists and it’s possible to reignite this in a new generation using games. Using AI, the company can shorten the game development time and lower the costs by leveraging large language models (LLMs) that are custom tailored to each world.

Infinite Realms can take the author’s work and put it into a custom LLM that is partially owned by the author and by the company. That LLM can be licensed out not only to other game studios but to players who want to make their own custom experiences.

It’s also interesting to test how small and efficient an LLM can be and still have intelligence. The LLM has a bunch of lore in it, but it also needs to have a base level of intelligence, or enough data to create a subconsious awareness so to speak so that it knows how to have a conversation about the lore. The LLM can have conversations with the fans, and the fans can feed more data and create more lore for the LLM.

“The possibilities are endless, and the same workflow and partnerships that we developed with Unleashed Games for creating worlds pretty much on the fly can allow us to build games super fast, in as little as six months, because we already have the gameplay sorted out,” Pereira said.

She said that in the past, people would buy books and maybe those books would be adapted into movies and television. Game of Thrones and Wheel of Time are some great examples.

“But with Infinite Realms, we’re building AI powered worlds that you can step inside and interact with some of these characters that you fell in love with when you were 15 years old,” Pereira said. “And by doing that, we create what we’re calling the Netflix of living worlds.”

I noted that the Wheel of Time’s owners have put all 14 books in the series into an LLM that they can make available for user-generated content and players. It can have encyclopedic answers for the fans questions, but it can also serve as the canon police for anyone creating a new experience with the lore.

Things that players create with the tools can be as simple as an ambient video or screensaver on a TV. Or it could be used to create a full game — the full range of potential experiences.

“We can see how this scales, as there are so many other IPs, and you can see us becoming a digital bookshelf,” she said. “You could go from one world to the other on the fly, and we open that up to players to be able to collect these books. So we, in turn, become a digital publisher, where we take these properties that have had them in print, and we’re essentially using them as the start of our transmedia strategy, and then turning them into playable experiences.”

Being respectful of IP

Infinite Realms wants to create AI LLMs around fantasy lore.

All of it will be done with the authors’ approval, and the LLMs themselves can govern what the players can or can’t do. Of course, J.R.R. Tolkien’s The Lord of the Rings is the biggest fantasy franchise, but there are others like Terry Brooks’ The Sword of Shannara, which has reached 40 million fans, down to smaller ones that have sold a few million. The latter are easier and less expensive to work with.

“We essentially become a digital publisher,” Pereira said. “We can deepen our relationships and use the data” to make better decisions on marketing and choosing new IPs.

She added, “This is a great cycle to where we could use our platform to help revive the book publishing industry.”

Pereira is raising a funding round and hopes to be able to accomplish that by getting traction with some of the fantasy authors.

Unleashed Games will likely seek its own money for Haven and Infinite Realms will grow its own business. The companies can use the same technology but still be positioned separately. Infinite Realms has 18 people and it has a partner among AI developers that is also helping.

To judge the market, Infinite Realms is creating ways to test the market for IPs by doing tests with fans.

“I’ve worked with IP holders, and that’s like the No. 1 thing that I’ve been hearing from a lot of IP holders is that they’re trying to find game studios to develop games for their IPs, but they’re unwilling to provide funding for it,” Pereira said.

At the same time, Pereira said, “We’re trying to find a way to re-architect how we think about AI so that it’s respectful of copyright and is constructed with the intention of protecting people’s work.”

Read More »

The Download: gene de-extinction, and Ukraine’s Starlink connection

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology. The short, strange history of gene de-extinction This week saw the release of some fascinating news about some very furry rodents—so-called “woolly mice”—created as part of an experiment to explore how we might one day resurrect the woolly mammoth. The idea of bringing back extinct species has gained traction thanks to advances in sequencing of ancient DNA. This ancient genetic data is deepening our understanding of the past—for instance, by shedding light on interactions among prehistoric humans. But researchers are becoming more ambitious. Rather than just reading ancient DNA, they want to use it—by inserting it into living organisms.
Because this idea is so new and attracting so much attention, I decided it would be useful to create a record of previous attempts to add extinct DNA to living organisms. And since the technology doesn’t have a name, let’s give it one: “chronogenics.” Read the full story. —Antonio Regalado
This article first appeared in The Checkup, MIT Technology Review’s weekly biotech newsletter. To receive it in your inbox every Thursday, and read articles like this first, sign up here.  If you’re interested in de-extinction, why not check out: + How much would you pay to see a woolly mammoth? We spoke to Sara Ord, director of species restoration at Colossal, the world’s first “de-extinction” company, about its big ambitions.+ Colossal is also a de-extinction company, which is trying to resurrect the dodo. Read the full story.+ DNA that was frozen for 2 million years has been sequenced. The ancient DNA fragments come from a Greenland ecosystem where mastodons roamed among flowering plants. It may hold clues to how to survive a warming climate. The must-reads I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology. 1 Ukraine is worried the US could sever its vital Starlink connectionIts satellite internet is vital to Ukraine’s drone operations. (WP $)+ Thankfully, there are alternative providers. (Wired $)+ Ukraine is due to start a fresh round of war-ending negotiations next week. (FT $)+ Meet the radio-obsessed civilian shaping Ukraine’s drone defense. (MIT Technology Review) 2 Israel’s military has trained a powerful AI model on intercepted Palestinian dataThe ChatGPT-like tool can answer queries about the people it’s monitoring. (The Guardian)

3 Donald Trump has suspended tariffs on Canada and MexicoUntil April 2, at least. (Reuters)+ It’s the second time Trump has rolled back import taxes in as many days. (BBC)+ How Trump’s tariffs could drive up the cost of batteries, EVs, and more. (MIT Technology Review) 4 Can someone check on NASA’s Athena lunar lander?While we know it reached the moon, it appears to have toppled over. (NYT $)+ If it remains in an incorrect position, it may be unable to complete its mission. (CNN)+ Its engineers aren’t sure exactly where it is on the moon, either. (NBC News) 5 Shutting down 2G is easier said than doneMillions of vulnerable people around the world still rely on it to communicate. (Rest of World) 6 The hunt for the world’s oldest functional computer codeSpoiler: it may no longer be on Earth. (New Scientist $) 7 Robots are set to compete with humans in a Beijing half marathon🦿My money’s on the flesh and blood competitors. (Insider $)+ Researchers taught robots to run. Now they’re teaching them to walk. (MIT Technology Review) 8 Where did it all go wrong for Skype?It was the world leading video-calling app—until it wasn’t. (The Verge)  9 Dating is out, matchmaking is inWhy swipe when a platform can do the hard work for you? (Wired $)+ Forget dating apps: Here’s how the net’s newest matchmakers help you find love. (MIT Technology Review) 10 Apps are back, baby! 📱It’s like the original smartphone app boom all over again. (Bloomberg $)
Quote of the day
“You can only get so much juice out of every lemon.” —Carl-Benedikt Frey, a professor of AI and work at Oxford University’s Internet Institute, explains why pushing AI as a means of merely increasing productivity won’t always work, the Financial Times reports. The big story The cost of building the perfect wave June 2024
For nearly as long as surfing has existed, surfers have been obsessed with the search for the perfect wave. While this hunt has taken surfers from tropical coastlines to icebergs, these days that search may take place closer to home. That is, at least, the vision presented by developers and boosters in the growing industry of surf pools, spurred by advances in wave-­generating technology that have finally created artificial waves surfers actually want to ride. But there’s a problem: some of these pools are in drought-ridden areas, and face fierce local opposition. At the core of these fights is a question that’s also at the heart of the sport: What is the cost of finding, or now creating, the perfect wave—and who will have to bear it? Read the full story. —Eileen Guo

Read More »

Norway Opens Application for One CO2 Storage Exploration Area

Norway’s Energy Ministry has designated another area of the North Sea for application for licenses to explore the potential of carbon dioxide (CO2) storage. The acreage comprises defined blocks on the Norwegian side of the sea, upstream regulator the Norwegian Offshore Directorate said in an online statement. This is the eighth time acreage is being offered for CO2 storage exploration or exploitation on the Norwegian continental shelf, it noted. The application window for the latest acreage offer closes April 23. “In line with the regulations on transportation and storage of CO2 into subsea reservoirs on the continental shelf, the ministry normally expects to award an exploration license prior to awarding an exploitation license in a relevant area”, the Energy Ministry said separately. Norway has so far awarded 13 CO2 storage licenses: 12 for exploration and one for exploitation. Energy Minister Terje Aasland commented, “The purpose of allocating land is to be able to offer stakeholders in Europe large-scale CO2 storage on commercial terms”. Licensing for CO2 storage is part of Norwegian regulations passed December 2014 to support CO2 storage to mitigate climate change.  “Norway has great potential for storage on the continental shelf”, the ministry added. The Norwegian continental shelf holds a theoretical CO2 storage capacity of 80 billion metric tons, representing about 1,600 years of Norwegian CO2 emissions at current levels, according to a statement by the ministry April 30, 2024. In the latest awards two consortiums with Norway’s majority state-owned Equinor ASA won two exploration licenses in the North Sea. Equinor and London-based Harbour Energy PLC together won a permit straddling blocks 15/8, 15/9, 15/11 and 15/12. The permit, EXL012, lasts four years with three phases. Harbour Energy Norge AS holds a 60 percent stake as operator while Equinor Low Carbon Solution AS has 40 percent, according to a work

Read More »

MP for Truro and Falmouth calls for Cornwall offshore wind strategy

A Labour politician in Cornwall has called for the region to ramp up its domestic offshore wind supply chain. Jayne Kirkham, member of parliament for Truro and Falmouth, said: “At a recent Celtic Sea Power event, I saw just how many brilliant companies are doing amazing things here.” She made the comments months after The Crown Estate entered the second stage of leasing acreage in the Celtic Seas last autumn. “Cornwall has a long history of industrial innovation,” Kirkham said while meeting with marine construction firm MintMech in Penryn. “We’ve got the heritage and the expertise, now we need a strategy that ensures Cornwall maximises the benefits of offshore wind.” The Crown Estate entered the latest phase in its fifth offshore wind leasing round to establish floating offshore wind farms in the Celtic Sea, off the south-west of England and South Wales coast, in August. The second phase of the leasing round was launched, in which bidders must lay out plans to deliver new wind farms and explain how they will benefit local communities. The round has the potential to source up to 4.5GW of new wind capacity and spur investment in the local supply chain. Kirkham expressed hope that Cornish companies will soon be busy on UK projects. She said there are ongoing conversations with the National Energy System Operator (NESO) about ensuring potential wind energy hubs are well connected to the grid. The minister also referenced The Crown Estate’s £50 million Supply Chain Development Fund, which was launched to ensure the UK is prepared to meet offshore wind demands. The first £5m from the fund was awarded in 2024. Kirkham met with directors of Penryn-based marine construction firm MintMech in Jubilee Wharf to discuss the role Cornwall can play in the expansion of the UK’s offshore wind industry.

Read More »

Payroll in USA Oil and Gas Totals $168 Billion in 2024

Payroll in the U.S. oil and gas industry totaled $168 billion in 2024. That’s what the Texas Independent Producers & Royalty Owners Association (TIPRO) said in its latest State of Energy report, which was released this week, highlighting that this figure was “an increase of nearly $5 billion compared to the previous year”. Texas had the highest oil and gas payroll in the country in 2024, according to the report, which pointed out that this figure stood at $62 billion. The report outlined that California was “a distant second” with an oil and gas payroll figure of $15 billion, and that Louisiana was third, with an oil and gas payroll figure of $10 billion. Gasoline Stations with Convenience Stores had the highest U.S. oil and gas payroll by industry figure last year, at $26.8 billion, the report showed. Support Activities for Oil and Gas Operations had the second highest U.S. oil and gas payroll by industry figure in 2024, at $23.9 billion, and Crude Petroleum Extraction had the third highest, at $19.1 billion, the report outlined. The number of U.S. oil and gas businesses totaled 165,110, subject to revisions, TIPRO’s latest report stated. It highlighted that direct oil and natural gas Gross Regional Product exceeded $1 trillion last year and said the U.S. oil and natural gas industry purchased goods and services from over 900 different U.S. industry sectors in the amount of $865 billion in 2024. According to the report, Texas had the highest number of oil and gas businesses in the nation last year, with 23,549. This was followed by California, with 9,486 oil and gas businesses, Florida, with 7,695 oil and gas businesses, Georgia, with 6,453 oil and gas businesses, and New York, with 5,768 oil and gas businesses, the report outlined. The report noted that, in

Read More »

Celebrating International Women’s Day with Axis Network’s Emma Behjat

With International Women’s Day on Saturday 8 March, AXIS Network has been promoting gender equity within the UK energy sector. Chairwoman Emma Behjat said: “At AXIS we push for gender equity within the workplace, for the energy sector. But IWD stands for more than just inequality in the workplace, it’s about women as a whole in every aspect of the world. The organisation recently gained sponsorship from OEUK, the NSTA and the NZTC. “With sponsorship from these governing bodies, we hope to accelerate progress by achieving greater awareness of our pledge, increase accountability and a sense of urgency pertaining to the actions needed to achieve gender equity by 2030,” she added. However, this year’s IWD comes amid a pushback against diversity, equity and inclusion (DEI) policies, especially in the US “There has certainly been a lot of push back on DEI, particularly vocally and visibly from the US in our media streams,” Behjat said. “As a (non-practicing) Muslim woman of colour and daughter of an immigrant, I am thankful to live here in the Scotland as our country’s desire to be inclusive remains with support and backing from the government.” But she added that it’s important to understand the road that led to this point. “Put simply, there is a very loud, vocal group who feel they have had their space, their options, their future taken from them to allow others in,” Behjat said. “Those volunteering and working in the DE&I space know that not to be true, the benefits to society and business is tested and positively impactful, but if we really reflect, has the efforts been enough to be inclusive of all?  Have we created space for everyone to be themselves and thrive? “Take for example men; society is much harder on men if they showed ‘feminine’ traits,

Read More »

SeAH Wind brings in three contractors for Hornsea 3 work

SeAH Wind has brought in Mammoet, Glacier Energy, and Hutchinson Engineering as key subcontractors for Ørsted’s 2.9GW Hornsea 3 Offshore Wind Project. SeAH Wind was previously brought in to supply monopiles for the project, which will be situated around 75 miles (120km) off the Norfolk Coast in the North Sea. With commercial production set to commence in the coming months, these subcontractors have been selected to support the operational and logistical services within the 120-acre site located on the South Bank of Teesworks. Mammoet has been appointed to provide self-propelled modular transporters (SPMTs) within the SeAH Wind facility. Their scope includes the transportation of can/cone structures and completed monopiles. By leveraging Mammoet’s expertise in heavy transport solutions in offshore wind site logistics, SeAH Wind ensures a seamless and efficient movement of monopiles, minimising downtime and improving overall production efficiency. Glacier Energy will conduct non-destructive testing (NDT) of welds throughout the manufacturing process, ensuring the highest quality standards. Their rigorous NDT inspections will enhance the reliability and durability of the monopiles, ensuring they meet both Ørsted’s stringent standards and international offshore wind regulations. Hutchinson Engineering has been tasked with supplying secondary steel components for the Hornsea 3 project. SeAH Wind sought a UK-based company capable of delivering these complex parts in compliance with stringent Ørsted drawings and specifications. Hutchinson Engineering’s previous experience on Ørsted projects, combined with their expertise, makes them an ideal partner for this phase of the project. SeAH Wind CEO Chris Sohn said: “These strategic appointments reflect our commitment to delivering market-leading XXXL Monopiles. With these trusted partners, we are confident that we will meet our project goals while upholding the highest standards of quality, safety, and efficiency.” Having begun construction in 2023, Hornsea 3 is expected to enter operations at the end of 2027. A previous £100-million

Read More »

EPL end date: ‘We need to make the changes now,’ says North Sea boss

As the UK government kicked off consultations on a successor tax regime to the Energy Profits Levy (EPL), Anasuria said that changes are needed “now” to save the North Sea. Anasuria Operating Company (AOC) chief executive Richard Beattie said: “I would say we need to make the changes now. “I do think the government is listening, I do think there has been positive engagement, I think we’ve seen shoots of that at the last budget.” Since its introduction in 2022 by then chancellor Rishi Sunak, the EPL, or windfall tax, has been a pain point for the UK oil and gas industry. Initially bringing the headline rate of tax imposed on North Sea operators to 75%, late last year the Labour government hiked rates by 3% as it closed investment incentives previously afforded to oil firms. In the Autumn Budget, chancellor Rachel Reeves also extended the end date of the EPL to March 2030, something it has stuck to as the government launched a consultation on the country’s tax regime and licencing regulations. However, waiting until the start of the next decade to rethink the North Sea’s controversial tax regime may be too late to save investment. “Every drinks reception I’m at, I’m talking to people and they’re talking about putting investment on hold, massive opportunities that they have that they want to invest in but with the uncertainty in terms of the number of changes that just brings natural suspicion,” Beattie added. He argued that, unlike how the UK was previously an appealing place to invest for oil and gas companies “it is simply unattractive” in the current global context. Addressing EPL ‘flaws’ When it launched the consultation process, the government said that authorities will work with industry, communities, trade unions and wider organisations to determine what the new

Read More »

The beautiful link between applying for jobs and our mental health

I don’t think anyone will doubt there’s a very strong distinction between what we do for work, and for our mental health. The jobs that we do, and everything that comes with them, play a huge part in our lives. From the aspirational and emotional side, the desire to develop, succeed and do something meaningful, to the physical side, with the need to provide and enjoy the life we want to live. To achieve the career utopia we seek, we need to head into the sometimes-dreaded recruitment process. Yep, we need to apply for jobs. But while it’s understandably approached with trepidation, what is sometimes overlooked is that through applying for jobs, a chance to work on our wellbeing and mental health arises. Which in the long run gives us better odds in achieving our career goal. © Supplied by Motive Offshore GrouMark Dalgarno, Recruitment Manager, Motive Offshore Group. I know there are challenges in finding a job, but to make that next career step, no matter why you are looking for something new, we must stop and check in with ourselves. This is your career. You’re doing this for you and your world. Others benefit from the work we do, but this is for you first and foremost. No career will progress, no application will be sent, no interview will be attended, without you, in whatever fashion you prefer, asking yourself: “What do I want to do next?” Questions like “Who am I?”, “What do I offer?”, “Why do I want to do this?” will follow. Deep, meaningful, thought-provoking, personal questions. Questions that lead you to answers and actions that positively impact your life. Which, if done meaningfully, can only be good for our mental health. Very rarely do we give ourselves time to check in with what’s really

Read More »

LandBridge at Loss for 2024 but Signals Shift in Q4

LandBridge Company LLC has reported a significant jump in net income for the fourth quarter of 2024 but remained in the red on an annual basis. LandBridge said its net income for the last quarter of the year was $8.2 million, recovering from a $2.8 million net loss for the previous quarter and increasing from the $2.5 million net income reported for the corresponding quarter a year prior. However, for the full year 2024, LandBridge reported a net loss of $41.5 million, compared to a net income of $63.2 million a year before. That is despite full-year revenues growing 51 percent to $110 million. “In 2024, we tripled the size of our land holdings, delivered high-double-digit revenue growth year-over-year, and demonstrated our ability to deliver industry-leading adjusted EBITDA and free cash flow margins. With more than 270,000 acres across the most active oil and natural gas development and production region of the prolific Permian Basin, we are uniquely positioned to capitalize on opportunities in energy and digital infrastructure to create sustainable value for our shareholders”, Jason Long, Chief Executive Officer, said. Revenue for the fourth quarter of 2024 amounted to $36.5 million, compared to $28.5 million for the third quarter of 2024 and $17.5 million for the fourth quarter of 2023. The increase from the previous quarter was mainly due to a rise in easements and other surface-related revenue by $8.2 million, along with oil and gas royalties increasing by $1.6 million and surface use royalties increasing by $0.7 million. These were partially countered by declines of $1.8 million in resource sales and $0.7 million in resource royalties sequentially, the company said. “Our triple-digit revenue growth during the fourth quarter is clear evidence of our momentum across the business. For 2025, we anticipate another year of strong revenue growth and

Read More »

European Commission Schedules New Round for Coordinated Gas Deals

The European Commission has announced a second midterm round for AggregateEU, a mechanism where international natural gas suppliers compete to book demand offered by companies in the EU and its Energy Community partner countries. In the new round buyers can express their demand for July 2025-October 2030. Buyers and sellers have until March 10 to register on the online portal of the AggregateEU service provider, PRISMA. Demand submission is scheduled for March 12-17. Vendors must file expressions of interest from March 18 to 21. The Commission will publish matchmaking results March 26, after which buyers and sellers bilaterally negotiate contracts. “This mid-term matching round reinforces the Commission’s continued commitment to the phase-out of Russian gas supplies to the EU”, the Commission’s Directorate-General for Energy said in an online statement. The first midterm round, whose demand submission phase closed last month, pooled a total of 34 billion cubic meters (1.1 trillion cubic feet) of demand from 19 companies including industrial players. Meanwhile offers totaled 97.4 billion cubic meters, almost triple the demand, according to the Commission. Aggregate EU was initially only meant for the 2023-24 winter season but the EU, citing lessons from the prolonged effects of the energy crisis, has made it a permanent mechanism under “Regulation (EU) 2024/1789 on the internal markets for renewable gas, natural gas and hydrogen”, adopted June 13, 2024. “In previous matching rounds, buyers and sellers could indicate a preferred terminal in the EU for the delivery of liquified natural gas”, the Directorate said when announcing the new midterm round. “To better reflect LNG trade practices and attract additional international suppliers, buyers and sellers can also indicate their preference to have the LNG delivered free-on-board”. Midterm rounds offer six-month contracts for potential suppliers during a buyer-seller partnership of up to five years. “In early 2024, with

Read More »

AI means the end of internet search as we’ve known it

We all know what it means, colloquially, to google something. You pop a few relevant words in a search box and in return get a list of blue links to the most relevant results. Maybe some quick explanations up top. Maybe some maps or sports scores or a video. But fundamentally, it’s just fetching information that’s already out there on the internet and showing it to you, in some sort of structured way.  But all that is up for grabs. We are at a new inflection point. The biggest change to the way search engines have delivered information to us since the 1990s is happening right now. No more keyword searching. No more sorting through links to click. Instead, we’re entering an era of conversational search. Which means instead of keywords, you use real questions, expressed in natural language. And instead of links, you’ll increasingly be met with answers, written by generative AI and based on live information from all across the internet, delivered the same way.  Of course, Google—the company that has defined search for the past 25 years—is trying to be out front on this. In May of 2023, it began testing AI-generated responses to search queries, using its large language model (LLM) to deliver the kinds of answers you might expect from an expert source or trusted friend. It calls these AI Overviews. Google CEO Sundar Pichai described this to MIT Technology Review as “one of the most positive changes we’ve done to search in a long, long time.”
AI Overviews fundamentally change the kinds of queries Google can address. You can now ask it things like “I’m going to Japan for one week next month. I’ll be staying in Tokyo but would like to take some day trips. Are there any festivals happening nearby? How will the surfing be in Kamakura? Are there any good bands playing?” And you’ll get an answer—not just a link to Reddit, but a built-out answer with current results.  More to the point, you can attempt searches that were once pretty much impossible, and get the right answer. You don’t have to be able to articulate what, precisely, you are looking for. You can describe what the bird in your yard looks like, or what the issue seems to be with your refrigerator, or that weird noise your car is making, and get an almost human explanation put together from sources previously siloed across the internet. It’s amazing, and once you start searching that way, it’s addictive.
And it’s not just Google. OpenAI’s ChatGPT now has access to the web, making it far better at finding up-to-date answers to your queries. Microsoft released generative search results for Bing in September. Meta has its own version. The startup Perplexity was doing the same, but with a “move fast, break things” ethos. Literal trillions of dollars are at stake in the outcome as these players jockey to become the next go-to source for information retrieval—the next Google. Not everyone is excited for the change. Publishers are completely freaked out. The shift has heightened fears of a “zero-click” future, where search referral traffic—a mainstay of the web since before Google existed—vanishes from the scene.  I got a vision of that future last June, when I got a push alert from the Perplexity app on my phone. Perplexity is a startup trying to reinvent web search. But in addition to delivering deep answers to queries, it will create entire articles about the news of the day, cobbled together by AI from different sources.  On that day, it pushed me a story about a new drone company from Eric Schmidt. I recognized the story. Forbes had reported it exclusively, earlier in the week, but it had been locked behind a paywall. The image on Perplexity’s story looked identical to one from Forbes. The language and structure were quite similar. It was effectively the same story, but freely available to anyone on the internet. I texted a friend who had edited the original story to ask if Forbes had a deal with the startup to republish its content. But there was no deal. He was shocked and furious and, well, perplexed. He wasn’t alone. Forbes, the New York Times, and Condé Nast have now all sent the company cease-and-desist orders. News Corp is suing for damages.  People are worried about what these new LLM-powered results will mean for our fundamental shared reality. It could spell the end of the canonical answer. It was precisely the nightmare scenario publishers have been so afraid of: The AI was hoovering up their premium content, repackaging it, and promoting it to its audience in a way that didn’t really leave any reason to click through to the original. In fact, on Perplexity’s About page, the first reason it lists to choose the search engine is “Skip the links.” But this isn’t just about publishers (or my own self-interest).  People are also worried about what these new LLM-powered results will mean for our fundamental shared reality. Language models have a tendency to make stuff up—they can hallucinate nonsense. Moreover, generative AI can serve up an entirely new answer to the same question every time, or provide different answers to different people on the basis of what it knows about them. It could spell the end of the canonical answer. But make no mistake: This is the future of search. Try it for a bit yourself, and you’ll see. 

Sure, we will always want to use search engines to navigate the web and to discover new and interesting sources of information. But the links out are taking a back seat. The way AI can put together a well-reasoned answer to just about any kind of question, drawing on real-time data from across the web, just offers a better experience. That is especially true compared with what web search has become in recent years. If it’s not exactly broken (data shows more people are searching with Google more often than ever before), it’s at the very least increasingly cluttered and daunting to navigate.  Who wants to have to speak the language of search engines to find what you need? Who wants to navigate links when you can have straight answers? And maybe: Who wants to have to learn when you can just know?  In the beginning there was Archie. It was the first real internet search engine, and it crawled files previously hidden in the darkness of remote servers. It didn’t tell you what was in those files—just their names. It didn’t preview images; it didn’t have a hierarchy of results, or even much of an interface. But it was a start. And it was pretty good.  Then Tim Berners-Lee created the World Wide Web, and all manner of web pages sprang forth. The Mosaic home page and the Internet Movie Database and Geocities and the Hampster Dance and web rings and Salon and eBay and CNN and federal government sites and some guy’s home page in Turkey. Until finally, there was too much web to even know where to start. We really needed a better way to navigate our way around, to actually find the things we needed.  And so in 1994 Jerry Yang created Yahoo, a hierarchical directory of websites. It quickly became the home page for millions of people. And it was … well, it was okay. TBH, and with the benefit of hindsight, I think we all thought it was much better back then than it actually was. But the web continued to grow and sprawl and expand, every day bringing more information online. Rather than just a list of sites by category, we needed something that actually looked at all that content and indexed it. By the late ’90s that meant choosing from a variety of search engines: AltaVista and AlltheWeb and WebCrawler and HotBot. And they were good—a huge improvement. At least at first.   But alongside the rise of search engines came the first attempts to exploit their ability to deliver traffic. Precious, valuable traffic, which web publishers rely on to sell ads and retailers use to get eyeballs on their goods. Sometimes this meant stuffing pages with keywords or nonsense text designed purely to push pages higher up in search results. It got pretty bad. 
And then came Google. It’s hard to overstate how revolutionary Google was when it launched in 1998. Rather than just scanning the content, it also looked at the sources linking to a website, which helped evaluate its relevance. To oversimplify: The more something was cited elsewhere, the more reliable Google considered it, and the higher it would appear in results. This breakthrough made Google radically better at retrieving relevant results than anything that had come before. It was amazing.  Google CEO Sundar Pichai describes AI Overviews as “one of the most positive changes we’ve done to search in a long, long time.”JENS GYARMATY/LAIF/REDUX For 25 years, Google dominated search. Google was search, for most people. (The extent of that domination is currently the subject of multiple legal probes in the United States and the European Union.)  
But Google has long been moving away from simply serving up a series of blue links, notes Pandu Nayak, Google’s chief scientist for search.  “It’s not just so-called web results, but there are images and videos, and special things for news. There have been direct answers, dictionary answers, sports, answers that come with Knowledge Graph, things like featured snippets,” he says, rattling off a litany of Google’s steps over the years to answer questions more directly.  It’s true: Google has evolved over time, becoming more and more of an answer portal. It has added tools that allow people to just get an answer—the live score to a game, the hours a café is open, or a snippet from the FDA’s website—rather than being pointed to a website where the answer may be.  But once you’ve used AI Overviews a bit, you realize they are different.  Take featured snippets, the passages Google sometimes chooses to highlight and show atop the results themselves. Those words are quoted directly from an original source. The same is true of knowledge panels, which are generated from information stored in a range of public databases and Google’s Knowledge Graph, its database of trillions of facts about the world. While these can be inaccurate, the information source is knowable (and fixable). It’s in a database. You can look it up. Not anymore: AI Overviews can be entirely new every time, generated on the fly by a language model’s predictive text combined with an index of the web. 
“I think it’s an exciting moment where we have obviously indexed the world. We built deep understanding on top of it with Knowledge Graph. We’ve been using LLMs and generative AI to improve our understanding of all that,” Pichai told MIT Technology Review. “But now we are able to generate and compose with that.” The result feels less like a querying a database than like asking a very smart, well-read friend. (With the caveat that the friend will sometimes make things up if she does not know the answer.)  “[The company’s] mission is organizing the world’s information,” Liz Reid, Google’s head of search, tells me from its headquarters in Mountain View, California. “But actually, for a while what we did was organize web pages. Which is not really the same thing as organizing the world’s information or making it truly useful and accessible to you.”  That second concept—accessibility—is what Google is really keying in on with AI Overviews. It’s a sentiment I hear echoed repeatedly while talking to Google execs: They can address more complicated types of queries more efficiently by bringing in a language model to help supply the answers. And they can do it in natural language. 
That will become even more important for a future where search goes beyond text queries. For example, Google Lens, which lets people take a picture or upload an image to find out more about something, uses AI-generated answers to tell you what you may be looking at. Google has even showed off the ability to query live video.  When it doesn’t have an answer, an AI model can confidently spew back a response anyway. For Google, this could be a real problem. For the rest of us, it could actually be dangerous. “We are definitely at the start of a journey where people are going to be able to ask, and get answered, much more complex questions than where we’ve been in the past decade,” says Pichai.  There are some real hazards here. First and foremost: Large language models will lie to you. They hallucinate. They get shit wrong. When it doesn’t have an answer, an AI model can blithely and confidently spew back a response anyway. For Google, which has built its reputation over the past 20 years on reliability, this could be a real problem. For the rest of us, it could actually be dangerous. In May 2024, AI Overviews were rolled out to everyone in the US. Things didn’t go well. Google, long the world’s reference desk, told people to eat rocks and to put glue on their pizza. These answers were mostly in response to what the company calls adversarial queries—those designed to trip it up. But still. It didn’t look good. The company quickly went to work fixing the problems—for example, by deprecating so-called user-generated content from sites like Reddit, where some of the weirder answers had come from. Yet while its errors telling people to eat rocks got all the attention, the more pernicious danger might arise when it gets something less obviously wrong. For example, in doing research for this article, I asked Google when MIT Technology Review went online. It helpfully responded that “MIT Technology Review launched its online presence in late 2022.” This was clearly wrong to me, but for someone completely unfamiliar with the publication, would the error leap out?  I came across several examples like this, both in Google and in OpenAI’s ChatGPT search. Stuff that’s just far enough off the mark not to be immediately seen as wrong. Google is banking that it can continue to improve these results over time by relying on what it knows about quality sources. “When we produce AI Overviews,” says Nayak, “we look for corroborating information from the search results, and the search results themselves are designed to be from these reliable sources whenever possible. These are some of the mechanisms we have in place that assure that if you just consume the AI Overview, and you don’t want to look further … we hope that you will still get a reliable, trustworthy answer.” In the case above, the 2022 answer seemingly came from a reliable source—a story about MIT Technology Review’s email newsletters, which launched in 2022. But the machine fundamentally misunderstood. This is one of the reasons Google uses human beings—raters—to evaluate the results it delivers for accuracy. Ratings don’t correct or control individual AI Overviews; rather, they help train the model to build better answers. But human raters can be fallible. Google is working on that too.  “Raters who look at your experiments may not notice the hallucination because it feels sort of natural,” says Nayak. “And so you have to really work at the evaluation setup to make sure that when there is a hallucination, someone’s able to point out and say, That’s a problem.” The new search Google has rolled out its AI Overviews to upwards of a billion people in more than 100 countries, but it is facing upstarts with new ideas about how search should work. Search Engine GoogleThe search giant has added AI Overviews to search results. These overviews take information from around the web and Google’s Knowledge Graph and use the company’s Gemini language model to create answers to search queries. What it’s good at Google’s AI Overviews are great at giving an easily digestible summary in response to even the most complex queries, with sourcing boxes adjacent to the answers. Among the major options, its deep web index feels the most “internety.” But web publishers fear its summaries will give people little reason to click through to the source material. PerplexityPerplexity is a conversational search engine that uses third-party largelanguage models from OpenAI and Anthropic to answer queries. Perplexity is fantastic at putting together deeper dives in response to user queries, producing answers that are like mini white papers on complex topics. It’s also excellent at summing up current events. But it has gotten a bad rep with publishers, who say it plays fast and loose with their content. ChatGPTWhile Google brought AI to search, OpenAI brought search to ChatGPT. Queries that the model determines will benefit from a web search automatically trigger one, or users can manually select the option to add a web search. Thanks to its ability to preserve context across a conversation, ChatGPT works well for performing searches that benefit from follow-up questions—like planning a vacation through multiple search sessions. OpenAI says users sometimes go “20 turns deep” in researching queries. Of these three, it makes links out to publishers least prominent. When I talked to Pichai about this, he expressed optimism about the company’s ability to maintain accuracy even with the LLM generating responses. That’s because AI Overviews is based on Google’s flagship large language model, Gemini, but also draws from Knowledge Graph and what it considers reputable sources around the web.  “You’re always dealing in percentages. What we have done is deliver it at, like, what I would call a few nines of trust and factuality and quality. I’d say 99-point-few-nines. I think that’s the bar we operate at, and it is true with AI Overviews too,” he says. “And so the question is, are we able to do this again at scale? And I think we are.” There’s another hazard as well, though, which is that people ask Google all sorts of weird things. If you want to know someone’s darkest secrets, look at their search history. Sometimes the things people ask Google about are extremely dark. Sometimes they are illegal. Google doesn’t just have to be able to deploy its AI Overviews when an answer can be helpful; it has to be extremely careful not to deploy them when an answer may be harmful.  “If you go and say ‘How do I build a bomb?’ it’s fine that there are web results. It’s the open web. You can access anything,” Reid says. “But we do not need to have an AI Overview that tells you how to build a bomb, right? We just don’t think that’s worth it.”  But perhaps the greatest hazard—or biggest unknown—is for anyone downstream of a Google search. Take publishers, who for decades now have relied on search queries to send people their way. What reason will people have to click through to the original source, if all the information they seek is right there in the search result?   Rand Fishkin, cofounder of the market research firm SparkToro, publishes research on so-called zero-click searches. As Google has moved increasingly into the answer business, the proportion of searches that end without a click has gone up and up. His sense is that AI Overviews are going to explode this trend.   “If you are reliant on Google for traffic, and that traffic is what drove your business forward, you are in long- and short-term trouble,” he says.  Don’t panic, is Pichai’s message. He argues that even in the age of AI Overviews, people will still want to click through and go deeper for many types of searches. “The underlying principle is people are coming looking for information. They’re not looking for Google always to just answer,” he says. “Sometimes yes, but the vast majority of the times, you’re looking at it as a jumping-off point.”  Reid, meanwhile, argues that because AI Overviews allow people to ask more complicated questions and drill down further into what they want, they could even be helpful to some types of publishers and small businesses, especially those operating in the niches: “You essentially reach new audiences, because people can now express what they want more specifically, and so somebody who specializes doesn’t have to rank for the generic query.”  “I’m going to start with something risky,” Nick Turley tells me from the confines of a Zoom window. Turley is the head of product for ChatGPT, and he’s showing off OpenAI’s new web search tool a few weeks before it launches. “I should normally try this beforehand, but I’m just gonna search for you,” he says. “This is always a high-risk demo to do, because people tend to be particular about what is said about them on the internet.”  He types my name into a search field, and the prototype search engine spits back a few sentences, almost like a speaker bio. It correctly identifies me and my current role. It even highlights a particular story I wrote years ago that was probably my best known. In short, it’s the right answer. Phew?  A few weeks after our call, OpenAI incorporated search into ChatGPT, supplementing answers from its language model with information from across the web. If the model thinks a response would benefit from up-to-date information, it will automatically run a web search (OpenAI won’t say who its search partners are) and incorporate those responses into its answer, with links out if you want to learn more. You can also opt to manually force it to search the web if it does not do so on its own. OpenAI won’t reveal how many people are using its web search, but it says some 250 million people use ChatGPT weekly, all of whom are potentially exposed to it.   “There’s an incredible amount of content on the web. There are a lot of things happening in real time. You want ChatGPT to be able to use that to improve its answers and to be a better super-assistant for you.” Kevin Weil, chief product officer, OpenAI According to Fishkin, these newer forms of AI-assisted search aren’t yet challenging Google’s search dominance. “It does not appear to be cannibalizing classic forms of web search,” he says.  OpenAI insists it’s not really trying to compete on search—although frankly this seems to me like a bit of expectation setting. Rather, it says, web search is mostly a means to get more current information than the data in its training models, which tend to have specific cutoff dates that are often months, or even a year or more, in the past. As a result, while ChatGPT may be great at explaining how a West Coast offense works, it has long been useless at telling you what the latest 49ers score is. No more.  “I come at it from the perspective of ‘How can we make ChatGPT able to answer every question that you have? How can we make it more useful to you on a daily basis?’ And that’s where search comes in for us,” Kevin Weil, the chief product officer with OpenAI, tells me. “There’s an incredible amount of content on the web. There are a lot of things happening in real time. You want ChatGPT to be able to use that to improve its answers and to be able to be a better super-assistant for you.” Today ChatGPT is able to generate responses for very current news events, as well as near-real-time information on things like stock prices. And while ChatGPT’s interface has long been, well, boring, search results bring in all sorts of multimedia—images, graphs, even video. It’s a very different experience.  Weil also argues that ChatGPT has more freedom to innovate and go its own way than competitors like Google—even more than its partner Microsoft does with Bing. Both of those are ad-dependent businesses. OpenAI is not. (At least not yet.) It earns revenue from the developers, businesses, and individuals who use it directly. It’s mostly setting large amounts of money on fire right now—it’s projected to lose $14 billion in 2026, by some reports. But one thing it doesn’t have to worry about is putting ads in its search results as Google does.  “For a while what we did was organize web pages. Which is not really the same thing as organizing the world’s information or making it truly useful and accessible to you,” says Google head of search, Liz Reid.WINNI WINTERMEYER/REDUX Like Google, ChatGPT is pulling in information from web publishers, summarizing it, and including it in its answers. But it has also struck financial deals with publishers, a payment for providing the information that gets rolled into its results. (MIT Technology Review has been in discussions with OpenAI, Google, Perplexity, and others about publisher deals but has not entered into any agreements. Editorial was neither party to nor informed about the content of those discussions.) But the thing is, for web search to accomplish what OpenAI wants—to be more current than the language model—it also has to bring in information from all sorts of publishers and sources that it doesn’t have deals with. OpenAI’s head of media partnerships, Varun Shetty, told MIT Technology Review that it won’t give preferential treatment to its publishing partners. Instead, OpenAI told me, the model itself finds the most trustworthy and useful source for any given question. And that can get weird too. In that very first example it showed me—when Turley ran that name search—it described a story I wrote years ago for Wired about being hacked. That story remains one of the most widely read I’ve ever written. But ChatGPT didn’t link to it. It linked to a short rewrite from The Verge. Admittedly, this was on a prototype version of search, which was, as Turley said, “risky.”  When I asked him about it, he couldn’t really explain why the model chose the sources that it did, because the model itself makes that evaluation. The company helps steer it by identifying—sometimes with the help of users—what it considers better answers, but the model actually selects them.  “And in many cases, it gets it wrong, which is why we have work to do,” said Turley. “Having a model in the loop is a very, very different mechanism than how a search engine worked in the past.” Indeed!  The model, whether it’s OpenAI’s GPT-4o or Google’s Gemini or Anthropic’s Claude, can be very, very good at explaining things. But the rationale behind its explanations, its reasons for selecting a particular source, and even the language it may use in an answer are all pretty mysterious. Sure, a model can explain very many things, but not when that comes to its own answers.  It was almost a decade ago, in 2016, when Pichai wrote that Google was moving from “mobile first” to “AI first”: “But in the next 10 years, we will shift to a world that is AI-first, a world where computing becomes universally available—be it at home, at work, in the car, or on the go—and interacting with all of these surfaces becomes much more natural and intuitive, and above all, more intelligent.”  We’re there now—sort of. And it’s a weird place to be. It’s going to get weirder. That’s especially true as these things we now think of as distinct—querying a search engine, prompting a model, looking for a photo we’ve taken, deciding what we want to read or watch or hear, asking for a photo we wish we’d taken, and didn’t, but would still like to see—begin to merge.  The search results we see from generative AI are best understood as a waypoint rather than a destination. What’s most important may not be search in itself; rather, it’s that search has given AI model developers a path to incorporating real-time information into their inputs and outputs. And that opens up all sorts of possibilities. “A ChatGPT that can understand and access the web won’t just be about summarizing results. It might be about doing things for you. And I think there’s a fairly exciting future there,” says OpenAI’s Weil. “You can imagine having the model book you a flight, or order DoorDash, or just accomplish general tasks for you in the future. It’s just once the model understands how to use the internet, the sky’s the limit.” This is the agentic future we’ve been hearing about for some time now, and the more AI models make use of real-time data from the internet, the closer it gets.  Let’s say you have a trip coming up in a few weeks. An agent that can get data from the internet in real time can book your flights and hotel rooms, make dinner reservations, and more, based on what it knows about you and your upcoming travel—all without your having to guide it. Another agent could, say, monitor the sewage output of your home for certain diseases, and order tests and treatments in response. You won’t have to search for that weird noise your car is making, because the agent in your vehicle will already have done it and made an appointment to get the issue fixed.  “It’s not always going to be just doing search and giving answers,” says Pichai. “Sometimes it’s going to be actions. Sometimes you’ll be interacting within the real world. So there is a notion of universal assistance through it all.” And the ways these things will be able to deliver answers is evolving rapidly now too. For example, today Google can not only search text, images, and even video; it can create them. Imagine overlaying that ability with search across an array of formats and devices. “Show me what a Townsend’s warbler looks like in the tree in front of me.” Or “Use my existing family photos and videos to create a movie trailer of our upcoming vacation to Puerto Rico next year, making sure we visit all the best restaurants and top landmarks.” “We have primarily done it on the input side,” he says, referring to the ways Google can now search for an image or within a video. “But you can imagine it on the output side too.” This is the kind of future Pichai says he is excited to bring online. Google has already showed off a bit of what that might look like with NotebookLM, a tool that lets you upload large amounts of text and have it converted into a chatty podcast. He imagines this type of functionality—the ability to take one type of input and convert it into a variety of outputs—transforming the way we interact with information.  In a demonstration of a tool called Project Astra this summer at its developer conference, Google showed one version of this outcome, where cameras and microphones in phones and smart glasses understand the context all around you—online and off, audible and visual—and have the ability to recall and respond in a variety of ways. Astra can, for example, look at a crude drawing of a Formula One race car and not only identify it, but also explain its various parts and their uses.  But you can imagine things going a bit further (and they will). Let’s say I want to see a video of how to fix something on my bike. The video doesn’t exist, but the information does. AI-assisted generative search could theoretically find that information somewhere online—in a user manual buried in a company’s website, for example—and create a video to show me exactly how to do what I want, just as it could explain that to me with words today. These are the kinds of things that start to happen when you put the entire compendium of human knowledge—knowledge that’s previously been captured in silos of language and format; maps and business registrations and product SKUs; audio and video and databases of numbers and old books and images and, really, anything ever published, ever tracked, ever recorded; things happening right now, everywhere—and introduce a model into all that. A model that maybe can’t understand, precisely, but has the ability to put that information together, rearrange it, and spit it back in a variety of different hopefully helpful ways. Ways that a mere index could not. That’s what we’re on the cusp of, and what we’re starting to see. And as Google rolls this out to a billion people, many of whom will be interacting with a conversational AI for the first time, what will that mean? What will we do differently? It’s all changing so quickly. Hang on, just hang on. 

Read More »

Subsea7 Scores Various Contracts Globally

Subsea 7 S.A. has secured what it calls a “sizeable” contract from Turkish Petroleum Offshore Technology Center AS (TP-OTC) to provide inspection, repair and maintenance (IRM) services for the Sakarya gas field development in the Black Sea. The contract scope includes project management and engineering executed and managed from Subsea7 offices in Istanbul, Türkiye, and Aberdeen, Scotland. The scope also includes the provision of equipment, including two work class remotely operated vehicles, and construction personnel onboard TP-OTC’s light construction vessel Mukavemet, Subsea7 said in a news release. The company defines a sizeable contract as having a value between $50 million and $150 million. Offshore operations will be executed in 2025 and 2026, Subsea7 said. Hani El Kurd, Senior Vice President of UK and Global Inspection, Repair, and Maintenance at Subsea7, said: “We are pleased to have been selected to deliver IRM services for TP-OTC in the Black Sea. This contract demonstrates our strategy to deliver engineering solutions across the full asset lifecycle in close collaboration with our clients. We look forward to continuing to work alongside TP-OTC to optimize gas production from the Sakarya field and strengthen our long-term presence in Türkiye”. North Sea Project Subsea7 also announced the award of a “substantial” contract by Inch Cape Offshore Limited to Seaway7, which is part of the Subsea7 Group. The contract is for the transport and installation of pin-pile jacket foundations and transition pieces for the Inch Cape Offshore Wind Farm. The 1.1-gigawatt Inch Cape project offshore site is located in the Scottish North Sea, 9.3 miles (15 kilometers) off the Angus coast, and will comprise 72 wind turbine generators. Seaway7’s scope of work includes the transport and installation of 18 pin-pile jacket foundations and 54 transition pieces with offshore works expected to begin in 2026, according to a separate news

Read More »

Driving into the future

Welcome to our annual breakthroughs issue. If you’re an MIT Technology Review superfan, you may already know that putting together our 10 Breakthrough Technologies (TR10) list is one of my favorite things we do as a publication. We spend months researching and discussing which technologies will make the list. We try to highlight a mix of items that reflect innovations happening in various fields. We look at consumer technologies, large industrial­-scale projects, biomedical advances, changes in computing, climate solutions, the latest in AI, and more.  We’ve been publishing this list every year since 2001 and, frankly, have a great track record of flagging things that are poised to hit a tipping point. When you look back over the years, you’ll find items like natural-language processing (2001), wireless power (2008), and reusable rockets (2016)—spot-on in terms of horizon scanning. You’ll also see the occasional miss, or moments when maybe we were a little bit too far ahead of ourselves. (See our Magic Leap entry from 2015.) But the real secret of the TR10 is what we leave off the list. It is hard to think of another industry, aside from maybe entertainment, that has as much of a hype machine behind it as tech does. Which means that being too conservative is rarely the wrong call. But it does happen.  Last year, for example, we were going to include robotaxis on the TR10. Autonomous vehicles have been around for years, but 2023 seemed like a real breakthrough moment; both Cruise and Waymo were ferrying paying customers around various cities, with big expansion plans on the horizon. And then, last fall, after a series of mishaps (including an incident when a pedestrian was caught under a vehicle and dragged), Cruise pulled its entire fleet of robotaxis from service. Yikes. 
The timing was pretty miserable, as we were in the process of putting some of the finishing touches on the issue. I made the decision to pull it. That was a mistake.  What followed turned out to be a banner year for the robotaxi. Waymo, which had previously been available only to a select group of beta testers, opened its service to the general public in San Francisco and Los Angeles in 2024. Its cars are now ubiquitous in the City by the Bay, where they have not only become a real competitor to the likes of Uber and Lyft but even created something of a tourist attraction. Which is no wonder, because riding in one is delightful. They are still novel enough to make it feel like a kind of magic. And as you can read, Waymo is just a part of this amazing story. 
The item we swapped into the robotaxi’s place was the Apple Vision Pro, an example of both a hit and a miss. We’d included it because it is truly a revolutionary piece of hardware, and we zeroed in on its micro-OLED display. Yet a year later, it has seemingly failed to find a market fit, and its sales are reported to be far below what Apple predicted. I’ve been covering this field for well over a decade, and I would still argue that the Vision Pro (unlike the Magic Leap vaporware of 2015) is a breakthrough device. But it clearly did not have a breakthrough year. Mea culpa.  Having said all that, I think we have an incredible and thought-provoking list for you this year—from a new astronomical observatory that will allow us to peer into the fourth dimension to new ways of searching the internet to, well, robotaxis. I hope there’s something here for everyone.

Read More »

Oil Holds at Highest Levels Since October

Crude oil futures slightly retreated but continue to hold at their highest levels since October, supported by colder weather in the Northern Hemisphere and China’s economic stimulus measures. That’s what George Pavel, General Manager at Naga.com Middle East, said in a market analysis sent to Rigzone this morning, adding that Brent and WTI crude “both saw modest declines, yet the outlook remains bullish as colder temperatures are expected to increase demand for heating oil”. “Beijing’s fiscal stimulus aims to rejuvenate economic activity and consumer demand, further contributing to fuel consumption expectations,” Pavel said in the analysis. “This economic support from China could help sustain global demand for crude, providing upward pressure on prices,” he added. Looking at supply, Pavel noted in the analysis that “concerns are mounting over potential declines in Iranian oil production due to anticipated sanctions and policy changes under the incoming U.S. administration”. “Forecasts point to a reduction of 300,000 barrels per day in Iranian output by the second quarter of 2025, which would weigh on global supply and further support prices,” he said. “Moreover, the U.S. oil rig count has decreased, indicating a potential slowdown in future output,” he added. “With supply-side constraints contributing to tightening global inventories, this situation is likely to reinforce the current market optimism, supporting crude prices at elevated levels,” Pavel continued. “Combined with the growing demand driven by weather and economic factors, these supply dynamics point to a favorable environment for oil prices in the near term,” Pavel went on to state. Rigzone has contacted the Trump transition team and the Iranian ministry of foreign affairs for comment on Pavel’s analysis. At the time of writing, neither have responded to Rigzone’s request yet. In a separate market analysis sent to Rigzone earlier this morning, Antonio Di Giacomo, Senior Market Analyst at

Read More »

What to expect from NaaS in 2025

Shamus McGillicuddy, vice president of research at EMA, says that network execs today have a fuller understanding of the potential benefits of NaaS, beyond simply a different payment model. NaaS can deliver access to new technologies faster and keep enterprises up-to-date as technologies evolve over time; it can help mitigate skills gaps for organizations facing a shortage of networking talent. For example, in a retail scenario, an organization can offload deployment and management of its Wi-Fi networks at all of its stores to a NaaS vendor, freeing up IT staffers for higher-level activities. Also, it can help organizations manage rapidly fluctuating demands on the network, he says. 2. Frameworks help drive adoption Industry standards can help accelerate the adoption of new technologies. MEF, a nonprofit industry forum, has developed a framework that combines standardized service definitions, extensive automation frameworks, security certifications, and multi-cloud integration capabilities—all aimed at enabling service providers to deliver what MEF calls a true cloud experience for network services. The blueprint serves as a guide for building an automated, federated ecosystem where enterprises can easily consume NaaS services from providers. It details the APIs, service definitions, and certification programs that MEF has developed to enable this vision. The four components of NaaS, according to the blueprint, are on-demand automated transport services, SD-WAN overlays and network slicing for application assurance, SASE-based security, and multi-cloud on-ramps. 3. The rise of campus/LAN NaaS Until very recently, the most popular use cases for NaaS were on-demand WAN connectivity, multi-cloud connectivity, SD-WAN, and SASE. However, campus/LAN NaaS, which includes both wired and wireless networks, has emerged as the breakout star in the overall NaaS market. Dell’Oro Group analyst Sian Morgan predicts: “In 2025, Campus NaaS revenues will grow over eight times faster than the overall LAN market. Startups offering purpose-built CNaaS technology will

Read More »

UK battery storage industry ‘back on track’

UK battery storage investor Gresham House Energy Storage Fund (LON:GRID) has said the industry is “back on track” as trading conditions improved, particularly in December. The UK’s largest fund specialising in battery energy storage systems (BESS) highlighted improvements in service by the UK government’s National Energy System Operator (NESO) as well as its renewed commitment to to the sector as part of clean power aims by 2030. It also revealed that revenues exceeding £60,000 per MW of electricity its facilities provided in the second half of 2024 meant it would meet or even exceed revenue targets. This comes after the fund said it had faced a “weak revenue environment” in the first part of the year. In April it reported a £110 million loss compared to a £217m profit the previous year and paused dividends. Fund manager Ben Guest said the organisation was “working hard” on refinancing  and a plan to “re-instate dividend payments”. In a further update, the fund said its 40MW BESS project at Shilton Lane, 11 miles from Glasgow, was  fully built and in the final stages of the NESO compliance process which expected to complete in February 2025. Fund chair John Leggate welcomed “solid progress” in company’s performance, “as well as improvements in NESO’s control room, and commitment to further change, that should see BESS increasingly well utilised”. He added: “We thank our shareholders for their patience as the battery storage industry gets back on track with the most environmentally appropriate and economically competitive energy storage technology (Li-ion) being properly prioritised. “Alongside NESO’s backing of BESS, it is encouraging to see the government’s endorsement of a level playing field for battery storage – the only proven, commercially viable technology that can dynamically manage renewable intermittency at national scale.” Guest, who in addition to managing the fund is also

Read More »

New open-source math model Light-R1-32B surpasses equivalent DeepSeek performance with only $1000 in training costs

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More A team of researchers has introduced Light-R1-32B, a new open-source AI model optimized for solving advanced math problems, making it available on Hugging Face under a permissive Apache 2.0 license — free for enterprises and researchers to take, deploy, fine-tune or modify as they wish, even for commercial purposes. The 32-billion parameter (number of model settings) model surpasses the performance of similarly sized (and even larger) open source models such as DeepSeek-R1-Distill-Llama-70B and DeepSeek-R1-Distill-Qwen-32B on third-party benchmark the American Invitational Mathematics Examination (AIME), which contains 15 math problems designed for extremely advanced students and has an allotted time limit of 3 hours for human users. Developed by Liang Wen, Fenrui Xiao, Xin He, Yunke Cai, Qi An, Zhenyu Duan, Yimin Du, Junchen Liu, Lifu Tang, Xiaowei Lv, Haosheng Zou, Yongchao Deng, Shousheng Jia, and Xiangzheng Zhang, the model surpasses previous open-source alternatives on competitive math benchmarks. Incredibly, the researchers completed the model’s training in fewer than six hours on 12 Nvidia H800 GPUs at an estimated total cost of $1,000. This makes Light-R1-32B one of the most accessible and practical approaches for developing high-performing math-specialized AI models. However, it’s important to remember the model was trained on a variant of Alibaba’s open source Qwen 2.5-32B-Instruct, which itself is presumed to have had much higher upfront training costs. Alongside the model, the team has released its training datasets, training scripts, and evaluation tools, providing a transparent and accessible framework for building math-focused AI models. The arrival of Light-R1-32B follows other similar efforts from rivals such as Microsoft with its Orca-Math series. A new math king emerges Light-R1-32B is designed to tackle complex mathematical reasoning, particularly on the AIME (American Invitational Mathematics Examination)

Read More »

Overcome Failing Document Ingestion & RAG Strategies with Agentic Knowledge Distillation

Introduction

Many generative AI use cases still revolve around Retrieval Augmented Generation (RAG), yet consistently fall short of user expectations. Despite the growing body of research on RAG improvements and even adding Agents into the process, many solutions still fail to return exhaustive results, miss information that is critical but infrequently mentioned in the documents, require multiple search iterations, and generally struggle to reconcile key themes across multiple documents. To top it all off, many implementations still rely on cramming as much “relevant” information as possible into the model’s context window alongside detailed system and user prompts. Reconciling all this information often exceeds the model’s cognitive capacity and compromises response quality and consistency.

This is where our Agentic Knowledge Distillation + Pyramid Search Approach comes into play. Instead of chasing the best chunking strategy, retrieval algorithm, or inference-time reasoning method, my team, Jim Brown, Mason Sawtell, Sandi Besen, and I, take an agentic approach to document ingestion.

We leverage the full capability of the model at ingestion time to focus exclusively on distilling and preserving the most meaningful information from the document dataset. This fundamentally simplifies the RAG process by allowing the model to direct its reasoning abilities toward addressing the user/system instructions rather than struggling to understand formatting and disparate information across document chunks. 

We specifically target high-value questions that are often difficult to evaluate because they have multiple correct answers or solution paths. These cases are where traditional RAG solutions struggle most and existing RAG evaluation datasets are largely insufficient for testing this problem space. For our research implementation, we downloaded annual and quarterly reports from the last year for the 30 companies in the DOW Jones Industrial Average. These documents can be found through the SEC EDGAR website. The information on EDGAR is accessible and able to be downloaded for free or can be queried through EDGAR public searches. See the SEC privacy policy for additional details, information on the SEC website is “considered public information and may be copied or further distributed by users of the web site without the SEC’s permission”. We selected this dataset for two key reasons: first, it falls outside the knowledge cutoff for the models evaluated, ensuring that the models cannot respond to questions based on their knowledge from pre-training; second, it’s a close approximation for real-world business problems while allowing us to discuss and share our findings using publicly available data. 

While typical RAG solutions excel at factual retrieval where the answer is easily identified in the document dataset (e.g., “When did Apple’s annual shareholder’s meeting occur?”), they struggle with nuanced questions that require a deeper understanding of concepts across documents (e.g., “Which of the DOW companies has the most promising AI strategy?”). Our Agentic Knowledge Distillation + Pyramid Search Approach addresses these types of questions with much greater success compared to other standard approaches we tested and overcomes limitations associated with using knowledge graphs in RAG systems. 

In this article, we’ll cover how our knowledge distillation process works, key benefits of this approach, examples, and an open discussion on the best way to evaluate these types of systems where, in many cases, there is no singular “right” answer.

Building the pyramid: How Agentic Knowledge Distillation works

Image by author and team depicting pyramid structure for document ingestion. Robots meant to represent agents building the pyramid.

Overview

Our knowledge distillation process creates a multi-tiered pyramid of information from the raw source documents. Our approach is inspired by the pyramids used in deep learning computer vision-based tasks, which allow a model to analyze an image at multiple scales. We take the contents of the raw document, convert it to markdown, and distill the content into a list of atomic insights, related concepts, document abstracts, and general recollections/memories. During retrieval it’s possible to access any or all levels of the pyramid to respond to the user request. 

How to distill documents and build the pyramid: 

Convert documents to Markdown: Convert all raw source documents to Markdown. We’ve found models process markdown best for this task compared to other formats like JSON and it is more token efficient. We used Azure Document Intelligence to generate the markdown for each page of the document, but there are many other open-source libraries like MarkItDown which do the same thing. Our dataset included 331 documents and 16,601 pages. 

Extract atomic insights from each page: We process documents using a two-page sliding window, which allows each page to be analyzed twice. This gives the agent the opportunity to correct any potential mistakes when processing the page initially. We instruct the model to create a numbered list of insights that grows as it processes the pages in the document. The agent can overwrite insights from the previous page if they were incorrect since it sees each page twice. We instruct the model to extract insights in simple sentences following the subject-verb-object (SVO) format and to write sentences as if English is the second language of the user. This significantly improves performance by encouraging clarity and precision. Rolling over each page multiple times and using the SVO format also solves the disambiguation problem, which is a huge challenge for knowledge graphs. The insight generation step is also particularly helpful for extracting information from tables since the model captures the facts from the table in clear, succinct sentences. Our dataset produced 216,931 total insights, about 13 insights per page and 655 insights per document.

Distilling concepts from insights: From the detailed list of insights, we identify higher-level concepts that connect related information about the document. This step significantly reduces noise and redundant information in the document while preserving essential information and themes. Our dataset produced 14,824 total concepts, about 1 concept per page and 45 concepts per document. 

Creating abstracts from concepts: Given the insights and concepts in the document, the LLM writes an abstract that appears both better than any abstract a human would write and more information-dense than any abstract present in the original document. The LLM generated abstract provides incredibly comprehensive knowledge about the document with a small token density that carries a significant amount of information. We produce one abstract per document, 331 total.

Storing recollections/memories across documents: At the top of the pyramid we store critical information that is useful across all tasks. This can be information that the user shares about the task or information the agent learns about the dataset over time by researching and responding to tasks. For example, we can store the current 30 companies in the DOW as a recollection since this list is different from the 30 companies in the DOW at the time of the model’s knowledge cutoff. As we conduct more and more research tasks, we can continuously improve our recollections and maintain an audit trail of which documents these recollections originated from. For example, we can keep track of AI strategies across companies, where companies are making major investments, etc. These high-level connections are super important since they reveal relationships and information that are not apparent in a single page or document.

Sample subset of insights extracted from IBM 10Q, Q3 2024 (page 4)

We store the text and embeddings for each layer of the pyramid (pages and up) in Azure PostgreSQL. We originally used Azure AI Search, but switched to PostgreSQL for cost reasons. This required us to write our own hybrid search function since PostgreSQL doesn’t yet natively support this feature. This implementation would work with any vector database or vector index of your choosing. The key requirement is to store and efficiently retrieve both text and vector embeddings at any level of the pyramid. 

This approach essentially creates the essence of a knowledge graph, but stores information in natural language, the way an LLM natively wants to interact with it, and is more efficient on token retrieval. We also let the LLM pick the terms used to categorize each level of the pyramid, this seemed to let the model decide for itself the best way to describe and differentiate between the information stored at each level. For example, the LLM preferred “insights” to “facts” as the label for the first level of distilled knowledge. Our goal in doing this was to better understand how an LLM thinks about the process by letting it decide how to store and group related information. 

Using the pyramid: How it works with RAG & Agents

At inference time, both traditional RAG and agentic approaches benefit from the pre-processed, distilled information ingested in our knowledge pyramid. The pyramid structure allows for efficient retrieval in both the traditional RAG case, where only the top X related pieces of information are retrieved or in the Agentic case, where the Agent iteratively plans, retrieves, and evaluates information before returning a final response. 

The benefit of the pyramid approach is that information at any and all levels of the pyramid can be used during inference. For our implementation, we used PydanticAI to create a search agent that takes in the user request, generates search terms, explores ideas related to the request, and keeps track of information relevant to the request. Once the search agent determines there’s sufficient information to address the user request, the results are re-ranked and sent back to the LLM to generate a final reply. Our implementation allows a search agent to traverse the information in the pyramid as it gathers details about a concept/search term. This is similar to walking a knowledge graph, but in a way that’s more natural for the LLM since all the information in the pyramid is stored in natural language.

Depending on the use case, the Agent could access information at all levels of the pyramid or only at specific levels (e.g. only retrieve information from the concepts). For our experiments, we did not retrieve raw page-level data since we wanted to focus on token efficiency and found the LLM-generated information for the insights, concepts, abstracts, and recollections was sufficient for completing our tasks. In theory, the Agent could also have access to the page data; this would provide additional opportunities for the agent to re-examine the original document text; however, it would also significantly increase the total tokens used. 

Here is a high-level visualization of our Agentic approach to responding to user requests:

Image created by author and team providing an overview of the agentic research & response process

Results from the pyramid: Real-world examples

To evaluate the effectiveness of our approach, we tested it against a variety of question categories, including typical fact-finding questions and complex cross-document research and analysis tasks. 

Fact-finding (spear fishing): 

These tasks require identifying specific information or facts that are buried in a document. These are the types of questions typical RAG solutions target but often require many searches and consume lots of tokens to answer correctly. 

Example task: “What was IBM’s total revenue in the latest financial reporting?”

Example response using pyramid approach: “IBM’s total revenue for the third quarter of 2024 was $14.968 billion [ibm-10q-q3-2024.pdf, pg. 4]

Total tokens used to research and generate response

This result is correct (human-validated) and was generated using only 9,994 total tokens, with 1,240 tokens in the generated final response. 

Complex research and analysis: 

These tasks involve researching and understanding multiple concepts to gain a broader understanding of the documents and make inferences and informed assumptions based on the gathered facts.

Example task: “Analyze the investments Microsoft and NVIDIA are making in AI and how they are positioning themselves in the market. The report should be clearly formatted.”

Example response:

Response generated by the agent analyzing AI investments and positioning for Microsoft and NVIDIA.

The result is a comprehensive report that executed quickly and contains detailed information about each of the companies. 26,802 total tokens were used to research and respond to the request with a significant percentage of them used for the final response (2,893 tokens or ~11%). These results were also reviewed by a human to verify their validity.

Snippet indicating total token usage for the task

Example task: “Create a report on analyzing the risks disclosed by the various financial companies in the DOW. Indicate which risks are shared and unique.”

Example response:

Part 1 of response generated by the agent on disclosed risks.

Part 2 of response generated by the agent on disclosed risks.

Similarly, this task was completed in 42.7 seconds and used 31,685 total tokens, with 3,116 tokens used to generate the final report. 

Snippet indicating total token usage for the task

These results for both fact-finding and complex analysis tasks demonstrate that the pyramid approach efficiently creates detailed reports with low latency using a minimal amount of tokens. The tokens used for the tasks carry dense meaning with little noise allowing for high-quality, thorough responses across tasks.

Benefits of the pyramid: Why use it?

Overall, we found that our pyramid approach provided a significant boost in response quality and overall performance for high-value questions. 

Some of the key benefits we observed include: 

Reduced model’s cognitive load: When the agent receives the user task, it retrieves pre-processed, distilled information rather than the raw, inconsistently formatted, disparate document chunks. This fundamentally improves the retrieval process since the model doesn’t waste its cognitive capacity on trying to break down the page/chunk text for the first time. 

Superior table processing: By breaking down table information and storing it in concise but descriptive sentences, the pyramid approach makes it easier to retrieve relevant information at inference time through natural language queries. This was particularly important for our dataset since financial reports contain lots of critical information in tables. 

Improved response quality to many types of requests: The pyramid enables more comprehensive context-aware responses to both precise, fact-finding questions and broad analysis based tasks that involve many themes across numerous documents. 

Preservation of critical context: Since the distillation process identifies and keeps track of key facts, important information that might appear only once in the document is easier to maintain. For example, noting that all tables are represented in millions of dollars or in a particular currency. Traditional chunking methods often cause this type of information to slip through the cracks. 

Optimized token usage, memory, and speed: By distilling information at ingestion time, we significantly reduce the number of tokens required during inference, are able to maximize the value of information put in the context window, and improve memory use. 

Scalability: Many solutions struggle to perform as the size of the document dataset grows. This approach provides a much more efficient way to manage a large volume of text by only preserving critical information. This also allows for a more efficient use of the LLMs context window by only sending it useful, clear information.

Efficient concept exploration: The pyramid enables the agent to explore related information similar to navigating a knowledge graph, but does not require ever generating or maintaining relationships in the graph. The agent can use natural language exclusively and keep track of important facts related to the concepts it’s exploring in a highly token-efficient and fluid way. 

Emergent dataset understanding: An unexpected benefit of this approach emerged during our testing. When asking questions like “what can you tell me about this dataset?” or “what types of questions can I ask?”, the system is able to respond and suggest productive search topics because it has a more robust understanding of the dataset context by accessing higher levels in the pyramid like the abstracts and recollections. 

Beyond the pyramid: Evaluation challenges & future directions

Challenges

While the results we’ve observed when using the pyramid search approach have been nothing short of amazing, finding ways to establish meaningful metrics to evaluate the entire system both at ingestion time and during information retrieval is challenging. Traditional RAG and Agent evaluation frameworks often fail to address nuanced questions and analytical responses where many different responses are valid.

Our team plans to write a research paper on this approach in the future, and we are open to any thoughts and feedback from the community, especially when it comes to evaluation metrics. Many of the existing datasets we found were focused on evaluating RAG use cases within one document or precise information retrieval across multiple documents rather than robust concept and theme analysis across documents and domains. 

The main use cases we are interested in relate to broader questions that are representative of how businesses actually want to interact with GenAI systems. For example, “tell me everything I need to know about customer X” or “how do the behaviors of Customer A and B differ? Which am I more likely to have a successful meeting with?”. These types of questions require a deep understanding of information across many sources. The answers to these questions typically require a person to synthesize data from multiple areas of the business and think critically about it. As a result, the answers to these questions are rarely written or saved anywhere which makes it impossible to simply store and retrieve them through a vector index in a typical RAG process. 

Another consideration is that many real-world use cases involve dynamic datasets where documents are consistently being added, edited, and deleted. This makes it difficult to evaluate and track what a “correct” response is since the answer will evolve as the available information changes. 

Future directions

In the future, we believe that the pyramid approach can address some of these challenges by enabling more effective processing of dense documents and storing learned information as recollections. However, tracking and evaluating the validity of the recollections over time will be critical to the system’s overall success and remains a key focus area for our ongoing work. 

When applying this approach to organizational data, the pyramid process could also be used to identify and assess discrepancies across areas of the business. For example, uploading all of a company’s sales pitch decks could surface where certain products or services are being positioned inconsistently. It could also be used to compare insights extracted from various line of business data to help understand if and where teams have developed conflicting understandings of topics or different priorities. This application goes beyond pure information retrieval use cases and would allow the pyramid to serve as an organizational alignment tool that helps identify divergences in messaging, terminology, and overall communication. 

Conclusion: Key takeaways and why the pyramid approach matters

The knowledge distillation pyramid approach is significant because it leverages the full power of the LLM at both ingestion and retrieval time. Our approach allows you to store dense information in fewer tokens which has the added benefit of reducing noise in the dataset at inference. Our approach also runs very quickly and is incredibly token efficient, we are able to generate responses within seconds, explore potentially hundreds of searches, and on average use

Read More »

Generative AI Is Declarative

ChatGPT launched in 2022 and kicked off the Generative Ai boom. In the two years since, academics, technologists, and armchair experts have written libraries worth of articles on the technical underpinnings of generative AI and about the potential capabilities of both current and future generative AI models.

Surprisingly little has been written about how we interact with these tools—the human-AI interface. The point where we interact with AI models is at least as important as the algorithms and data that create them. “There is no success where there is no possibility of failure, no art without the resistance of the medium” (Raymond Chandler). In that vein, it’s useful to examine human-AI interaction and the strengths and weaknesses inherent in that interaction. If we understand the “resistance in the medium” then product managers can make smarter decisions about how to incorporate generative AI into their products. Executives can make smarter decisions about what capabilities to invest in. Engineers and designers can build around the tools’ limitations and showcase their strengths. Everyday people can know when to use generative AI and when not to.

Imagine walking into a restaurant and ordering a cheeseburger. You don’t tell the chef how to grind the beef, how hot to set the grill, or how long to toast the bun. Instead, you simply describe what you want: “I’d like a cheeseburger, medium rare, with lettuce and tomato.” The chef interprets your request, handles the implementation, and delivers the desired outcome. This is the essence of declarative interaction—focusing on the what rather than the how.

Now, imagine interacting with a Large Language Model (LLM) like ChatGPT. You don’t have to provide step-by-step instructions for how to generate a response. Instead, you describe the result you’re looking for: “A user story that lets us implement A/B testing for the Buy button on our website.” The LLM interprets your prompt, fills in the missing details, and delivers a response. Just like ordering a cheeseburger, this is a declarative mode of interaction.

Explaining the steps to make a cheeseburger is an imperative interaction. Our LLM prompts sometimes feel imperative. We might phrase our prompts like a question: ”What is the tallest mountain on earth?” This is equivalent to describing “the answer to the question ‘What is the tallest mountain on earth?’” We might phrase our prompt as a series of instructions: ”Write a summary of the attached report, then read it as if you are a product manager, then type up some feedback on the report.” But, again, we’re describing the result of a process with some context for what that process is. In this case, it is a sequence of descriptive results—the report then the feedback.

This is a more useful way to think about LLMs and generative AI. In some ways it is more accurate; the neural network model behind the curtain doesn’t explain why or how it produced one output instead of another. More importantly though, the limitations and strengths of generative AI make more sense and become more predictable when we think of these models as declarative.

LLMs as a declarative mode of interaction

Computer scientists use the term “declarative” to describe coding languages. SQL is one of the most common. The code describes the output table and the procedures in the database figure out how to retrieve and combine the data to produce the result. LLMs share many of the benefits of declarative languages like SQL or declarative interactions like ordering a cheeseburger.

Focus on desired outcome: Just as you describe the cheeseburger you want, you describe the output you want from the LLM. For example, “Summarize this article in three bullet points” focuses on the result, not the process.

Abstraction of implementation: When you order a cheeseburger, you don’t need to know how the chef prepares it. When submitting SQL code to a server, the server figures out where the data lives, how to fetch it, and how to aggregate it based on your description. You as the user don’t need to know how. With LLMs, you don’t need to know how the model generates the response. The underlying mechanisms are abstracted away.

Filling in missing details: If you don’t specify onions on your cheeseburger, the chef won’t include them. If you don’t specify a field in your SQL code, it won’t show up in the output table. This is where LLMs differ slightly from declarative coding languages like SQL. If you ask ChatGPT to create an image of “a cheeseburger with lettuce and tomato” it may also show the burger on a sesame seed bun or include pickles, even if that wasn’t in your description. The details you omit are inferred by the LLM using the “average” or “most likely” detail depending on the context, with a bit of randomness thrown in. Ask for the cheeseburger image six times; it may show you three burgers with cheddar cheese, two with Swiss, and one with pepper jack.

Like other forms of declarative interaction, LLMs share one key limitation. If your description is vague, ambiguous, or lacks enough detail, then the result may not be what you hoped to see. It is up to the user to describe the result with sufficient detail.

This explains why we often iterate to get what we’re looking for when using LLMs and generative AI. Going back to our cheeseburger analogy, the process to generate a cheeseburger from an LLM may look like this.

“Make me a cheeseburger, medium rare, with lettuce and tomatoes.” The result also has pickles and uses cheddar cheese. The bun is toasted. There’s mayo on the top bun.

“Make the same thing but this time no pickles, use pepper jack cheese, and a sriracha mayo instead of plain mayo.” The result now has pepper jack, no pickles. The sriracha mayo is applied to the bottom bun and the bun is no longer toasted.

“Make the same thing again, but this time, put the sriracha mayo on the top bun. The buns should be toasted.” Finally, you have the cheeseburger you’re looking for.

This example demonstrates one of the main points of friction with human-AI interaction. Human beings are really bad at describing what they want with sufficient detail on the first attempt.

When we asked for a cheeseburger, we had to refine our description to be more specific (the type of cheese). In the second generation, some of the inferred details (whether the bun was toasted) changed from one iteration to the next, so then we had to add that specificity to our description as well. Iteration is an important part of AI-human generation.

Insight: When using generative AI, we need to design an iterative human-AI interaction loop that enables people to discover the details of what they want and refine their descriptions accordingly.

To iterate, we need to evaluate the results. Evaluation is extremely important with generative AI. Say you’re using an LLM to write code. You can evaluate the code quality if you know enough to understand it or if you can execute it and inspect the results. On the other hand, hypothetical questions can’t be tested. Say you ask ChatGPT, “What if we raise our product prices by 5 percent?” A seasoned expert could read the output and know from experience if a recommendation doesn’t take into account important details. If your product is property insurance, then increasing premiums by 5 percent may mean pushback from regulators, something an experienced veteran of the industry would know. For non-experts in a topic, there’s no way to tell if the “average” details inferred by the model make sense for your specific use case. You can’t test and iterate.

Insight: LLMs work best when the user can evaluate the result quickly, whether through execution or through prior knowledge.

The examples so far involve general knowledge. We all know what a cheeseburger is. When you start asking about non-general information—like when you can make dinner reservations next week—you delve into new points of friction.

In the next section we’ll think about different types of information, what we can expect the AI to “know”, and how this impacts human-AI interaction.

What did the AI know, and when did it know it?

Above, I explained how generative AI is a declarative mode of interaction and how that helps understand its strengths and weaknesses. Here, I’ll identify how different types of information create better or worse human-AI interactions.

Understanding the information available

When we describe what we want to an LLM, and when it infers missing details from our description, it draws from different sources of information. Understanding these sources of information is important. Here’s a useful taxonomy for information types:

General information used to train the base model.

Non-general information that the base model is not aware of.

Fresh information that is new or changes rapidly, like stock prices or current events.

Non-public information, like facts about you and where you live or about your company, its employees, its processes, or its codebase.

General information vs. non-general information

LLMs are built on a massive corpus of written word data. A large part of GPT-3 was trained on a combination of books, journals, Wikipedia, Reddit, and CommonCrawl (an open-source repository of web crawl data). You can think of the models as a highly compressed version of that data, organized in a gestalt manner—all the like things are close together. When we submit a prompt, the model takes the words we use (and any words added to the prompt behind the scenes) and finds the closest set of related words based on how those things appear in the data corpus. So when we say “cheeseburger” it knows that word is related to “bun” and “tomato” and “lettuce” and “pickles” because they all occur in the same context throughout many data sources. Even when we don’t specify pickles, it uses this gestalt approach to fill in the blanks.

This training information is general information, and a good rule of thumb is this: if it was in Wikipedia a year ago then the LLM “knows” about it. There could be new articles on Wikipedia, but that didn’t exist when the model was trained. The LLM doesn’t know about that unless told.

Now, say you’re a company using an LLM to write a product requirements document for a new web app feature. Your company, like most companies, is full of its own lingo. It has its own lore and history scattered across thousands of Slack messages, emails, documents, and some tenured employees who remember that one meeting in Q1 last year. The LLM doesn’t know any of that. It will infer any missing details from general information. You need to supply everything else. If it wasn’t in Wikipedia a year ago, the LLM doesn’t know about it. The resulting product requirements document may be full of general facts about your industry and product but could lack important details specific to your firm.

This is non-general information. This includes personal info, anything kept behind a log-in or paywall, and non-digital information. This non-general information permeates our lives, and incorporating it is another source of friction when working with generative AI.

Non-general information can be incorporated into a generative AI application in three ways:

Through model fine-tuning (supplying a large corpus to the base model to expand its reference data).

Retrieved and fed it to the model at query time (e.g., the retrieval augmented generation or “RAG” technique).

Supplied by the user in the prompt.

Insight: When designing any human-AI interactions, you should think about what non-general information is required, where you will get it, and how you will expose it to the AI.

Fresh information

Any information that changes in real-time or is new can be called fresh information. This includes new facts like current events but also frequently changing facts like your bank account balance. If the fresh information is available in a database or some searchable source, then it needs to be retrieved and incorporated into the application. To retrieve the information from a database, the LLM must create a query, which may require specific details that the user didn’t include.

Here’s an example. I have a chatbot that gives information on the stock market. You, the user, type the following: “What is the current price of Apple? Has it been increasing or decreasing recently?”

The LLM doesn’t have the current price of Apple in its training data. This is fresh, non-general information. So, we need to retrieve it from a database.

The LLM can read “Apple”, know that you’re talking about the computer company, and that the ticker symbol is AAPL. This is all general information.

What about the “increasing or decreasing” part of the prompt? You did not specify over what period—increasing in the past day, month, year? In order to construct a database query, we need more detail. LLMs are bad at knowing when to ask for detail and when to fill it in. The application could easily pull the wrong data and provide an unexpected or inaccurate answer. Only you know what these details should be, depending on your intent. You must be more specific in your prompt.

A designer of this LLM application can improve the user experience by specifying required parameters for expected queries. We can ask the user to explicitly input the time range or design the chatbot to ask for more specific details if not provided. In either case, we need to have a specific type of query in mind and explicitly design how to handle it. The LLM will not know how to do this unassisted.

Insight: If a user is expecting a more specific type of output, you need to explicitly ask for enough detail. Too little detail could produce a poor quality output.

Non-public information

Incorporating non-public information into an LLM prompt can be done if that information can be accessed in a database. This introduces privacy issues (should the LLM be able to access my medical records?) and complexity when incorporating multiple non-public sources of information.

Let’s say I have a chatbot that helps you make dinner reservations. You, the user, type the following: “Help me make dinner reservations somewhere with good Neapolitan pizza.”

The LLM knows what a Neapolitan pizza is and can infer that “dinner” means this is for an evening meal.

To do this task well, it needs information about your location, the restaurants near you and their booking status, or even personal details like dietary restrictions. Assuming all that non-public information is available in databases, bringing them all together into the prompt takes a lot of engineering work.

Even if the LLM could find the “best” restaurant for you and book the reservation, can you be confident it has done that correctly? You never specified how many people you need a reservation for. Since only you know this information, the application needs to ask for it upfront.

If you’re designing this LLM-based application, you can make some thoughtful choices to help with these problems. We could ask about a user’s dietary restrictions when they sign up for the app. Other information, like the user’s schedule that evening, can be given in a prompting tip or by showing the default prompt option “show me reservations for two for tomorrow at 7PM”. Promoting tips may not feel as automagical as a bot that does it all, but they are a straightforward way to collect and integrate the non-public information.

Some non-public information is large and can’t be quickly collected and processed when the prompt is given. These need to be fine-tuned in batch or retrieved at prompt time and incorporated. A chatbot that answers information about a company’s HR policies can obtain this information from a corpus of non-public HR documents. You can fine-tune the model ahead of time by feeding it the corpus. Or you can implement a retrieval augmented generation technique, searching a corpus for relevant documents and summarizing the results. Either way, the response will only be as accurate and up-to-date as the corpus itself.

Insight: When designing an AI application, you need to be aware of non-public information and how to retrieve it. Some of that information can be pulled from databases. Some needs to come from the user, which may require prompt suggestions or explicitly asking.

If you understand the types of information and treat human-AI interaction as declarative, you can more easily predict which AI applications will work and which ones won’t. In the next section we’ll look at OpenAI’s Operator and deep research products. Using this framework, we can see where these applications fall short, where they work well, and why.

Critiquing OpenAI’s Operator and deep research through a declarative lens

I have now explained how thinking of generative AI as declarative helps us understand its strengths and weaknesses. I also identified how different types of information create better or worse human-AI interactions.

Now I’ll apply these ideas by critiquing two recent products from OpenAI—Operator and deep research. It’s important to be honest about the shortcomings of AI applications. Bigger models trained on more data or using new techniques might one day solve some issues with generative AI. But other issues arise from the human-AI interaction itself and can only be addressed by making appropriate design and product choices.

These critiques demonstrate how the framework can help identify where the limitations are and how to address them.

The limitations of Operator

Journalist Casey Newton of Platformer reviewed Operator in an article that was largely positive. Newton has covered AI extensively and optimistically. Still, Newton couldn’t help but point out some of Operator’s frustrating limitations.

[Operator] can take action on your behalf in ways that are new to AI systems — but at the moment it requires a lot of hand-holding, and may cause you to throw up your hands in frustration. 

My most frustrating experience with Operator was my first one: trying to order groceries. “Help me buy groceries on Instacart,” I said, expecting it to ask me some basic questions. Where do I live? What store do I usually buy groceries from? What kinds of groceries do I want? 

It didn’t ask me any of that. Instead, Operator opened Instacart in the browser tab and begin searching for milk in grocery stores located in Des Moines, Iowa.

The prompt “Help me buy groceries on Instacart,” viewed declaratively, describes groceries being purchased using Instacart. It doesn’t have a lot of the information someone would need to buy groceries, like what exactly to buy, when it would be delivered, and to where.

It’s worth repeating: LLMs are not good at knowing when to ask additional questions unless explicitly programmed to do so in the use case. Newton gave a vague request and expected follow-up questions. Instead, the LLM filled in all the missing details with the “average”. The average item was milk. The average location was Des Moines, Iowa. Newton doesn’t mention when it was scheduled to be delivered, but if the “average” delivery time is tomorrow, then that was likely the default.

If we engineered this application specifically for ordering groceries, keeping in mind the declarative nature of AI and the information it “knows”, then we could make thoughtful design choices that improve functionality. We would need to prompt the user to specify when and where they want groceries up front (non-public information). With that information, we could find an appropriate grocery store near them. We would need access to that grocery store’s inventory (more non-public information). If we have access to the user’s previous orders, we could also pre-populate a cart with items typical to their order. If not, we may add a few suggested items and guide them to add more. By limiting the use case, we only have to deal with two sources of non-public information. This is a more tractable problem than Operator’s “agent that does it all” approach.

Newton also mentions that this process took eight minutes to complete, and “complete” means that Operator did everything up to placing the order. This is a long time with very little human-in-the-loop iteration. Like we said before, an iteration loop is very important for human-AI interaction. A better-designed application would generate smaller steps along the way and provide more frequent interaction. We could prompt the user to describe what to add to their shopping list. The user might say, “Add barbeque sauce to the list,” and see the list update. If they see a vinegar-based barbecue sauce, they can refine that by saying, “Replace that with a barbeque sauce that goes well with chicken,” and might be happier when it’s replaced by a honey barbecue sauce. These frequent iterations make the LLM a creative tool rather than a does-it-all agent. The does-it-all agent looks automagical in marketing, but a more guided approach provides more utility with a less frustrating and more delightful experience.

Elsewhere in the article, Newton gives an example of a prompt that Operator performed well: “Put together a lesson plan on the Great Gatsby for high school students, breaking it into readable chunks and then creating assignments and connections tied to the Common Core learning standard.” This prompt describes an output using much more specificity. It also solely relies on general information—the Great Gatsby, the Common Core standard, and a general sense of what assignments are. The general-information use case lends itself better to AI generation, and the prompt is explicit and detailed in its request. In this case, very little guidance was given to create the prompt, so it worked better. (In fact, this prompt comes from Ethan Mollick who has used it to evaluate AI chatbots.)

This is the risk of general-purpose AI applications like Operator. The quality of the result relies heavily on the use case and specificity provided by the user. An application with a more specific use case allows for more design guidance and can produce better output more reliably.

The limitations of deep research

Newton also reviewed deep research, which, according to OpenAI’s website, is an “agent that uses reasoning to synthesize large amounts of online information and complete multi-step research tasks for you.”

Deep research came out after Newton’s review of Operator. Newton chose an intentionally tricky prompt that prods at some of the tool’s limitations regarding fresh information and non-general information: “I wanted to see how OpenAI’s agent would perform given that it was researching a story that was less than a day old, and for which much of the coverage was behind paywalls that the agent would not be able to access. And indeed, the bot struggled more than I expected.”

Near the end of the article, Newton elaborates on some of the shortcomings he noticed with deep research.

OpenAI’s deep research suffers from the same design problem that almost all AI products have: its superpowers are completely invisible and must be harnessed through a frustrating process of trial and error.

Generally speaking, the more you already know about something, the more useful I think deep research is. This may be somewhat counterintuitive; perhaps you expected that an AI agent would be well suited to getting you up to speed on an important topic that just landed on your lap at work, for example. 

In my early tests, the reverse felt true. Deep research excels for drilling deep into subjects you already have some expertise in, letting you probe for specific pieces of information, types of analysis, or ideas that are new to you.

The “frustrating trial and error” shows a mismatch between Newton’s expectations and a necessary aspect of many generative AI applications. A good response requires more information than the user will probably give in the first attempt. The challenge is to design the application and set the user’s expectations so that this interaction is not frustrating but exciting.

Newton’s more poignant criticism is that the application requires already knowing something about the topic for it to work well. From the perspective of our framework, this makes sense. The more you know about a topic, the more detail you can provide. And as you iterate, having knowledge about a topic helps you observe and evaluate the output. Without the ability to describe it well or evaluate the results, the user is less likely to use the tool to generate good output.

A version of deep research designed for lawyers to perform legal research could be powerful. Lawyers have an extensive and common vocabulary for describing legal matters, and they’re more likely to see a result and know if it makes sense. Generative AI tools are fallible, though. So, the tool should focus on a generation-evaluation loop rather than writing a final draft of a legal document.

The article also highlights many improvements compared to Operator. Most notably, the bot asked clarifying questions. This is the most impressive aspect of the tool. Undoubtedly, it helps that deep search has a focused use-case of retrieving and summarizing general information instead of a does-it-all approach. Having a focused use case narrows the set of likely interactions, letting you design better guidance into the prompt flow.

Good application design with generative AI

Designing effective generative AI applications requires thoughtful consideration of how users interact with the technology, the types of information they need, and the limitations of the underlying models. Here are some key principles to guide the design of generative AI tools:

1. Constrain the input and focus on providing details

Applications are inputs and outputs. We want the outputs to be useful and pleasant. By giving a user a conversational chatbot interface, we allow for a vast surface area of potential inputs, making it a challenge to guarantee useful outputs. One strategy is to limit or guide the input to a more manageable subset.

For example, FigJam, a collaborative whiteboarding tool, uses pre-set template prompts for timelines, Gantt charts, and other common whiteboard artifacts. This provides some structure and predictability to the inputs. Users still have the freedom to describe further details like color or the content for each timeline event. This approach ensures that the AI has enough specificity to generate meaningful outputs while giving users creative control.

2. Design frequent iteration and evaluation into the tool

Iterating in a tight generation-evaluation loop is essential for refining outputs and ensuring they meet user expectations. OpenAI’s Dall-E is great at this. Users quickly iterate on image prompts and refine their descriptions to add additional detail. If you type “a picture of a cheeseburger on a plate”, you may then add more detail by specifying “with pepperjack cheese”.

AI code generating tools work well because users can run a generated code snippet immediately to see if it works, enabling rapid iteration and validation. This quick evaluation loop produces better results and a better coder experience. 

Designers of generative AI applications should pull the user in the loop early, often, in a way that is engaging rather than frustrating. Designers should also consider the user’s knowledge level. Users with domain expertise can iterate more effectively.

Referring back to the FigJam example, the prompts and icons in the app quickly communicate “this is what we call a mind map” or “this is what we call a gantt chart” for users who want to generate these artifacts but don’t know the terms for them. Giving the user some basic vocabulary can help them better generate desired results quickly with less frustration.

3. Be mindful of the types of information needed

LLMs excel at tasks involving general knowledge already in the base training set. For example, writing class assignments involves absorbing general information, synthesizing it, and producing a written output, so LLMs are very well-suited for that task.

Use cases that require non-general information are more complex. Some questions the designer and engineer should ask include:

Does this application require fresh information? Maybe this is knowledge of current events or a user’s current bank account balance. If so, that information needs to be retrieved and incorporated into the model.

How much non-general information does the LLM need to know? If it’s a lot of information—like a corpus of company documentation and communication—then the model may need to be fine tuned in batch ahead of time. If the information is relatively small, a retrieval augmented generation (RAG) approach at query time may suffice. 

How many sources of non-general information—small and finite or potentially infinite? General purpose agents like Operator face the challenge of potentially infinite non-general information sources. Depending on what the user requires, it could need to access their contacts, restaurant reservation lists, financial data, or even other people’s calendars. A single-purpose restaurant reservation chatbot may only need access to Yelp, OpenTable, and the user’s calendar. It’s much easier to reconcile access and authentication for a handful of known data sources.

Is there context-specific information that can only come from the user? Consider our restaurant reservation chatbot. Is the user making reservations for just themselves? Probably not. “How many people and who” is a detail that only the user can provide, an example of non-public information that only the user knows. We shouldn’t expect the user to provide this information upfront and unguided. Instead, we can use prompt suggestions so they include the information. We may even be able to design the LLM to ask these questions when the detail is not provided.

4. Focus on specific use cases

Broad, all-purpose chatbots often struggle to deliver consistent results due to the complexity and variability of user needs. Instead, focus on specific use cases where the AI’s shortcomings can be mitigated through thoughtful design.

Narrowing the scope helps us address many of the issues above.

We can identify common requests for the use case and incorporate those into prompt suggestions.

We can design an iteration loop that works well with the type of thing we’re generating.

We can identify sources of non-general information and devise solutions to incorporate it into the model or prompt.

5. Translation or summary tasks work well

A common task for ChatGPT is to rewrite something in a different style, explain what some computer code is doing, or summarize a long document. These tasks involve converting a set of information from one form to another.

We have the same concerns about non-general information and context. For instance, a Chatbot asked to explain a code script doesn’t know the system that script is part of unless that information is provided.

But in general, the task of transforming or summarizing information is less prone to missing details. By definition, you have provided the details it needs. The result should have the same information in a different or more condensed form.

The exception to the rules

There is a case when it doesn’t matter if you break any or all of these rules—when you’re just having fun. LLMs are creative tools by nature. They can be an easel to paint on, a sandbox to build in, a blank sheet to scribe. Iteration is still important; the user wants to see the thing they’re creating as they create it. But unexpected results due to lack of information or omitted details may add to the experience. If you ask for a cheeseburger recipe, you might get some funny or interesting ingredients. If the stakes are low and the process is its own reward, don’t worry about the rules.

Read More »

Enhancing AI agents with long-term memory: Insights into LangMem SDK, Memobase and the A-MEM Framework

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More AI agents can automate many tasks that enterprises want to perform. One downside, though, is that they tend to be forgetful. Without long-term memory, agents must either finish a task in a single session or be constantly re-prompted.  So, as enterprises continue to explore use cases for AI agents and how to implement them safely, the companies enabling development of agents must consider how to make them less forgetful. Long-term memory will make agents much more valuable in a workflow, able to remember instructions even for complex tasks that require several turns to complete. Manvinder Singh, VP of AI product management at Redis, told VentureBeat that memory makes agents more robust.  “Agentic memory is crucial for enhancing [agents’] efficiency and capabilities since LLMs are inherently stateless — they don’t remember things like prompts, responses or chat histories,” Singh said in an email. “Memory allows AI agents to recall past interactions, retain information and maintain context to deliver more coherent, personalized responses, and more impactful autonomy.” Companies like LangChain have begun offering options to extend agentic memory. LangChain’s LangMem SDK helps developers build agents with tools “to extract information from conversation, optimize agent behavior through prompt updates, and maintain long-term memory about behaviors, facts, and events.” Other options include Memobase, an open-source tool launched in January to give agents “user-centric memory” so apps remember and adapt. CrewAI also has tooling around long-term agentic memory, while OpenAI’s Swarm requires users to bring their memory model.  Mike Mason, chief AI officer at tech consultancy Thoughtworks, told VentureBeat in an email that better agentic memory changes how companies use agents. “Memory transforms AI agents from simple, reactive tools into dynamic, adaptive assistants,” Mason said. “Without

Read More »

The Download: AI can cheat at chess, and the future of search

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology. AI reasoning models can cheat to win chess games The news: Facing defeat in chess, the latest generation of AI reasoning models sometimes cheat without being instructed to do so. The finding suggests that the next wave of AI models could be more likely to seek out deceptive ways of doing whatever they’ve been asked to do. And worst of all? There’s no simple way to fix it. How they did it: Researchers from the AI research organization Palisade Research instructed seven large language models to play hundreds of games of chess against Stockfish, a powerful open-source chess engine. The research suggests that the more sophisticated the AI model, the more likely it is to spontaneously try to “hack” the game in an attempt to beat its opponent. Older models would do this kind of thing only after explicit nudging from the team. Read the full story.
—Rhiannon Williams
MIT Technology Review Narrated: AI search could break the web At its best, AI search can infer a user’s intent, amplify quality content, and synthesize information from diverse sources. But if AI search becomes our primary portal to the web, it threatens to disrupt an already precarious digital economy.Today, the production of content online depends on a fragile set of incentives tied to virtual foot traffic: ads, subscriptions, donations, sales, or brand exposure. By shielding the web behind an all-knowing chatbot, AI search could deprive creators of the visits and “eyeballs” they need to survive. This is our latest story to be turned into a MIT Technology Review Narrated podcast, which we’re publishing each week on Spotify and Apple Podcasts. Just navigate to MIT Technology Review Narrated on either platform, and follow us to get all our new content as it’s released. Join us to discuss disruption in the AI model market Join MIT Technology Review’s AI writers as they discuss the latest upheaval in the AI marketplace. Editor in chief Mat Honan will be joined by Will Douglas Heaven, our senior AI editor, and James O’Donnell, our AI and hardware reporter, to dive into how new developments in AI model development are reshaping competition, raising questions for investors, challenging industry assumptions, and accelerating timelines for AI adoption and innovation. Make sure you register here—it kicks off at 12.30pm ET today. The must-reads

I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology. 1 A judge has denied Elon Musk’s attempt to halt OpenAI’s for-profit plansBut other aspects of the lawsuit have been permitted to proceed. (CNBC)+ The court will fast-track a trial later this year. (FT $) 2 ChatGPT isn’t going to dethrone GoogleAt least not any time soon. (Insider $)+ AI means the end of internet search as we’ve known it. (MIT Technology Review)3 Beijing is going all in on AIChina is treating the technology as key to boosting its economy—and lessening its reliance on overseas trade. (WSJ $)+ DeepSeek is, naturally, the jewel in its crown. (Reuters)+ Four Chinese AI startups to watch beyond DeepSeek. (MIT Technology Review) 4  A pair of reinforcement learning pioneers have won the Turing AwardAndrew Barto and Richard Sutton’s technique underpins today’s chatbots. (Axios)+ The former professor and student wrote the literal book on reinforcement learning. (NYT $)+ The pair will share a million dollar prize. (New Scientist $) 5 US apps are being used to groom and exploit minors in Colombia Better internet service is making it easier for sex traffickers to find and sell young girls. (Bloomberg $)+ An AI companion site is hosting sexually charged conversations with underage celebrity bots. (MIT Technology Review)6 Europe is on high alert following undersea cable attacksIt’s unclear whether improving Russian-American relations will help. (The Guardian)+ These stunning images trace ships’ routes as they move. (MIT Technology Review) 7 Jeff Bezos is cracking the whip at Blue OriginHe’s implementing a tougher, Amazon-like approach to catch up with rival SpaceX. (FT $) 8 All hail the return of DiggThe news aggregator is staging a comeback, over a decade after it was split into parts. (Inc)+ It’s been acquired by its original founder Kevin Rose and Reddit co-founder Alexis Ohanian. (TechCrunch)+ Digg wants to resurrect the community-first social platform. (The Verge)+ How to fix the internet. (MIT Technology Review) 9 We’re still learning about how memory works 🧠Greater understanding could pave the way to better treatments for anxiety and chronic pain. (Knowable Magazine)+ A memory prosthesis could restore memory in people with damaged brains. (MIT Technology Review)
10 AI can’t replace your personalityDespite what Big Tech seems to be peddling. (NY Mag $)
Quote of the day “That is just a lot of money [to invest] on a handshake.” —US District Judge Yvonne Gonzalez Rogers questions why Elon Musk invested tens of millions of dollars in OpenAI without a written contract, Associated Press reports. The big story People are worried that AI will take everyone’s jobs. We’ve been here before.
January 2024 It was 1938, and the pain of the Great Depression was still very real. Unemployment in the US was around 20%. New machinery was transforming factories and farms, and everyone was worried about jobs. Were the impressive technological achievements that were making life easier for many also destroying jobs and wreaking havoc on the economy? To make sense of it all, Karl T. Compton, the president of MIT from 1930 to 1948 and one of the leading scientists of the day, wrote in the December 1938 issue of this publication about the “Bogey of Technological Unemployment.” His essay concisely framed the debate over jobs and technical progress in a way that remains relevant, especially given today’s fears over the impact of artificial intelligence. It’s a worthwhile reminder that worries over the future of jobs are not new and are best addressed by applying an understanding of economics, rather than conjuring up genies and monsters. Read the full story.—David Rotman

Read More »

Salesforce launches Agentforce 2dx, letting AI run autonomously across enterprise systems

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Salesforce announced Agentforce 2dx today, a major update to its digital labor platform that enables autonomous AI agents to work proactively behind the scenes across enterprise systems without constant human supervision. The announcement marks a substantial evolution from the company’s previous approach, where agents primarily operated within chat interfaces and required explicit user prompts. The new system aims to embed AI agents that can anticipate needs, monitor data changes, and take action autonomously across any business process. “Companies today have more work than workers, and Agentforce is stepping in to fill the gap,” said Adam Evans, EVP and GM of Salesforce’s AI Platform, in a statement sent to VentureBeat. “By extending digital labor beyond CRM, we’re making it easier than ever for businesses to embed agentic AI into any workflow or application.” Autonomous AI agents now work without human prompting The most transformative aspect of today’s announcement is the shift from purely reactive AI interactions to proactive agents that can operate autonomously in the background. This change allows companies to deploy AI labor that doesn’t just wait for user commands but actively monitors systems and initiates processes when needed. “What surprised me the most is the pace — the speed of creation and speed of iteration,” said Rob Seaman, SVP and GM of Slack, in a recent interview with VentureBeat. “The number of people now that can create technology that can help solve employee or customer problems has dramatically expanded because it’s topics and instructions, not C++, Java, Python, or Hack.” The announcement comes at a critical moment in the evolution of AI agents, as enterprises move beyond experimentation toward deploying autonomous systems that can handle increasingly complex workflows without human

Read More »

Avista, PG&E, Ameren AI demonstrations show big potential – but are other utilities ready?

Utilities and system operators are discovering new ways for artificial intelligence and machine learning to help meet reliability threats in the face of growing loads, utilities and analysts say. There has been an “explosion into public consciousness of generative AI models,” according to a 2024 Electric Power Research Institute, or EPRI, paper. The explosion has resulted in huge 2025 AI financial commitments like the $500 billion U.S. Stargate Project and the $206 billion European Union fund. And utilities are beginning to realize the new possibilities. “Utility executives who were skeptical of AI even five years ago are now using cloud computing, drones, and AI in innovative projects,” said Electric Power Research Institute Executive Director, AI and Quantum, Jeremy Renshaw. “Utilities rapid adoption may make what is impossible today standard operating practice in a few years.” Concerns remain that artificial intelligence and machine learning, or AI/ML, algorithms, could bypass human decision-making and cause the reliability failures they are intended to avoid. “But any company that has not taken its internal knowledge base into a generative AI model that can be queried as needed is not leveraging the data it has long paid to store,” said NVIDIA Senior Managing Director Marc Spieler. For now, humans will remain in the loop and AI/ML algorithms will allow better decision-making by making more, and more relevant, data available faster, he added. In real world demonstrations, utilities and software providers are using AI/ML algorithms to improve tasks as varied as nuclear power plant design and electric vehicle, or EV, charging. But utilities and regulators must face the conundrum of making proprietary data more accessible for the new digital intelligence to increase reliability and reduce customer costs while also protecting it.    The old renewed The power system has already put AI/ML algorithms to work in cybersecurity applications

Read More »

Infinite Realms turns fantasy books into living, breathing game worlds with help of AI

Infinite Realms wants to turn beloved fantasy books with big followings into living, breathing game worlds.

The company, which was born from the game studio startup Unleashed Games, wants to license fantasy books from bestselling authors and then turn their creations into games, said Irena Pereira, CEO of Infinite Worlds. It’s not unlike part of the plot of Electronic Arts’ new game, Split Fiction.

Pereira said she came upon the plan with chief marketing officer Vanessa Camones while talking with a seasoned venture capitalist. Unleashed will continue to build a World of Warcraft-like adventure fantasy game called Haven. But Infinite Realms bring together the worlds of fantasy authors, the creativity of small game developers (or even players), and the speedy development of AI tools, Pereira said.

Infinite Realms started out as the back end for Unleashed, but now it is being spun off on its own.

“Infinite Realms is a backend AI-driven engine that can intake book manuscripts and turn them into living, breathing worlds that you can play,” Pereira said. “We’ll be able to license out these intellectual properties to any game studio for them to make their own games based on these IPs. It’s essentially a AI-driven licensing engine for IPs.”

Addressing the industry’s biggest creativity problems

Irena Pereira demos Haven for Rob Foote at GDC 2024.

Pereira said the company is addressing some of the industry’s big problems. Making games is too expensive, original IP is risky, and gamers are getting tired of sequels. Platform fees are taking the profits out of the business. The result is layoffs among game developers and unhappy players.

“The way to solve this problem is to literally hack distribution, by finding new ways to get to players in terms of connecting them with their favorite worlds. These might might not have the economics that are considered worthy of investment by an EA or a Microsoft because the revenues are too small, but they’re the right size for us to get access to the IP that have large built-in audiences,” Pereira said.

She added, “We want to connect fans with their favorite authors.”

And she said that some of the authors are her personal friends. They have sold as many as 40 million books, their IPs have won awards and they’ve been on the New York Times Bestseller lists. Some fans have been obsessed with these IPs for decades and consider them to be core to their own personalities.

“The people who love these books are mega fans and would jump at the opportunity to play any of these stories,” Pereira said. “So we’ve built an engine that can take these books and turn it into a game experience, and then we create this wonderful virtuous cycle where these book lovers go into our game, and then we use that to drive a bigger audiences, which turns back and drives more book sales to properties that we know resonate but might have been sitting on a shelf collecting dust for the last 20 years because they’ve been lost to time.”

Reigniting forgotten worlds

Infinite Realms is combining fantasy book sales and AI and UGC.

The company knows that those communities and the fandom still exists and it’s possible to reignite this in a new generation using games. Using AI, the company can shorten the game development time and lower the costs by leveraging large language models (LLMs) that are custom tailored to each world.

Infinite Realms can take the author’s work and put it into a custom LLM that is partially owned by the author and by the company. That LLM can be licensed out not only to other game studios but to players who want to make their own custom experiences.

It’s also interesting to test how small and efficient an LLM can be and still have intelligence. The LLM has a bunch of lore in it, but it also needs to have a base level of intelligence, or enough data to create a subconsious awareness so to speak so that it knows how to have a conversation about the lore. The LLM can have conversations with the fans, and the fans can feed more data and create more lore for the LLM.

“The possibilities are endless, and the same workflow and partnerships that we developed with Unleashed Games for creating worlds pretty much on the fly can allow us to build games super fast, in as little as six months, because we already have the gameplay sorted out,” Pereira said.

She said that in the past, people would buy books and maybe those books would be adapted into movies and television. Game of Thrones and Wheel of Time are some great examples.

“But with Infinite Realms, we’re building AI powered worlds that you can step inside and interact with some of these characters that you fell in love with when you were 15 years old,” Pereira said. “And by doing that, we create what we’re calling the Netflix of living worlds.”

I noted that the Wheel of Time’s owners have put all 14 books in the series into an LLM that they can make available for user-generated content and players. It can have encyclopedic answers for the fans questions, but it can also serve as the canon police for anyone creating a new experience with the lore.

Things that players create with the tools can be as simple as an ambient video or screensaver on a TV. Or it could be used to create a full game — the full range of potential experiences.

“We can see how this scales, as there are so many other IPs, and you can see us becoming a digital bookshelf,” she said. “You could go from one world to the other on the fly, and we open that up to players to be able to collect these books. So we, in turn, become a digital publisher, where we take these properties that have had them in print, and we’re essentially using them as the start of our transmedia strategy, and then turning them into playable experiences.”

Being respectful of IP

Infinite Realms wants to create AI LLMs around fantasy lore.

All of it will be done with the authors’ approval, and the LLMs themselves can govern what the players can or can’t do. Of course, J.R.R. Tolkien’s The Lord of the Rings is the biggest fantasy franchise, but there are others like Terry Brooks’ The Sword of Shannara, which has reached 40 million fans, down to smaller ones that have sold a few million. The latter are easier and less expensive to work with.

“We essentially become a digital publisher,” Pereira said. “We can deepen our relationships and use the data” to make better decisions on marketing and choosing new IPs.

She added, “This is a great cycle to where we could use our platform to help revive the book publishing industry.”

Pereira is raising a funding round and hopes to be able to accomplish that by getting traction with some of the fantasy authors.

Unleashed Games will likely seek its own money for Haven and Infinite Realms will grow its own business. The companies can use the same technology but still be positioned separately. Infinite Realms has 18 people and it has a partner among AI developers that is also helping.

To judge the market, Infinite Realms is creating ways to test the market for IPs by doing tests with fans.

“I’ve worked with IP holders, and that’s like the No. 1 thing that I’ve been hearing from a lot of IP holders is that they’re trying to find game studios to develop games for their IPs, but they’re unwilling to provide funding for it,” Pereira said.

At the same time, Pereira said, “We’re trying to find a way to re-architect how we think about AI so that it’s respectful of copyright and is constructed with the intention of protecting people’s work.”

Read More »

The Download: gene de-extinction, and Ukraine’s Starlink connection

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology. The short, strange history of gene de-extinction This week saw the release of some fascinating news about some very furry rodents—so-called “woolly mice”—created as part of an experiment to explore how we might one day resurrect the woolly mammoth. The idea of bringing back extinct species has gained traction thanks to advances in sequencing of ancient DNA. This ancient genetic data is deepening our understanding of the past—for instance, by shedding light on interactions among prehistoric humans. But researchers are becoming more ambitious. Rather than just reading ancient DNA, they want to use it—by inserting it into living organisms.
Because this idea is so new and attracting so much attention, I decided it would be useful to create a record of previous attempts to add extinct DNA to living organisms. And since the technology doesn’t have a name, let’s give it one: “chronogenics.” Read the full story. —Antonio Regalado
This article first appeared in The Checkup, MIT Technology Review’s weekly biotech newsletter. To receive it in your inbox every Thursday, and read articles like this first, sign up here.  If you’re interested in de-extinction, why not check out: + How much would you pay to see a woolly mammoth? We spoke to Sara Ord, director of species restoration at Colossal, the world’s first “de-extinction” company, about its big ambitions.+ Colossal is also a de-extinction company, which is trying to resurrect the dodo. Read the full story.+ DNA that was frozen for 2 million years has been sequenced. The ancient DNA fragments come from a Greenland ecosystem where mastodons roamed among flowering plants. It may hold clues to how to survive a warming climate. The must-reads I’ve combed the internet to find you today’s most fun/important/scary/fascinating stories about technology. 1 Ukraine is worried the US could sever its vital Starlink connectionIts satellite internet is vital to Ukraine’s drone operations. (WP $)+ Thankfully, there are alternative providers. (Wired $)+ Ukraine is due to start a fresh round of war-ending negotiations next week. (FT $)+ Meet the radio-obsessed civilian shaping Ukraine’s drone defense. (MIT Technology Review) 2 Israel’s military has trained a powerful AI model on intercepted Palestinian dataThe ChatGPT-like tool can answer queries about the people it’s monitoring. (The Guardian)

3 Donald Trump has suspended tariffs on Canada and MexicoUntil April 2, at least. (Reuters)+ It’s the second time Trump has rolled back import taxes in as many days. (BBC)+ How Trump’s tariffs could drive up the cost of batteries, EVs, and more. (MIT Technology Review) 4 Can someone check on NASA’s Athena lunar lander?While we know it reached the moon, it appears to have toppled over. (NYT $)+ If it remains in an incorrect position, it may be unable to complete its mission. (CNN)+ Its engineers aren’t sure exactly where it is on the moon, either. (NBC News) 5 Shutting down 2G is easier said than doneMillions of vulnerable people around the world still rely on it to communicate. (Rest of World) 6 The hunt for the world’s oldest functional computer codeSpoiler: it may no longer be on Earth. (New Scientist $) 7 Robots are set to compete with humans in a Beijing half marathon🦿My money’s on the flesh and blood competitors. (Insider $)+ Researchers taught robots to run. Now they’re teaching them to walk. (MIT Technology Review) 8 Where did it all go wrong for Skype?It was the world leading video-calling app—until it wasn’t. (The Verge)  9 Dating is out, matchmaking is inWhy swipe when a platform can do the hard work for you? (Wired $)+ Forget dating apps: Here’s how the net’s newest matchmakers help you find love. (MIT Technology Review) 10 Apps are back, baby! 📱It’s like the original smartphone app boom all over again. (Bloomberg $)
Quote of the day
“You can only get so much juice out of every lemon.” —Carl-Benedikt Frey, a professor of AI and work at Oxford University’s Internet Institute, explains why pushing AI as a means of merely increasing productivity won’t always work, the Financial Times reports. The big story The cost of building the perfect wave June 2024
For nearly as long as surfing has existed, surfers have been obsessed with the search for the perfect wave. While this hunt has taken surfers from tropical coastlines to icebergs, these days that search may take place closer to home. That is, at least, the vision presented by developers and boosters in the growing industry of surf pools, spurred by advances in wave-­generating technology that have finally created artificial waves surfers actually want to ride. But there’s a problem: some of these pools are in drought-ridden areas, and face fierce local opposition. At the core of these fights is a question that’s also at the heart of the sport: What is the cost of finding, or now creating, the perfect wave—and who will have to bear it? Read the full story. —Eileen Guo

Read More »

Norway Opens Application for One CO2 Storage Exploration Area

Norway’s Energy Ministry has designated another area of the North Sea for application for licenses to explore the potential of carbon dioxide (CO2) storage. The acreage comprises defined blocks on the Norwegian side of the sea, upstream regulator the Norwegian Offshore Directorate said in an online statement. This is the eighth time acreage is being offered for CO2 storage exploration or exploitation on the Norwegian continental shelf, it noted. The application window for the latest acreage offer closes April 23. “In line with the regulations on transportation and storage of CO2 into subsea reservoirs on the continental shelf, the ministry normally expects to award an exploration license prior to awarding an exploitation license in a relevant area”, the Energy Ministry said separately. Norway has so far awarded 13 CO2 storage licenses: 12 for exploration and one for exploitation. Energy Minister Terje Aasland commented, “The purpose of allocating land is to be able to offer stakeholders in Europe large-scale CO2 storage on commercial terms”. Licensing for CO2 storage is part of Norwegian regulations passed December 2014 to support CO2 storage to mitigate climate change.  “Norway has great potential for storage on the continental shelf”, the ministry added. The Norwegian continental shelf holds a theoretical CO2 storage capacity of 80 billion metric tons, representing about 1,600 years of Norwegian CO2 emissions at current levels, according to a statement by the ministry April 30, 2024. In the latest awards two consortiums with Norway’s majority state-owned Equinor ASA won two exploration licenses in the North Sea. Equinor and London-based Harbour Energy PLC together won a permit straddling blocks 15/8, 15/9, 15/11 and 15/12. The permit, EXL012, lasts four years with three phases. Harbour Energy Norge AS holds a 60 percent stake as operator while Equinor Low Carbon Solution AS has 40 percent, according to a work

Read More »

MP for Truro and Falmouth calls for Cornwall offshore wind strategy

A Labour politician in Cornwall has called for the region to ramp up its domestic offshore wind supply chain. Jayne Kirkham, member of parliament for Truro and Falmouth, said: “At a recent Celtic Sea Power event, I saw just how many brilliant companies are doing amazing things here.” She made the comments months after The Crown Estate entered the second stage of leasing acreage in the Celtic Seas last autumn. “Cornwall has a long history of industrial innovation,” Kirkham said while meeting with marine construction firm MintMech in Penryn. “We’ve got the heritage and the expertise, now we need a strategy that ensures Cornwall maximises the benefits of offshore wind.” The Crown Estate entered the latest phase in its fifth offshore wind leasing round to establish floating offshore wind farms in the Celtic Sea, off the south-west of England and South Wales coast, in August. The second phase of the leasing round was launched, in which bidders must lay out plans to deliver new wind farms and explain how they will benefit local communities. The round has the potential to source up to 4.5GW of new wind capacity and spur investment in the local supply chain. Kirkham expressed hope that Cornish companies will soon be busy on UK projects. She said there are ongoing conversations with the National Energy System Operator (NESO) about ensuring potential wind energy hubs are well connected to the grid. The minister also referenced The Crown Estate’s £50 million Supply Chain Development Fund, which was launched to ensure the UK is prepared to meet offshore wind demands. The first £5m from the fund was awarded in 2024. Kirkham met with directors of Penryn-based marine construction firm MintMech in Jubilee Wharf to discuss the role Cornwall can play in the expansion of the UK’s offshore wind industry.

Read More »

Payroll in USA Oil and Gas Totals $168 Billion in 2024

Payroll in the U.S. oil and gas industry totaled $168 billion in 2024. That’s what the Texas Independent Producers & Royalty Owners Association (TIPRO) said in its latest State of Energy report, which was released this week, highlighting that this figure was “an increase of nearly $5 billion compared to the previous year”. Texas had the highest oil and gas payroll in the country in 2024, according to the report, which pointed out that this figure stood at $62 billion. The report outlined that California was “a distant second” with an oil and gas payroll figure of $15 billion, and that Louisiana was third, with an oil and gas payroll figure of $10 billion. Gasoline Stations with Convenience Stores had the highest U.S. oil and gas payroll by industry figure last year, at $26.8 billion, the report showed. Support Activities for Oil and Gas Operations had the second highest U.S. oil and gas payroll by industry figure in 2024, at $23.9 billion, and Crude Petroleum Extraction had the third highest, at $19.1 billion, the report outlined. The number of U.S. oil and gas businesses totaled 165,110, subject to revisions, TIPRO’s latest report stated. It highlighted that direct oil and natural gas Gross Regional Product exceeded $1 trillion last year and said the U.S. oil and natural gas industry purchased goods and services from over 900 different U.S. industry sectors in the amount of $865 billion in 2024. According to the report, Texas had the highest number of oil and gas businesses in the nation last year, with 23,549. This was followed by California, with 9,486 oil and gas businesses, Florida, with 7,695 oil and gas businesses, Georgia, with 6,453 oil and gas businesses, and New York, with 5,768 oil and gas businesses, the report outlined. The report noted that, in

Read More »

Stay Ahead with the Paperboy Newsletter

Your weekly dose of insights into AI, Bitcoin mining, Datacenters and Energy indusrty news. Spend 3-5 minutes and catch-up on 1 week of news.

Smarter with ONMINE

Streamline Your Growth with ONMINE