Stay Ahead, Stay ONMINE

Like humans, AI is forcing institutions to rethink their purpose

Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more Cognitive migration is not just an individual journey; it is also a collective and institutional one. As AI reshapes the terrain of thought, judgment and coordination, the very foundations of our schools, […]

Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more


Cognitive migration is not just an individual journey; it is also a collective and institutional one. As AI reshapes the terrain of thought, judgment and coordination, the very foundations of our schools, governments, corporations and civic systems are being called into question.

Institutions, like people, now face the challenge of rapid change: “Rethinking” their purpose, adapting their structures and rediscovering what makes them essential in a world where machines can increasingly think, decide and produce. Like people who are undergoing cognitive migration, institutions — and the people who run them — must reassess what they were made for.

Discontinuity

Institutions are designed to promote continuity. Their purpose is to endure, to offer structure, legitimacy and coherence across time. It is those very attributes that contribute to trust. We rely on institutions not only to deliver services and enforce norms, but to provide a sense of order in a complex world. They are the long-arc vessels of civilization, meant to hold steady as individuals come and go. Without viable institutions, society risks upheaval and an increasingly uncertain future.

But today, many of our core institutions are reeling. Having long served as the scaffolding of modern life, they are being tested in ways that feel not only sudden, but systemic.

Some of this pressure comes from AI, which is rapidly reshaping the cognitive terrain on which these institutions were built. But AI is not the only force. The past two decades have brought rising public distrust, partisan fragmentation and challenges to institutional legitimacy that predate the generative AI technological wave. From increasing income inequality, to attacks on scientific process and consensus, to politicized courts, to declining university enrollments, the erosion of trust in our institutions has multiple causes, as well as compounding effects.

In this context, the arrival of increasingly capable AI systems is not merely another challenge. It is an accelerant, fuel to the fire of institutional disruption. This disruption demands that institutions adapt their operations and revisit foundational assumptions. What are institutions for in a world where credentialing, reasoning and coordination are no longer exclusively human domains? All this institutional reinvention needs to take place at a pace that defies their very purpose and nature.

This is the institutional dimension of cognitive migration: A shift not just in individuals find meaning and value, but in how our collective societal structures must evolve to support a new era. And as with all migrations, the journey will be uneven, contested and deeply consequential.

The architecture of the old regime

The institutions in place now were not designed for this moment. Most were forged in the Industrial Age and refined during the Digital Revolution. Their operating models reflect the logic of earlier cognitive regimes: stable processes, centralized expertise and the tacit assumption that human intelligence would remain preeminent.

Schools, corporations, courts and government agencies are structured to manage people and information on a large scale. They rely on predictability, expert credentials and well-defined hierarchies of decision-making. These are traditional strengths that — even when considered bureaucratic — have historically offered a foundation for trust, consistency and broad participation within complex societies.

But the assumptions beneath these structures are under strain. AI systems now perform tasks once reserved for knowledge workers, including summarizing documents, analyzing data, writing legal briefs, performing research, creating lesson plans and teaching, coding applications and building and executing marketing campaigns. Beyond automation, a deeper disruption is underway: The people running these institutions are expected to defend their continued relevance in a world where knowledge itself is no longer as highly valued or even a uniquely human asset.

The relevance of some institutions is called into question from outside challengers including tech platforms, alternative credentialing models and decentralized networks. This essentially means that the traditional gatekeepers of trust, expertise and coordination are being challenged by faster, flatter and often more digitally native alternatives. In some cases, even long-standing institutional functions such as adjudicating disputes are being questioned, ignored, or bypassed altogether.

This does not mean institutional collapse is inevitable. But it does suggest that the current paradigm of stable, slow-moving and authority-based structures may not endure. At a minimum, institutions are under intense pressure to change. If institutions are to remain relevant and play a vital role in the age of AI, they must become more adaptive, transparent and attuned to the values that cannot readily be encoded in algorithms: human dignity, ethical deliberation and long-term stewardship.

The choice ahead is not whether institutions will change, but how. Will they resist, ossify and fall into irrelevance? Will they be forcibly restructured to meet transient agendas? Or will they deliberately reimagine themselves as co-evolving partners in a world of shared intelligence and shifting value?

First steps of institutional migration

A growing number of institutions are beginning to adapt. These responses are varied and often tentative, signs of motion more than full transformation. These are green shoots; taken together, they suggest that the cognitive migration of institutions may already be underway.

Yet there is a deeper challenge beneath these experiments: Many institutions are still bound by outdated methods of operating. The environment, however, has changed. AI and other factors are redrawing the landscape, and institutions are only beginning to recalibrate.

One example of change comes from an Arizona-based charter school where AI plays a leading role in daily instruction. Branded as Unbound Academy, the school uses AI platforms to deliver core academic content in condensed, focused sessions tailored for each child. This shows promise to improve academic achievement while also allowing students time later in the day to work on life skills, project-based learning and interpersonal development. In this model, teachers are reframed as guides and mentors, not content deliverers. It is an early glimpse of what institutional migration might look like in education: Not just digitizing the old classroom, but redesigning its structure, human roles and priorities around what AI can do.

The World Bank reported on a pilot program in Nigeria that used AI to support learning through an after-school program. The results revealed “overwhelmingly positive effects on learning outcomes,” with AI serving as a virtual tutor and teachers providing support. Testing showed students achieved “nearly two years of typical learning in just six weeks.”

Similar signals are emerging elsewhere. In government, a growing number of public agencies are experimenting with AI systems to improve responsiveness: triaging constituent inquiries, drafting preliminary communications or analyzing public sentiment. Leading AI labs such as OpenAI are now tailoring their tools for government use. These nascent efforts offer a glimpse into how institutions might reallocate human effort and attention toward interpretation, discretion and trust-building; functions that remain profoundly human.

While most of these initiatives are framed in terms of productivity, they raise deeper questions about the evolving role of the human within decision-making structures. In other words, what is the future of human work? The conventional wisdom viewpoint voiced by futurist Melanie Subin in a CBS interview is that “AI is going to change jobs, replace tasks and change the nature of work. But as with the Industrial Revolution and many other technological advancements we have seen over the past 100 years, there will still be a role for people; that role may just change.”

That seeming evolution stands in stark contrast to the poignant prediction from Dario Amodei, CEO of Anthropic, one of the world’s most powerful creators of AI technologies. In his view, AI could eliminate half of all entry-level white-collar jobs and spike unemployment to 10 to 20% in the next 1 to 5 years. “We, as the producers of this technology, have a duty and an obligation to be honest about what is coming,” he said in an interview with Axios. His draconian prediction could happen, although perhaps not as quickly as he suggests, as diffusion of new technology across society can often take longer than is expected.

Nevertheless, the potential for AI to displace workers has long been known. As early as 2019, Kevin Roose wrote about conversations he had with corporate executives at a World Economic Forum meeting. “They’ll never admit it in public,” he wrote, “but many of your bosses want machines to replace you as soon as possible.”

In 2025, Roose reported that there are signs this is beginning to occur. “In interview after interview, I’m hearing that firms are making rapid progress toward automating entry-level work, and that AI companies are racing to build ‘virtual workers’ that can replace junior employees at a fraction of the cost.”

Across all institutional domains, there are green shoots of transformation. But the throughline remains fragmented, merely early signals of change and not yet blueprints. The deeper challenge is to move from experimentation to structural reinvention. In the interim, there could be a lot of collateral damage, not only to those who lose their jobs but also to the overall effectiveness of institutions amidst turmoil.

How can institutions move from experimentation to integration, from reactive adoption to principled design? And can this be done at a pace that adequately reflects the rate of change? Recognizing the need is only the beginning. The real challenge is designing for it.

Institutional design principles for the next era

If AI acceleration continues, this will lead to immense pressure on institutions to respond. If institutions can move at pace, the question becomes: How can they move from reactive adoption to principled design? They need not just innovation, but informed vision and principled intention. Institutions must be reimagined from the ground up, built not just for efficiency or scale, but for adaptability, trust and long-term societal coherence.

This requires design principles that are neither technocratic nor nostalgic, but grounded in the realities of the migration underway, based on shared intelligence, human vulnerability and with a goal of creating a more humane society. That in mind, here are three practical design principles.

Build for responsiveness, not longevity

Institutions must be designed to move beyond fixed hierarchies and slow feedback loops. In a world reshaped by real-time information and AI-augmented decision-making, responsiveness and adaptability become core competencies. This means flattening decision layers where possible, empowering frontline actors with tools and trust and investing in data systems that surface insights quickly, without outsourcing judgment to algorithms alone. Responsiveness is not just about speed. It is about sensing change early and acting with moral clarity.

Integrate AI where it frees humans to focus on the human

AI should be deployed not as a replacement strategy, but as a refocusing tool. The most forward-looking institutions will utilize AI to absorb repetitive tasks and administrative burdens, thus freeing human capacity for interpretation, trust-building, care, creativity and strategic thinking. In education, this might mean AI-created and presented lessons that allow teachers to spend more time with struggling students. In government, it could mean greater automated processing that gives human staff more time to solve complex cases with empathy and discretion. The goal should not be to fully automate institutions. It is instead to humanize them. This principle encourages using AI as a support beam, not a substitute.

Keep humans in the loop where it matters most

Institutions that endure will be those that make room for human judgment at critical points of interpretation, escalation and ethics. This means designing systems where human-in-the-loop is not a checkbox, but a structural feature that is clearly defined, legally protected and socially valued. Whether in justice systems, healthcare or public service, the presence of a human voice and moral perspective must remain central where stakes are high, and values are contested. AI can inform, but humans must still decide.

These principles are not meant to be static rules, but directional choices. They are starting points for reimagining how institutions can remain human-centered in a machine-enhanced world. They reflect a commitment to modernization without moral abandonment, to speed without shallowness or callousness and to intelligence shared between humans and machines.

Beyond adaptation: Institutions and question of purpose

In times of disruption, individuals often ask: ‘What was I made for?’ We must ask the same of our institutions. As AI upends our cognitive terrain and accelerates the pace of change, the relevance of our core institutions is no longer guaranteed by tradition, function or status. They, too, are subject to the forces of cognitive migration. Like individuals, their future must include decisions about whether to resist, retreat or transform.

As generative AI systems take on tasks of reasoning, research, writing and coordination, the foundational assumptions of institutional authority including expertise, hierarchy and predictability begin to fracture. But what follows cannot be a hollowing out, because the fundamental purpose of institutions is too essential to abandon. It must be a re-founding.

Our institutions should not be replaced by machines. They should instead become more human: More responsive to complexity, anchored in ethical deliberation, capable of holding long-term visions in a short-term world. Institutions that do not adapt with intention may not survive the turbulence ahead. The dynamism of the 21st century will not wait.

This is the institutional dimension of cognitive migration: A reckoning with identity, value and function in a world where intelligence is no longer our exclusive domain. The institutions that endure will be those that migrate not just in form, but in soul, crossing into new terrain with tools that serve humanity.

For those shaping schools, companies or civic structures, the path forward lies not in resisting AI, but in redefining what only humans and human institutions can truly offer.

Gary Grossman is EVP of technology practice at Edelman and global lead of the Edelman AI Center of Excellence. 

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Nvidia aims to bring AI to wireless

Key features of ARC-Compact include: Energy Efficiency: Utilizing the L4 GPU (72-watt power footprint) and an energy-efficient ARM CPU, ARC-Compact aims for a total system power comparable to custom baseband unit (BBU) solutions currently in use. 5G vRAN support: It fully supports 5G TDD, FDD, massive MIMO, and all O-RAN

Read More »

Netgear’s enterprise ambitions grow with SASE acquisition

Addressing the SME security gap The acquisition directly addresses a portfolio gap that Netgear (Nasdaq:NTGR) has identified through customer feedback.  According to Badjate, customers have been saying that they like the Netgear products, but they also really need more security capabilities. Netgear’s target market focuses on organizations with fewer than

Read More »

IBM’s cloud crisis deepens: 54 services disrupted in latest outage

Rawat said IBM’s incident response appears slow and ineffective, hinting at procedural or resource limitations. The situation also raises concerns about IBM Cloud’s adherence to zero trust principles, its automation in threat response, and the overall enforcement of security controls. “The recent IBM Cloud outages are part of a broader

Read More »

ScottishPower Allots About $300MM for UK Power Grid Modernization

ScottishPower Energy Networks (SPEN), Iberdrola’s distribution company in the United Kingdom, will invest more than EUR 262 million ($298.8 million) in the modernization of the United Kingdom’s electricity grid. SPEN said in a media release that six partners will continue working on the maintenance and upgrade of more than 20,000 kilometers (12,400 miles) of overhead lines across the network over the next four years. SPEN partners include Scottland-based Aureos, Gaeltec, and PLPC, which will support the six license districts in central and southern Scotland (Ayrshire and Clyde South, Central and Fife, Dumfries and Galloway, Edinburgh and Borders, Glasgow and Clyde North, Lanarkshire). The company said it is also partnering with Emerald Power, IES, and Network Plus, which will support the license districts in Mid-Cheshire, Merseyside, Dee Valley and Mid Wales, Wirral and North Wales. “Ensuring we have the partners, resources, and technical skills in place to deliver on our bold and ambitious plans for our network is vital for the modern and resilient grid needed to support the doubling of demand”, Nicola Connelly, SPEN CEO, said. “These contracts not only support significant investment in our overhead line network, they allow us to build on the solid foundations created with our supply chain partners and give certainty and confidence to further invest in their skills and people.  It’s a win-win on both sides and we look forward to working together to make a long and lasting difference for all our communities – from Anstruther to Anglesey”. The contracts will support over 500 jobs – including more than 50 new linesmen roles – nationwide, with companies based in and around ScottishPower’s Scotland and Manweb license areas. “This is an extremely significant milestone for Emerald Power and provides the opportunity to further invest in our business – recruiting, training, and upskilling the resources needed to deliver

Read More »

Fennex to Deploy AI-Powered Safety System across EnQuest’s UK Operations

Fennex Ltd. has bagged a multi-year deal from EnQuest plc to deploy the flagship AI-powered Behaviour-Based Safety System (BBSS) across EnQuest’s UK operations.   EnQuest oversees a varied portfolio of offshore assets in the North Sea, which includes Thistle, Heather, Magnus, and the Kraken FPSO, along with the Sullom Voe Terminal located onshore Shetland, recognized as one of the largest oil terminals in Europe, Fennex noted in a media release. Fennex added that the BBSS is already live across all of EnQuest’s UK offshore assets and the Sullom Voe Terminal. This rapid deployment was achieved in just eight weeks. “BBSS is now deployed across all EnQuest’s UK-operated offshore assets, and for the first time at a major onshore terminal”, Adrian Brown, Managing Director at Fennex, said. “EnQuest was eager to roll out the platform quickly, and thanks to strong collaboration, we were able to go live both offshore and onshore in record time”. “We identified BBSS as an opportunity to make a step change in operational safety through making it easier and more user-friendly for personnel to participate and allowing us to make more effective use of the resulting leading data. It provides full visibility of engagement in our safety reporting and real-time data, giving us immediate insight into reported issues and the ability to act swiftly for the best outcomes in our operations”, EnQuest’s Director of HSE and Wells, Ian McKimmie, added. As the collaboration progresses, Fennex and EnQuest are working together to reveal even more value – leveraging advanced analytics, behavioral insights, and AI-driven predictive safety tools to foster a culture of proactive, intelligence-led safety, Fennex said. To contact the author, email [email protected] WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to

Read More »

Canada’s Oil Sands Emissions Intensity Falls for Sixth Year

Canada’s oil sands industry reduced its emissions per barrel for the sixth straight year in 2023, even as one growing portion of the sector moved in the opposite direction, according to new Alberta government data released Thursday. The emissions intensity of all oil sands sites fell to the equivalent of 0.399 metric tons of carbon dioxide per cubic meter of bitumen produced, down from 0.404 in 2022, the data show. The gain reflects improvements at oil sands mines, where bitumen is dug from the ground. However, in situ oil sands, which use wells similar to traditional oil producers, saw emissions per barrel rise.  Even with the efficiency improvement, total emissions rose to the equivalent 80.1 million metric tons of carbon dioxide, up from 78.8 million in 2022, the data show. That’s the highest in data back to 2011. The oil sands’ declining energy intensity — to the lowest in data stretching back to 2011 — is welcome news for an industry that has struggled with a reputation for being climate unfriendly, prompting some investors to shun it altogether. However, the rising emissions intensity at well sites presents a challenge for the sector, as the method’s lower costs make it increasingly popular among producers. While the average intensity of oil sands producers is higher than the average for the global oil industry overall, drillers ‘emissions profiles vary widely around the world, said Kevin Birn, chief analyst for Canadian oil markets for S&P Global. The oil sands “fits well within the range of carbon intensity of oil and gas we see in the world,” Birn said in an interview. All of the oil sands mines reduced their emissions intensity with Canadian Natural Resources Ltd.’s Horizon making the biggest gain for the year.  In situ production facilities, which include the more than 250,000 barrel-a-day Suncor

Read More »

ICYMI: Energy Secretary Wright, EPA Administrator Zeldin, and Interior Secretary Burgum Visit Alaska to Advance Unleashing American Energy

In Case You Missed It— U.S. Secretary of Energy Chris Wright traveled to Alaska earlier this week, where he joined U.S. Department of the Interior Secretary Doug Burgum and U.S. Environmental Protection Agency (EPA) Administrator Lee Zeldin to advance President Trump’s bold agenda to unleash Alaska’s extraordinary resource potential. In Anchorage and on the North Slope, Secretary Wright met with state, local and tribal leaders as well as Alaskan energy workers, held bilateral meetings with foreign energy ministers to advance American energy projects like Alaska LNG, and participated in a keynote fireside chat at the Alaska Sustainable Energy Conference. “I was proud to join my colleagues, Doug Burgum and Lee Zeldin, in Alaska to help fulfill President Trump’s mission of unleashing Alaska to create a safe, prosperous, and opportunity-filled future,” said Secretary Wright. “Alaska has always been a big bold place, a big bold idea, full of big bold people. Unfortunately, the last administration especially and many before them sought to shut down Alaska in the name of climate nonsense. Those days are over. The Trump Administration is fully committed to restoring the rights and liberties of Alaskans, because when Alaska is unleashed, the entire country and world benefits.” Click here to watch a highlight video of the trip and here to download and view photos from the visit. Roundtable in Anchorage On Sunday, Secretaries Wright and Burgum and Administrator Zeldin joined Alaska Governor Mike Dunleavy and Senator Dan Sullivan (R-Alaska) for a roundtable discussion in Anchorage with leaders from across Alaska’s resource development industry. WATCH: Alaska Resources Roundtable with Secretary Burgum, Secretary Wright and Administrator Zeldin Town hall in Utqiagvik (Barrow), Alaska The delegation then traveled to the capital of Alaska’s North Slope Borough, Utqiagvik, Alaska, for a town hall with the predominantly Alaska-Native community. The cabinet members heard

Read More »

Kenya Says Tullow Asset Sale ‘Raises Concern’ for Oil Sector

Kenya lawmakers said a lack of detail over Tullow Oil Plc’s sale of oil assets to local company Gulf Energy “raises concerns” for the industry in the East African nation. Tullow ended a decade-long quest to develop discoveries in April, when it sold the inland fields for $120 million to Gulf Energy, a Nairobi-based oil and gas trader. Tullow has announced divestments in recent months toward reducing its debt below $1 billion this year. The deal for the assets has included few details, “which raises concerns about the continuity, governance, and strategic direction of the country’s nascent oil industry,” the national assembly energy committee said in a report tabled on Wednesday.  “There is limited information regarding the terms of the exit, the implications of the transaction on Kenya’s commercial oil prospects, and its potential impact on the completion and timely approval of the Field Development Plan,” it said. Gulf Energy is an oil supplier, electricity generator and infrastructure developer, according to its website. It didn’t immediately respond to emails seeking comment. The sale of working interests in Kenya to Gulf Energy “continues to progress well,” a Tullow spokesperson said in a reply to questions. “We have a high level of confidence in it completing with the receipt of the first two payments totaling $80 million expected during 2025.” Passage of the development plan, which is still pending, became one of the obstacles Tullow faced with the project. Partners TotalEnergies SE and Africa Oil Corp. exited the joint venture in 2023 and the company failed to find a strategic partner to replace them.      The payment profile of the Gulf Energy deal “isn’t too demanding,” though it’s unclear who is backing the company, said Ashley Kelty, an analyst at Panmure Liberum. WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect

Read More »

Oil Rises as Solid USA Jobs Data Pushes Algos to Drop Short Bets

Oil rose as stronger-than-expected US jobs data eased concerns about an economic slowdown that would crimp demand, spurring algorithmic traders to reduce short positions.  West Texas Intermediate climbed almost 2% to settle above $64 a barrel, notching the largest weekly gain since November. Crude followed equities higher after US job growth in May narrowly surpassed economist forecasts, allaying concerns of near-term demand deterioration. The figures also pushed economy-sensitive diesel futures to a two-week high.  “Trading is relatively quiet today, with macroeconomic factors continuing to drive the narrative,” said Rebecca Babin, a senior energy trader at CIBC Private Wealth Group. “The unemployment data is easing concerns that demand will sharply decline due to tariff uncertainty.” The positive economic data spurred commodity trading advisers to ease off of their bearish tilt. The funds, which can accelerate price momentum, liquidated short positions to sit at negative 9% short in WTI on Friday, compared with 64% short on June 5, according to data from Bridgeton Research Group. The rally was supported by enduring risk-on sentiment from optimistic signs on trade talks between the US and China, the world’s largest importer of crude. President Donald Trump and his Chinese counterpart, Xi Jinping, agreed to further negotiations over tariffs and supplies of rare earth minerals. The positive signals come against the backdrop of an oil market that has been increasingly rangebound in recent weeks. Prices have traded in a $5 band since the middle of May, and a gauge of volatility for US crude futures is at the lowest since early April.  Oil has been buffeted in Trump’s second term as trade tensions between the world’s two largest economies menace demand. At the same time, the OPEC+ alliance has been adding barrels back to the market at a faster-than-expected rate, further clouding an already weak outlook for the second half

Read More »

LiquidStack launches cooling system for high density, high-powered data centers

The CDU is serviceable from the front of the unit, with no rear or end access required, allowing the system to be placed against the wall. The skid-mounted system can come with rail and overhead piping pre-installed or shipped as separate cabinets for on-site assembly. The single-phase system has high-efficiency dual pumps designed to protect critical components from leaks and a centralized design with separate pump and control modules reduce both the number of components and complexity. “AI will keep pushing thermal output to new extremes, and data centers need cooling systems that can be easily deployed, managed, and scaled to match heat rejection demands as they rise,” said Joe Capes, CEO of LiquidStack in a statement. “With up to 10MW of cooling capacity at N, N+1, or N+2, the GigaModular is a platform like no other—we designed it to be the only CDU our customers will ever need. It future-proofs design selections for direct-to-chip liquid cooling without traditional limits or boundaries.”

Read More »

Enterprises face data center power design challenges

” Now, with AI, GPUs need data to do a lot of compute and send that back to another GPU. That connection needs to be close together, and that is what’s pushing the density, the chips are more powerful and so on, but the necessity of everything being close together is what’s driving this big revolution,” he said. That revolution in new architecture is new data center designs. Cordovil said that instead of putting the power shelves within the rack, system administrators are putting a sidecar next to those racks and loading the sidecar with the power system, which serves two to four racks. This allows for more compute per rack and lower latency since the data doesn’t have to travel as far. The problem is that 1 mW racks are uncharted territory and no one knows how to manage the power, which is considerable now. ”There’s no user manual that says, hey, just follow this and everything’s going to be all right. You really need to push the boundaries of understanding how to work. You need to start designing something somehow, so that is a challenge to data center designers,” he said. And this brings up another issue: many corporate data centers have power plugs that are like the ones that you have at home, more or less, so they didn’t need to have an advanced electrician certification. “We’re not playing with that power anymore. You need to be very aware of how to connect something. Some of the technicians are going to need to be certified electricians, which is a skills gap in the market that we see in most markets out there,” said Cordovil. A CompTIA A+ certification will teach you the basics of power, but not the advanced skills needed for these increasingly dense racks. Cordovil

Read More »

HPE Nonstop servers target data center, high-throughput applications

HPE has bumped up the size and speed of its fault-tolerant Nonstop Compute servers. There are two new servers – the 8TB, Intel Xeon-based Nonstop Compute NS9 X5 and Nonstop Compute NS5 X5 – aimed at enterprise customers looking to upgrade their transaction processing network infrastructure or support larger application workloads. Like other HPE Nonstop systems, the two new boxes include compute, software, storage, networking and database resources as well as full-system clustering and HPE’s specialized Nonstop operating system. The flagship NS9 X5 features support for dual-fabric HDR200 InfiniBand interconnect, which effectively doubles the interconnect bandwidth between it and other servers compared to the current NS8 X4, according to an HPE blog detailing the new servers. It supports up to 270 networking ports per NS9 X system, can be clustered with up to 16 other NS9 X5s, and can support 25 GbE network connectivity for modern data center integration and high-throughput applications, according to HPE.

Read More »

AI boom exposes infrastructure gaps: APAC’s data center demand to outstrip supply by 42%

“Investor confidence in data centres is expected to strengthen over the remainder of the decade,” the report said. “Strong demand and solid underlying fundamentals fuelled by AI and cloud services growth will provide a robust foundation for investors to build scale.” Enterprise strategies must evolve With supply constrained and prices rising, CBRE recommended that enterprises rethink data center procurement models. Waiting for optimal sites or price points is no longer viable in many markets. Instead, enterprises should pursue early partnerships with operators that have robust development pipelines and focus on securing power-ready land. Build-to-suit models are becoming more relevant, especially for larger capacity requirements. Smaller enterprise facilities — those under 5MW — may face sustainability challenges in the long term. The report suggested that these could become “less relevant” as companies increasingly turn to specialized colocation and hyperscale providers. Still, traditional workloads will continue to represent up to 50% of total demand through 2030, preserving value in existing facilities for non-AI use cases, the report added. The region’s projected 15 to 25 GW gap is more than a temporary shortage — it signals a structural shift, CBRE said. Enterprises that act early to secure infrastructure, invest in emerging markets, and align with power availability will be best positioned to meet digital transformation goals. “Those that wait may find themselves locked out of the digital infrastructure they need to compete,” the report added.

Read More »

Cisco bolsters DNS security package

The software can block domains associated with phishing, malware, botnets, and other high-risk categories such as cryptomining or new domains that haven’t been reported previously. It can also create custom block and allow lists and offers the ability to pinpoint compromised systems using real-time security activity reports, Brunetto wrote. According to Cisco, many organizations leave DNS resolution to their ISP. “But the growth of direct enterprise internet connections and remote work make DNS optimization for threat defense, privacy, compliance, and performance ever more important,” Cisco stated. “Along with core security hygiene, like a patching program, strong DNS-layer security is the leading cost-effective way to improve security posture. It blocks threats before they even reach your firewall, dramatically reducing the alert pressure your security team manages.” “Unlike other Secure Service Edge (SSE) solutions that have added basic DNS security in a ‘checkbox’ attempt to meet market demand, Cisco Secure Access – DNS Defense embeds strong security into its global network of 50+ DNS data centers,” Brunetto wrote. “Among all SSE solutions, only Cisco’s features a recursive DNS architecture that ensures low-latency, fast DNS resolution, and seamless failover.”

Read More »

HPE Aruba unveils raft of new switches for data center, campus modernization

And in large-scale enterprise environments embracing collapsed-core designs, the switch acts as a high-performance aggregation layer. It consolidates services, simplifies network architecture, and enforces security policies natively, reducing complexity and operational cost, Gray said. In addition, the switch offers the agility and security required at colocation facilities and edge sites. Its integrated Layer 4 stateful security and automation-ready platform enable rapid deployment while maintaining robust control and visibility over distributed infrastructure, Gray said. The CX 10040 significantly expands the capacity it can provide and the roles it can serve for enterprise customers, according to one industry analyst. “From the enterprise side, this expands on the feature set and capabilities of the original 10000, giving customers the ability to run additional services directly in the network,” said Alan Weckel, co-founder and analyst with The 650 Group. “It helps drive a lower TCO and provide a more secure network.”  Aimed as a VMware alternative Gray noted that HPE Aruba is combining its recently announced Morpheus VM Essentials plug-in package, which offers a hypervisor-based package aimed at hybrid cloud virtualization environments, with the CX 10040 to deliver a meaningful alternative to Broadcom’s VMware package. “If customers want to get out of the business of having to buy VM cloud or Cloud Foundation stuff and all of that, they can replace the distributed firewall, microsegmentation and lots of the capabilities found in the old VMware NSX [networking software] and the CX 10k, and Morpheus can easily replace that functionality [such as VM orchestration, automation and policy management],” Gray said. The 650 Group’s Weckel weighed in on the idea of the CX 10040 as a VMware alternative:

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »