Stay Ahead, Stay ONMINE

Nvidia launches Blackwell RTX Pro for workstations and servers

Nvidia today announced the Nvidia RTX Pro Blackwell series — a new generation of workstation and server graphics processing units. Nvidia said the GPUs redefine workflows for AI, technical, creative, engineering and design professionals with accelerated computing, AI inference, ray tracing and neural rendering technologies. The company unveiled the news during the keynote speech of Jensen Huang, CEO of Nvidia, at the GTC 2025 event. For everything from agentic AI, simulation, extended reality, 3D design and complex visual effects to developing physical AI powering autonomous robots, vehicles and smart spaces, the RTX Pro Blackwell series provides professionals across industries with compute power, memory capacity and data throughput right at their fingertips — from their desktop, on the go with mobile workstations or powered by data center GPUs. The new lineup includes: Data center GPU: Nvidia RTX PRO 6000 Blackwell Server Edition Desktop GPUs: Nvidia RTX PRO 6000 Blackwell Workstation Edition, Nvidia RTX PRO 6000 Blackwell Max-Q Workstation Edition, Nvidia RTX PRO 5000 Blackwell, Nvidia RTX PRO 4500 Blackwell and Nvidia RTX PRO 4000 Blackwell Laptop GPUs: Nvidia RTX PRO 5000 Blackwell, Nvidia RTX PRO 4000 Blackwell, Nvidia RTX PRO 3000 Blackwell, Nvidia RTX PRO 2000 Blackwell, Nvidia RTX PRO 1000 Blackwell and Nvidia RTX PRO 500 Blackwell “Software developers, data scientists, artists, designers and engineers need powerful AI and graphics performance to push the boundaries of visual computing and simulation, helping tackle incredible industry challenges,” said Bob Pette, vice president of enterprise platforms at Nvidia, in a statement. “Bringing Nvidia Blackwell to workstations and servers will take productivity, performance and speed to new heights, accelerating AI inference serving, data science, visualization and content creation.” Nvidia Blackwell technology comes to workstations and data centers Nvidia RTX PRO Blackwell Desktop GPUs. RTX Pro Blackwell GPUs unlock the potential of generative, agentic and physical AI by delivering exceptional performance, efficiency and scale. Nvidia RTX Pro Blackwell GPUs feature: Nvidia Streaming Multiprocessor: Offers up to 1.5 times faster throughput and new neural shaders that integrate AI inside of programmable shaders to drive the next decade of AI-augmented graphics innovations. Fourth-Generation RT Cores: Delivers up to two times the performance of the previous generation to create photoreal, physically accurate scenes and complex 3D designs with optimizations for Nvidia RTX Mega Geometry. Fifth-Generation Tensor Cores: Delivers up to 4,000 AI trillion operations per second and adds support for FP4 precision and Nvidia DLSS 4 Multi Frame Generation, enabling a new era of AI-powered graphics and the ability to run and prototype larger AI models faster. Larger, Faster GDDR7 Memory: Boosts bandwidth and capacity — up to 96GB for workstations and servers and up to 24GB on laptops. This enables applications to run faster and work with larger, more complex datasets for everything from tackling massive 3D and AI projects to exploring large-scale virtual reality environments. Ninth-Generation Nvidia NVENC: Accelerates video encoding speed and improves quality for professional video applications with added support for 4:2:2 encoding. Sixth-Generation Nvidia NVDEC: Provides up to double the H.264 decoding throughput and offers support for 4:2:2 H.264 and HEVC decode. Professionals can benefit from high-quality video playback, accelerate video data ingestion and use advanced AI-powered video editing features. Fifth-Generation PCIe: Support for fifth-generation PCI Express provides double the bandwidth over the previous generation, improving data transfer speeds from CPU memory and unlocking faster performance for data-intensive tasks. DisplayPort 2.1: Drives high-resolution displays at up to 4K at 480Hz and 8K at 165Hz. Increased bandwidth enables seamless multi-monitor setups, while high dynamic range and higher color depth support deliver more precise color accuracy for tasks like video editing, 3D design and live broadcasting. Multi-Instance GPU (MIG): The RTX PRO 6000 data center and desktop GPUs and 5000 series desktop GPUs feature MIG technology, enabling secure partitioning of a single GPU into up to four instances (6000 series) or two instances (5000 series). Nvidia said Fault isolation is designed to prevent workload interference for secure, efficient resource allocation for diverse workloads, maximizing performance and flexibility. The new laptop GPUs also support the latest Nvidia Blackwell Max-Q technologies, which intelligently and continually optimize laptop performance and power efficiency with AI. Nvidia RTX PRO Blackwell Laptop GPUs. With neural rendering and AI-augmented tools, Nvidia RTX PRO Blackwell GPUs enable the creation of stunning visuals, digital twins of real-world environments and immersive experiences with unprecedented speed and efficiency. The GPUs are built to elevate 3D computer-aided design and building information model workflows, offering designers and engineers exceptional performance for complex modeling, rendering and visualization. Designed for enterprise data center deployments, the RTX PRO 6000 Blackwell Server Edition features a passively cooled thermal design and can be configured with up to eight GPUs per server. For workloads that require the compute density and scale that data centers offer, the RTX PRO 6000 Blackwell Server Edition delivers powerful performance for next-generation AI, scientific and visual computing applications across industries such as healthcare, manufacturing, retail and media and entertainment. In addition, this powerful data center GPU can be combined with Nvidia vGPU software to power AI workloads across virtualized environments and deliver high-performance virtual workstation instances to remote users. Nvidia vGPU support for the Nvidia RTX PRO 6000 Blackwell Server Edition GPU is expected in the latter half of this year. “Foster + Partners has tested the Nvidia RTX PRO 6000 Blackwell Max-Q Workstation Edition GPU on Cyclops, our GPU-based ray-tracing product,” said Martha Tsigkari, head of applied research and development and senior partner at Foster + Partners, in a statement. “The new Nvidia Blackwell GPU has managed to outperform everything we have tested before. For example, when using it with Cyclops, it has performed at 5x the speed of Nvidia RTX A6000 GPUs. Rendering speeds also increased five times, allowing tools like Cyclops to provide feedback on how well our design solutions perform in real time as we design them and resulting in intuitive yet informed decision-making from early conceptual stages.” “Nvidia RTX PRO 6000 Blackwell Workstation Edition GPUs enable incredibly sharp and photorealistic graphics,” said Jeff Hammoud, chief design officer at Rivian. “In conjunction with a Varjo XR4 headset and Autodesk VRED, the system delivered the level of crispness necessary for immersive automotive design reviews. With Nvidia Blackwell support for PCIe Gen 5, we used two powerful 600W GPUs via VR SLI, allowing us to achieve the highest pixel density and the most stunning visuals we have ever experienced in VR.” RTX PRO GPUs run on the Nvidia AI platform and feature larger memory capacity and the latest Tensor Cores to accelerate a deep ecosystem of AI-accelerated applications built on Nvidia CUDA and RTX technology. With everything from the latest AI-based content creation tools and new reasoning models, such as the Nvidia Llama Nemotron Reason family of models and Nvidia NIM microservices unveiled today, inferencing is faster than ever. And with over 400 Nvidia CUDA-X libraries, developers can easily build, optimize, deploy and scale new AI applications, from workstations to the data center or cloud. Enterprises can fast-track their AI development and deployments by prototyping locally with an Nvidia RTX PRO GPU and the Nvidia Omniverse and Nvidia AI Enterprise platforms, Nvidia Blueprints and Nvidia NIM, which gives access to easy-to-use inference microservices backed by enterprise-level support. They can also run these applications at scale on the ultimate universal data center GPU for AI and visual computing, delivering breakthrough acceleration for the most demanding compute-intensive enterprise workloads with the RTX Pro 6000 Blackwell Server Edition. Availability The Nvidia RTX Pro 6000 Blackwell Server Edition will soon be available in server configurations from leading data center system partners including Cisco, Dell Technologies, Hewlett Packard Enterprise, Lenovo and Supermicro. Cloud service providers and GPU cloud providers including AWS, Google Cloud, Microsoft Azure and CoreWeave will be among the first to offer instances powered by the Nvidia RTX Pro 6000 Blackwell Server Edition later this year. In addition, the server edition GPU will be available in data center platforms from ASUS, GIGABYTE, Ingrasys, Quanta Cloud Technology (QCT) and other global system partners. The Nvidia RTX PRO 6000 Blackwell Workstation Edition and Nvidia RTX PRO 6000 Blackwell Max-Q Workstation Edition will be available through global distribution partners such as PNY and TD Synnex starting in April, with availability from manufacturers, such as BOXX, Dell, HP Inc., Lambda and Lenovo, starting in May. The Nvidia RTX PRO 5000, RTX PRO 4500 and RTX PRO 4000 Blackwell GPUs will be available in the summer from Boxx, Dell, HP and Lenovo and through global distribution partners. Nvidia RTX PRO Blackwell laptop GPUs will be available from Dell, HP, Lenovo and Razer starting later this year.

Nvidia today announced the Nvidia RTX Pro Blackwell series — a new generation of workstation and server graphics processing units.

Nvidia said the GPUs redefine workflows for AI, technical, creative, engineering and design professionals with accelerated computing, AI inference, ray tracing and neural rendering technologies. The company unveiled the news during the keynote speech of Jensen Huang, CEO of Nvidia, at the GTC 2025 event.

For everything from agentic AI, simulation, extended reality, 3D design and complex visual effects to developing physical AI powering autonomous robots, vehicles and smart spaces, the RTX Pro Blackwell series provides professionals across industries with compute power, memory capacity and data throughput right at their fingertips — from their desktop, on the go with mobile workstations or powered by data center GPUs.

The new lineup includes:

  • Data center GPU: Nvidia RTX PRO 6000 Blackwell Server Edition
  • Desktop GPUs: Nvidia RTX PRO 6000 Blackwell Workstation Edition, Nvidia RTX
  • PRO 6000 Blackwell Max-Q Workstation Edition, Nvidia RTX PRO 5000 Blackwell,
  • Nvidia RTX PRO 4500 Blackwell and Nvidia RTX PRO 4000 Blackwell
  • Laptop GPUs: Nvidia RTX PRO 5000 Blackwell, Nvidia RTX PRO 4000 Blackwell,
  • Nvidia RTX PRO 3000 Blackwell, Nvidia RTX PRO 2000 Blackwell, Nvidia RTX PRO
  • 1000 Blackwell and Nvidia RTX PRO 500 Blackwell

“Software developers, data scientists, artists, designers and engineers need powerful AI and graphics performance to push the boundaries of visual computing and simulation, helping tackle incredible industry challenges,” said Bob Pette, vice president of enterprise platforms at Nvidia, in a statement. “Bringing Nvidia Blackwell to workstations and servers will take productivity, performance and speed to new heights, accelerating AI inference serving, data science, visualization and content creation.”

Nvidia Blackwell technology comes to workstations and data centers

Nvidia RTX PRO Blackwell Desktop GPUs.

RTX Pro Blackwell GPUs unlock the potential of generative, agentic and physical AI by delivering exceptional performance, efficiency and scale.

Nvidia RTX Pro Blackwell GPUs feature:

  • Nvidia Streaming Multiprocessor: Offers up to 1.5 times faster throughput and new neural shaders that integrate AI inside of programmable shaders to drive the next decade of AI-augmented graphics innovations.
  • Fourth-Generation RT Cores: Delivers up to two times the performance of the previous generation to create photoreal, physically accurate scenes and complex 3D designs with optimizations for Nvidia RTX Mega Geometry.
  • Fifth-Generation Tensor Cores: Delivers up to 4,000 AI trillion operations per second and adds support for FP4 precision and Nvidia DLSS 4 Multi Frame Generation, enabling a new era of AI-powered graphics and the ability to run and prototype larger AI models faster.
  • Larger, Faster GDDR7 Memory: Boosts bandwidth and capacity — up to 96GB for workstations and servers and up to 24GB on laptops. This enables applications to run faster and work with larger, more complex datasets for everything from tackling massive 3D and AI projects to exploring large-scale virtual reality environments.
  • Ninth-Generation Nvidia NVENC: Accelerates video encoding speed and improves quality for professional video applications with added support for 4:2:2 encoding.
  • Sixth-Generation Nvidia NVDEC: Provides up to double the H.264 decoding throughput and offers support for 4:2:2 H.264 and HEVC decode. Professionals can benefit from high-quality video playback, accelerate video data ingestion and use advanced AI-powered video editing features.
  • Fifth-Generation PCIe: Support for fifth-generation PCI Express provides double the bandwidth over the previous generation, improving data transfer speeds from CPU memory and unlocking faster performance for data-intensive tasks.
  • DisplayPort 2.1: Drives high-resolution displays at up to 4K at 480Hz and 8K at 165Hz. Increased bandwidth enables seamless multi-monitor setups, while high dynamic range and higher color depth support deliver more precise color accuracy for tasks like video editing, 3D design and live broadcasting.
  • Multi-Instance GPU (MIG): The RTX PRO 6000 data center and desktop GPUs and 5000 series desktop GPUs feature MIG technology, enabling secure partitioning of a single GPU into up to four instances (6000 series) or two instances (5000 series).

Nvidia said Fault isolation is designed to prevent workload interference for secure, efficient resource allocation for diverse workloads, maximizing performance and flexibility.

The new laptop GPUs also support the latest Nvidia Blackwell Max-Q technologies, which intelligently and continually optimize laptop performance and power efficiency with AI.

Nvidia RTX PRO Blackwell Laptop GPUs.

With neural rendering and AI-augmented tools, Nvidia RTX PRO Blackwell GPUs enable the creation of stunning visuals, digital twins of real-world environments and immersive experiences with unprecedented speed and efficiency. The GPUs are built to elevate 3D computer-aided design and building information model workflows, offering designers and engineers exceptional performance for complex modeling, rendering and visualization.

Designed for enterprise data center deployments, the RTX PRO 6000 Blackwell Server Edition features a passively cooled thermal design and can be configured with up to eight GPUs per server. For workloads that require the compute density and scale that data centers offer, the RTX PRO 6000 Blackwell Server Edition delivers powerful performance for next-generation AI, scientific and visual computing applications across industries such as healthcare, manufacturing, retail and media and entertainment.

In addition, this powerful data center GPU can be combined with Nvidia vGPU software to power AI workloads across virtualized environments and deliver high-performance virtual workstation instances to remote users. Nvidia vGPU support for the Nvidia RTX PRO 6000 Blackwell Server Edition GPU is expected in the latter half of this year.

“Foster + Partners has tested the Nvidia RTX PRO 6000 Blackwell Max-Q Workstation Edition GPU on Cyclops, our GPU-based ray-tracing product,” said Martha Tsigkari, head of applied research and development and senior partner at Foster + Partners, in a statement. “The new Nvidia Blackwell GPU has managed to outperform everything we have tested before. For example, when using it with Cyclops, it has performed at 5x the speed of Nvidia RTX A6000 GPUs. Rendering speeds also increased five times, allowing tools like Cyclops to provide feedback on how well our design solutions perform in real time as we design them and resulting in intuitive yet informed decision-making from early conceptual stages.”

“Nvidia RTX PRO 6000 Blackwell Workstation Edition GPUs enable incredibly sharp and photorealistic graphics,” said Jeff Hammoud, chief design officer at Rivian. “In conjunction with a Varjo XR4 headset and Autodesk VRED, the system delivered the level of crispness necessary for immersive automotive design reviews. With Nvidia Blackwell support for PCIe Gen 5, we used two powerful 600W GPUs via VR SLI, allowing us to achieve the highest pixel density and the most stunning visuals we have ever experienced in VR.”

RTX PRO GPUs run on the Nvidia AI platform and feature larger memory capacity and the latest Tensor Cores to accelerate a deep ecosystem of AI-accelerated applications built on Nvidia CUDA and RTX technology. With everything from the latest AI-based content creation tools and new reasoning models, such as the Nvidia Llama Nemotron Reason family of models and Nvidia NIM microservices unveiled today, inferencing is faster than ever. And with over 400 Nvidia CUDA-X libraries, developers can easily build, optimize, deploy and scale new AI applications, from workstations to the data center or cloud.

Enterprises can fast-track their AI development and deployments by prototyping locally with an Nvidia RTX PRO GPU and the Nvidia Omniverse and Nvidia AI Enterprise platforms, Nvidia Blueprints and Nvidia NIM, which gives access to easy-to-use inference microservices backed by enterprise-level support. They can also run these applications at scale on the ultimate universal data center GPU for AI and visual computing, delivering breakthrough acceleration for the most demanding compute-intensive enterprise workloads with the RTX Pro 6000 Blackwell Server Edition.

Availability

The Nvidia RTX Pro 6000 Blackwell Server Edition will soon be available in server configurations from leading data center system partners including Cisco, Dell Technologies, Hewlett Packard Enterprise, Lenovo and Supermicro.

Cloud service providers and GPU cloud providers including AWS, Google Cloud, Microsoft Azure and CoreWeave will be among the first to offer instances powered by the Nvidia RTX Pro 6000 Blackwell Server Edition later this year. In addition, the server edition GPU will be available in data center platforms from ASUS, GIGABYTE, Ingrasys, Quanta Cloud Technology (QCT) and other global system partners.

The Nvidia RTX PRO 6000 Blackwell Workstation Edition and Nvidia RTX PRO 6000 Blackwell Max-Q Workstation Edition will be available through global distribution partners such as PNY and TD Synnex starting in April, with availability from manufacturers, such as BOXX, Dell, HP Inc., Lambda and Lenovo, starting in May.

The Nvidia RTX PRO 5000, RTX PRO 4500 and RTX PRO 4000 Blackwell GPUs will be available in the summer from Boxx, Dell, HP and Lenovo and through global distribution partners.

Nvidia RTX PRO Blackwell laptop GPUs will be available from Dell, HP, Lenovo and Razer starting later this year.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

IonQ, Alice & Bob roll out quantum breakthroughs

“And therefore, does bring us closer to escape velocity,” he added. “When this will happen, and we’ll be able to present an Alice & Bob’s logical qubit under threshold depends on a variety of factors, but we can openly say that this is the current big work happening in the

Read More »

Oil Dips as Economic Gloom Undercuts Middle East Supply Risk

Oil slipped as broader market weakness and concerns about a global glut of crude overshadowed escalating tensions in the Middle East. West Texas Intermediate fell 1% to settle below $67 a barrel, snapping a two-session winning streak. US equities weakened a day ahead of a Federal Reserve decision set to provide clarity on the economic effects of ongoing trade wars. Oil earlier rose as much as 1.7% after Israel conducted military strikes across Gaza, while the US increases pressure on Iran.  “Crude has been pricing in only a minimal geopolitical risk premium as tensions between Israel and Hamas resurface,” said Rebecca Babin, a senior energy trader at CIBC Private Wealth Group. “Most traders view the premiums as selling opportunities, positioning for inventory builds later in the year and increasing macroeconomic risks.” Crude is on track for a quarterly loss as an escalating global trade war threatens demand while OPEC and its allies are set to raise production starting in April. The global market was already set for a glut, according to the International Energy Agency. Russian President Vladimir Putin agreed in a phone call with US President Donald Trump to limit attacks on Ukrainian energy assets, but stopped short of committing to a broader 30-day ceasefire. Still, some market participants have been quick to hedge against a pickup in geopolitical risk. Premiums on bearish put options declined relative to bullish calls on Monday, and the day’s trading was dominated by a flurry of $100-a-barrel wagers.  Prices: WTI for April delivery fell 1% to settle at $66.90 a barrel in New York. Brent for May settlement slipped 0.7% to settle at $70.56 a barrel. WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate or insulting comments

Read More »

Sembcorp Is Said to Conduct Strategic Review of Business

Sembcorp Industries Ltd. is carrying out a strategic review of its business, with various possibilities under consideration, including potentially going private or selling some assets, people with knowledge of the matter said. The Singaporean firm, which is backed by state-owned Temasek Holdings Pte, is working with a financial adviser on the review, according to the people, who asked not to be identified because the matter is private.  Sembcorp’s stock rose as much as 4.9%, the steepest intraday gain since Aug 7. Considerations are at an early stage and Sembcorp may decide not to pursue any deal, the people said.  A spokesperson for Sembcorp said the company always assesses options to enhance shareholder value, including investments, divestment, restructuring, listing and capital management.  “As part of such assessments, we work with consultants from time to time, as the needs arise” the spokesperson said. “The assessed options may or may not result in any form of transaction.”  “We will make the necessary announcements in accordance with the requirements in the SGX-ST Listing Manual where an initiative is material and has progressed to a stage where there is a high level of certainty,” the spokesperson added.  Sembcorp’s businesses include wind, solar, hydro and energy storage operations in Southeast Asia, China, India, the UK and Middle East, as well as gas and diesel-fired power plants in similar geographies, urban solutions such as waste and water management, and corporate decarbonization services.  Past restructurings include Covid-hit Sembcorp Marine’s demerger in 2020. Sembcorp Industries’ chief executive officer at the time said the move would help it focus on the energy and urban sectors. Sembcorp Industries also completed divestments in Chile, China and Panama that year. In 2024, it announced the sale of Sembcorp Environment Pte for S$405 million. Sembcorp’s shares have climbed more than 250% since the end of 2020, giving the

Read More »

But first, eat your vegetables: A guide to smarter load management

For more than a decade, the phrase “first, eat your energy efficiency vegetables” has been used to encourage consumers and utilities to prioritize reducing energy use before making changes to how energy is sourced. This principle remains true today, but with a new twist: electrification. As utilities navigate the evolving energy landscape, ensuring that efficiency and electrification work hand-in-hand is critical to reducing emissions, optimizing grid performance, and keeping customer bills in check. The Foundation Energy efficiency (EE) has long been the go-to strategy for utilities looking to lower energy consumption, reduce peak demand, and help customers save money. Electrification, on the other hand, can increase energy demand—especially if customers adopt electric vehicles (EVs) and heat pumps without implementing efficiency measures first. But when done right, electrification can be a powerful tool for decarbonization while still maintaining the benefits of energy efficiency. Why Smart Load Management Matters Electrification without efficiency is like skipping the vegetables and going straight for dessert—it may taste good in the short term but leads to problems down the line. Increased electricity consumption from EVs, heat pumps, and other electric technologies can put additional stress on the grid. The key is to integrate energy efficiency strategies with electrification efforts to balance load growth and avoid excessive peak demand. Targeting Customers Who Are Ready for the Next Step Not every customer is ready for a full-scale energy retrofit, but many are open to incremental improvements. The good news? Data shows that those who adopt smart thermostats or EVs are often inclined to pursue further energy efficiency improvements, and vice versa. Utilities can leverage this customer propensity by using targeted marketing and strategic program design to enhance energy efficiency and demand response (DR) participation. Here’s how: 1. Promote Heat Pumps to Homes with Recent Insulation Work A home

Read More »

APA Hits Oil Discovery on Alaskan North Slope

APA Corp. and its partners have declared a “significant” petroleum discovery in the Lagniappe area on Alaska’s North Slope. Drilled to about 10,500 feet, the Sockeye-2 well “encountered a high-quality reservoir with approximately 25 feet of net oil pay in one blocky, Paleocene-aged sand with an average porosity of 20 percent”, Houston, Texas-based APA said in an online statement. “As compared to recent regional field analogs in the Brookian play, the porosity and permeability are both better than expected, with the permeability to be confirmed through a planned flow test”, the oil and gas explorer and producer said. “Additional zones of potential pay were also encountered in the shallower Staines Tongue formation. “The Sockeye prospect is amplitude supported across 25,000 to 30,000 acres, and confirms the partners’ geologic and geophysical models, derisking numerous additional prospects in the area. “Wireline logging is complete and additional data collection is underway, including acquiring core and flow testing the well”. APA holds a 50 percent stake in the leasehold through subsidiary Apache Corp. Operator Lagniappe Alaska LLC of Armstrong Oil & Gas Inc. owns 25 percent. Santos Ltd. has the remaining 25 percent. Sockeye-2 is the second successful exploratory well drilled by the consortium on their 325,411-acre position on state lands. The first discovery, King Street-1, struck oil in two Brookian zones. “The Sockeye-2 well further demonstrates the potential of the play, presenting an exciting opportunity in an active area of the North Slope with significant existing infrastructure”, commented Bill Armstrong, chief executive of Armstrong Oil & Gas. The leasehold sits east of Prudhoe Bay, one of the biggest oilfields in the United States. In a separate statement, Santos said, “The exploration well cost is carried by APA as part of a 2023 farm-in agreement”. On a separate upstream development in Alaska, Santos also said phase

Read More »

New York state offers $5M for existing building energy innovations

The New York State Energy Research and Development Authority has allocated $5 million to support building solutions that can be demonstrated in existing buildings, including small- to medium-sized commercial, multifamily and single-family buildings, the department announced Friday. Through the Enabling Innovative Clean Energy Building Solutions program, the funding aims to accelerate the market-readiness of new and emerging clean energy and energy efficiency technologies that can help to manage future energy use, enhance energy efficiency and lower operating costs in buildings across the state, NYSERDA says. “By addressing critical areas like clean heating and cooling, energy efficiency, and building decarbonization, NYSERDA … will not only help the State meet its climate goals but also support long-term resilience and cost savings for tenants and building owners across New York State,” Richard Yancey, CEO of the Building Energy Exchange, said in a statement.  The program seeks applications for projects and programs that demonstrate the “latest innovations in integrated systems, materials and technologies that can greatly improve building or home performance, reduce costs, and/or improve resiliency,” NYSERDA said. The maximum funding for each selected proposal is $1 million, with eligible projects being required to have been previously awarded funding from a federal, state or local government entity — besides NYSERDA — foundation or non-profit to develop an innovative clean building technology product or solution, per the release. Eligible funding areas include solutions related to building envelopes, such as components or materials that improve performance or enable building envelope retrofit; clean heating and cooling; intelligent grid-interactive building controls; low-global warming potential refrigerant applications and leak detection; or thermal storage.  Clean heating and cooling solutions include new heat pump equipment, ground source systems and hybrid systems, with grid-interactive building controls including those that enable electric load and energy asset management. Thermal storage solutions that integrate heating,

Read More »

Microgrids called a low-burden way to ensure backup power

Microgrids don’t pose an immediate threat to diesel generators as the way most managers ensure their facilities have backup power, but Allan Schurr, the chief commercial officer at microgrid company Enchanted Rock, expects more operators to look at what he says is a cleaner, administratively easier approach that microgrids offer.  “Diesel generators are in short supply and sometimes it’s two years before you can get a diesel generator delivered,” Schurr told Facilities Dive. “By having a microgrid … you can [have something] more reliable and cleaner” and have it up and running within a year.” A microgrid is a self-contained electrical network that companies like Schurr’s install on their client’s property. Enchanted Rock builds its network using half-megawatt natural gas generators that it gangs together based on the amount of load the client wants to back up. For a hospital, it might be a network of 10 generators. For a grocery store, it might be two or three.  “Our standardization of equipment is part of our differentiator,” said Schurr, referring to his company’s use of a single-size generator. “That allows us to dial in exactly the amount of backup power that’s needed for any facility size.” Power from the microgrid is cleaner, Schurr says. In the case of Enchanted Rock, it’s cleaner because it’s generated from natural gas. But microgrids can also use solar, wind and other non-fossil-fuel burning sources.  The cleaner power might be important to organizations that make sustainability a priority but it can also help lower energy costs by enabling the microgrid to run power outside of emergency situations and sell it as a supplemental power source to the local utility grid. That revenue goes to the microgrid and helps offset what the property owner pays for its backup power.  “So, [our clients] can get the air

Read More »

Jio teams with AMD, Cisco and Nokia to build AI-enabled telecom platform

Its accessible and flexible design will allow telecom providers, across the globe, to add AI capabilities to various parts of their infrastructure, including radio access networks (RAN), security systems, data centers, and network routing. This will make networks more manageable and responsive to issues, resulting in a smoother user experience. Jio will be the first to try out this new platform in hopes of setting an example for other telecom providers around the world. According to Jio, the goal is to create a model that others can easily follow and put into action going forward. Oommen expects the first innovation from this alliance to come in this calendar year. For Cisco, this alliance is a continuation of the strategy put in place several years ago to be more interoperable and open. Historically, Cisco has been accused of being closed and proprietary and keeping customers locked into their products. While I think the competitive chatter was far greater than the reality, there was some truth to it. This is something Jeetu Patel, who took on the role of chief product officer last year, has changed. “Over the past three to four years, we have made a huge amount of progress in this area. We have partnerships with Microsoft in security; our collaboration products work with Microsoft Teams and Zoom; our XDR takes in telemetry from competitors; and we have partnerships with multiple AI companies,” Patel told me. The willingness to be open is something all vendors should embrace as openness creates better competition and drives innovation and brings costs down. The democratization of AI creates a better world, like the way making the Internet ubiquitous did. The Open Telecom AI Platform allows more companies to participate in AI and share in the upside creating a “rising tide” that will fundamentally transform

Read More »

Enterprises reevaluate virtualization strategies amid Broadcom uncertainty

This dilemma of whether to absorb the Broadcom price hikes or embark on the arduous and risky journey of untangling from the VMware ecosystem is triggering a broader C-level conversation around virtualization strategy. “For enterprises navigating this uncertainly, the challenge isn’t just finding a replacement for VMware. IT shops of all sizes see Broadcom’s actions as an opportunity to rethink their approach to virtualization, cloud strategy and IT modernization,” says Steve McDowell, chief analyst at NAND Research. Elliot says that server virtualization has been taken for granted for a long time, and the Broadcom-driven wake-up call is forcing organizations to reevaluate their virtualization strategies at the board level. “That kind of strategic conversation hasn’t happened for years. Customers are saying, ‘What can we do as this platform emerges from VMware. How do we think about this relative to our multi-cloud strategy and private cloud and the efficiencies we can gain? Let’s talk about risk reduction. Let’s talk about platform strategy.’ This is an opportunity to identify business value. It’s triggering this plethora of swim lanes.” Check the waters before diving in While there are multiple alternatives to the VMware platform, none of them are as good from a feature perspective, and there’s a risk associated with moving off a tried-and-true platform. In estimating the cost of a large-scale VMware migration, Gartner cautions: “VMware’s server virtualization platform has become the point of integration for its customers across server, storage and network infrastructure in the data center. Equally, it is a focus of IT operational duties including workload provisioning, backup and disaster recovery. Migrating from VMware’s server virtualization platform would require untangling many aspects of these investments.” It would take a midsize enterprise at least two years to untangle much of its dependency upon VMware, and it could take a large enterprise

Read More »

5 alternatives to VMware vSphere virtualization platform

Nutanix – which is actively courting disgruntled VMware customers – provides storage services that aggregate storage in a global pool that enables any VM to access and consume storage resources. Features include compression, deduplication, high-availability and snapshots. Enterprises running high-performance databases often require external storage arrays, and Nutanix has addressed that need by certifying storage with SAP HANA and Oracle RAC. (Read more: Cisco, Nutanix strengthen joint HCI package) 4. Scale Computing Platform Scale provides an all-in-one hardware and software package that includes all software licenses. Software features offered at no additional charge include high-availability clustering, built-in disaster recovery, replication and software-defined storage. Scale also offers a tool to automate migrations off vSphere, a centralized management feature for HCI clusters, and the ability to mix and match dissimilar hardware appliances in a cluster. In addition, all storage is pooled. Last summer, Scale Computing said in a quarterly earnings announcement that sales have taken off, thanks in part to Broadcom’s changes to VMware sales operations.  5. VergeIO VergeIO takes HCI to the next level with something it calls ultraconverged infrastructure (UCI). This means VergeIO can not only virtualize the normal stack of compute, networking and storage, it can also implement multi-tenancy, creating multiple virtual data centers (VDCs). Each VDC has its own compute, network, storage, management and VergeOS assigned to it. Enterprises can manage and use each VDC much like the virtual private clouds offered by the hyperscalers. VergeIO says this model creates greater workload density, which means lower costs, improved availability, and simplified IT.

Read More »

IBM laying foundation for mainframe as ultimate AI server

“It will truly change what customers are able to do with AI,” Stowell said. IBM’s mainframe processors The next generation of processors is expected to continue a long history of generation-to-generation improvements, IBM stated in a new white paper on AI and the mainframe. “They are projected to clock in at 5.5 GHz. and include ten 36 MB level 2 caches. They’ll feature built-in low-latency data processing for accelerated I/O as well as a completely redesigned cache and chip-interconnection infrastructure for more on-chip cache and compute capacity,” IBM wrote.  Today’s mainframes also have extensions and accelerators that integrate with the core systems. These specialized add-ons are designed to enable the adoption of technologies such as Java, cloud and AI by accelerating computing paradigms that are essential for high-volume, low-latency transaction processing, IBM wrote.  “The next crop of AI accelerators are expected to be significantly enhanced—with each accelerator designed to deliver 4 times more compute power, reaching 24 trillion operations per second (TOPS),” IBM wrote. “The I/O and cache improvements will enable even faster processing and analysis of large amounts of data and consolidation of workloads running across multiple servers, for savings in data center space and power costs. And the new accelerators will provide increased capacity to enable additional transaction clock time to perform enhanced in-transaction AI inferencing.” In addition, the next generation of the accelerator architecture is expected to be more efficient for AI tasks. “Unlike standard CPUs, the chip architecture will have a simpler layout, designed to send data directly from one compute engine, and use a range of lower- precision numeric formats. These enhancements are expected to make running AI models more energy efficient and far less memory intensive. As a result, mainframe users can leverage much more complex AI models and perform AI inferencing at a greater scale

Read More »

VergeIO enhances VergeFabric network virtualization offering

VergeIO is not, however, using an off-the-shelf version of KVM. Rather, it is using what Crump referred to as a heavily modified KVM hypervisor base, with significant proprietary enhancements while still maintaining connections to the open-source community. VergeIO’s deployment profile is currently 70% on premises and about 30% via bare-metal service providers, with a particularly strong following among cloud service providers that host applications for their customers. The software requires direct hardware access due to its low-level integration with physical resources. “Since November of 2023, the normal number one customer we’re attracting right now is guys that have had a heart attack when they got their VMware renewal license,” Crump said. “The more of the stack you own, the better our story becomes.” A 2024 report from Data Center Intelligence Group (DCIG) identified VergeOS as one of the top 5 alternatives to VMware. “VergeIO starts by installing VergeOS on bare metal servers,” the report stated. “It then brings the servers’ hardware resources under its management, catalogs these resources, and makes them available to VMs. By directly accessing and managing the server’s hardware resources, it optimizes them in ways other hypervisors often cannot.” Advanced networking features in VergeFabric VergeFabric is the networking component within the VergeOS ecosystem, providing software-defined networking capabilities as an integrated service rather than as a separate virtual machine or application.

Read More »

Podcast: On the Frontier of Modular Edge AI Data Centers with Flexnode’s Andrew Lindsey

The modular data center industry is undergoing a seismic shift in the age of AI, and few are as deeply embedded in this transformation as Andrew Lindsey, Co-Founder and CEO of Flexnode. In a recent episode of the Data Center Frontier Show podcast, Lindsey joined Editor-in-Chief Matt Vincent and Senior Editor David Chernicoff to discuss the evolution of modular data centers, the growing demand for high-density liquid-cooled solutions, and the industry factors driving this momentum. A Background Rooted in Innovation Lindsey’s career has been defined by the intersection of technology and the built environment. Prior to launching Flexnode, he worked at Alpha Corporation, a top 100 engineering and construction management firm founded by his father in 1979. His early career involved spearheading technology adoption within the firm, with a focus on high-security infrastructure for both government and private clients. Recognizing a massive opportunity in the data center space, Lindsey saw a need for an innovative approach to infrastructure deployment. “The construction industry is relatively uninnovative,” he explained, citing a McKinsey study that ranked construction as the second least-digitized industry—just above fishing and wildlife, which remains deliberately undigitized. Given the billions of square feet of data center infrastructure required in a relatively short timeframe, Lindsey set out to streamline and modernize the process. Founded four years ago, Flexnode delivers modular data centers with a fully integrated approach, handling everything from site selection to design, engineering, manufacturing, deployment, operations, and even end-of-life decommissioning. Their core mission is to provide an “easy button” for high-density computing solutions, including cloud and dedicated GPU infrastructure, allowing faster and more efficient deployment of modular data centers. The Rising Momentum for Modular Data Centers As Vincent noted, Data Center Frontier has closely tracked the increasing traction of modular infrastructure. Lindsey has been at the forefront of this

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »