Stay Ahead, Stay ONMINE

Why Data Scientists Should Care about Containers — and Stand Out with This Knowledge

“I train models, analyze data and create dashboards — why should I care about Containers?” Many people who are new to the world of data science ask themselves this question. But imagine you have trained a model that runs perfectly on your laptop. However, error messages keep popping up in the cloud when others access […]

“I train models, analyze data and create dashboards — why should I care about Containers?”

Many people who are new to the world of data science ask themselves this question. But imagine you have trained a model that runs perfectly on your laptop. However, error messages keep popping up in the cloud when others access it — for example because they are using different library versions.

This is where containers come into play: They allow us to make machine learning models, data pipelines and development environments stable, portable and scalable — regardless of where they are executed.

Let’s take a closer look.

Table of Contents
1 — Containers vs. Virtual Machines: Why containers are more flexible than VMs
2 — Containers & Data Science: Do I really need Containers? And 4 reasons why the answer is yes.
3 — First Practice, then Theory: Container creation even without much prior knowledge
4 — Your 101 Cheatsheet: The most important Docker commands & concepts at a glance
Final Thoughts: Key takeaways as a data scientist
Where Can You Continue Learning?

1 — Containers vs. Virtual Machines: Why containers are more flexible than VMs

Containers are lightweight, isolated environments. They contain applications with all their dependencies. They also share the kernel of the host operating system, making them fast, portable and resource-efficient.

I have written extensively about virtual machines (VMs) and virtualization in ‘Virtualization & Containers for Data Science Newbiews’. But the most important thing is that VMs simulate complete computers and have their own operating system with their own kernel on a hypervisor. This means that they require more resources, but also offer greater isolation.

Both containers and VMs are virtualization technologies.

Both make it possible to run applications in an isolated environment.

But in the two descriptions, you can also see the 3 most important differences:

  • Architecture: While each VM has its own operating system (OS) and runs on a hypervisor, containers share the kernel of the host operating system. However, containers still run in isolation from each other. A hypervisor is the software or firmware layer that manages VMs and abstracts the operating system of the VMs from the physical hardware. This makes it possible to run multiple VMs on a single physical server.
  • Resource consumption: As each VM contains a complete OS, it requires a lot of memory and CPU. Containers, on the other hand, are more lightweight because they share the host OS.
  • Portability: You have to customize a VM for different environments because it requires its own operating system with specific drivers and configurations that depend on the underlying hardware. A container, on the other hand, can be created once and runs anywhere a container runtime is available (Linux, Windows, cloud, on-premise). Container runtime is the software that creates, starts and manages containers — the best-known example is Docker.
Created by the author

You can experiment faster with Docker — whether you’re testing a new ML model or setting up a data pipeline. You can package everything in a container and run it immediately. And you don’t have any “It works on my machine”-problems. Your container runs the same everywhere — so you can simply share it.

2 — Containers & Data Science: Do I really need Containers? And 4 reasons why the answer is yes.

As a data scientist, your main task is to analyze, process and model data to gain valuable insights and predictions, which in turn are important for management.

Of course, you don’t need to have the same in-depth knowledge of containers, Docker or Kubernetes as a DevOps Engineer or a Site Reliability Engineer (SRE). Nevertheless, it is worth having container knowledge at a basic level — because these are 4 examples of where you will come into contact with it sooner or later:

Model deployment

You are training a model. You not only want to use it locally but also make it available to others. To do this, you can pack it into a container and make it available via a REST API.

Let’s look at a concrete example: Your trained model runs in a Docker container with FastAPI or Flask. The server receives the requests, processes the data and returns ML predictions in real-time.

Reproducibility and easier collaboration

ML models and pipelines require specific libraries. For example, if you want to use a deep learning model like a Transformer, you need TensorFlow or PyTorch. If you want to train and evaluate classic machine learning models, you need Scikit-Learn, NumPy and Pandas. A Docker container now ensures that your code runs with exactly the same dependencies on every computer, server or in the cloud. You can also deploy a Jupyter Notebook environment as a container so that other people can access it and use exactly the same packages and settings.

Cloud integration

Containers include all packages, dependencies and configurations that an application requires. They therefore run uniformly on local computers, servers or cloud environments. This means you don’t have to reconfigure the environment.

For example, you write a data pipeline script. This works locally for you. As soon as you deploy it as a container, you can be sure that it will run in exactly the same way on AWS, Azure, GCP or the IBM Cloud.

Scaling with Kubernetes

Kubernetes helps you to orchestrate containers. But more on that below. If you now get a lot of requests for your ML model, you can scale it automatically with Kubernetes. This means that more instances of the container are started.

3 — First Practice, then Theory: Container creation even without much prior knowledge

Let’s take a look at an example that anyone can run through with minimal time — even if you haven’t heard much about Docker and containers. It took me 30 minutes.

We’ll set up a Jupyter Notebook inside a Docker container, creating a portable, reproducible Data Science environment. Once it’s up and running, we can easily share it with others and ensure that everyone works with the exact same setup.

0 — Install Docker Dekstop and create a project directory

To be able to use containers, we need Docker Desktop. To do this, we download Docker Desktop from the official website.

Now we create a new folder for the project. You can do this directly in the desired folder. I do this via Terminal — on Windows with Windows + R and open CMD.

We use the following command:

Screenshot taken by the author

1. Create a Dockerfile

Now we open VS Code or another editor and create a new file with the name ‘Dockerfile’. We save this file without an extension in the same directory. Why doesn’t it need an extension?

We add the following code to this file:

# Use the official Jupyter notebook image with SciPy
FROM jupyter/scipy-notebook:latest  

# Set the working directory inside the container
WORKDIR /home/jovyan/work  

# Copy all local files into the container
COPY . .

# Start Jupyter Notebook without token
CMD ["start-notebook.sh", "--NotebookApp.token=''"]

We have thus defined a container environment for Jupyter Notebook that is based on the official Jupyter SciPy Notebook image.

First, we define with FROM on which base image the container is built. jupyter/scipy-notebook:latest is a preconfigured Jupyter notebook image and contains libraries such as NumPy, SiPy, Matplotlib or Pandas. Alternatively, we could also use a different image here.

With WORKDIR we set the working directory within the container. /home/jovyan/work is the default path used by Jupyter. User jovyan is the default user in Jupyter Docker images. Another directory could also be selected — but this directory is best practice for Jupyter containers.

With COPY . . we copy all files from the local directory — in this case the Dockerfile, which is located in the jupyter-docker directory — to the working directory /home/jovyan/work in the container.

With CMD [“start-notebook.sh”, “ — NotebookApp.token=‘’’”] we specify the default start command for the container, specify the start script for Jupyter Notebook and define that the notebook is started without a token — this allows us to access it directly via the browser.

2. Create the Docker image

Next, we will build the Docker image. Make sure you have the previously installed Docker desktop open. We now go back to the terminal and use the following command:

cd jupyter-docker
docker build -t my-jupyter .

With cd jupyter-docker we navigate to the folder we created earlier. With docker build we create a Docker image from the Dockerfile. With -t my-jupyter we give the image a name. The dot means that the image will be built based on the current directory. What does that mean? Note the space between the image name and the dot.

The Docker image is the template for the container. This image contains everything needed for the application such as the operating system base (e.g. Ubuntu, Python, Jupyter), dependencies such as Pandas, Numpy, Jupyter Notebook, the application code and the startup commands. When we “build” a Docker image, this means that Docker reads the Dockerfile and executes the steps that we have defined there. The container can then be started from this template (Docker image).

We can now watch the Docker image being built in the terminal.

Screenshot taken by the author

We use docker images to check whether the image exists. If the output my-jupyter appears, the creation was successful.

docker images

If yes, we see the data for the created Docker image:

Screenshot taken by the author

3. Start Jupyter container

Next, we want to start the container and use this command to do so:

docker run -p 8888:8888 my-jupyter

We start a container with docker run. First, we enter the specific name of the container that we want to start. And with -p 8888:8888 we connect the local port (8888) with the port in the container (8888). Jupyter runs on this port. I do not understand.

Alternatively, you can also perform this step in Docker desktop:

Screenshot taken by the author

4. Open Jupyter Notebook & create a test notebook

Now we open the URL [http://localhost:8888](http://localhost:8888/) in the browser. You should now see the Jupyter Notebook interface.

Here we will now create a Python 3 notebook and insert the following Python code into it.

import numpy as np
import matplotlib.pyplot as plt

x = np.linspace(0, 10, 100)
y = np.sin(x)

plt.plot(x, y)
plt.title("Sine Wave")
plt.show()

Running the code will display the sine curve:

Screenshot taken by the author

5. Terminate the container

At the end, we end the container either with ‘CTRL + C’ in the terminal or in Docker Desktop.

With docker ps we can check in the terminal whether containers are still running and with docker ps -a we can display the container that has just been terminated:

Screenshot taken by the author

6. Share your Docker image

If you now want to upload your Docker image to a registry, you can do this with the following command. This will upload your image to Docker Hub (you need a Docker Hub account for this). You can also upload it to a private registry of AWS Elastic Container, Google Container, Azure Container or IBM Cloud Container.

docker login

docker tag my-jupyter your-dockerhub-name/my-jupyter:latest

docker push dein-dockerhub-name/mein-jupyter:latest

If you then open Docker Hub and go to your repositories in your profile, the image should be visible.

This was a very simple example to get started with Docker. If you want to dive a little deeper, you can deploy a trained ML model with FastAPI via a container.

4 — Your 101 Cheatsheet: The most important Docker commands & concepts at a glance

You can actually think of a container like a shipping container. Regardless of whether you load it onto a ship (local computer), a truck (cloud server) or a train (data center) — the content always remains the same.

The most important Docker terms

  • Container: Lightweight, isolated environment for applications that contains all dependencies.
  • Docker: The most popular container platform that allows you to create and manage containers.
  • Docker Image: A read-only template that contains code, dependencies and system libraries.
  • Dockerfile: Text file with commands to create a Docker image.
  • Kubernetes: Orchestration tool to manage many containers automatically.

The basic concepts behind containers

  • Isolation: Each container contains its own processes, libraries and dependencies
  • Portability: Containers run wherever a container runtime is installed.
  • Reproducibility: You can create a container once and it runs exactly the same everywhere.

The most basic Docker commands

docker --version # Check if Docker is installed
docker ps # Show running containers
docker ps -a # Show all containers (including stopped ones)
docker images # List of all available images
docker info # Show system information about the Docker installation

docker run hello-world # Start a test container
docker run -d -p 8080:80 nginx # Start Nginx in the background (-d) with port forwarding
docker run -it ubuntu bash # Start interactive Ubuntu container with bash

docker pull ubuntu # Load an image from Docker Hub
docker build -t my-app . # Build an image from a Dockerfile

Final Thoughts: Key takeaways as a data scientist

👉 With Containers you can solve the “It works on my machine” problem. Containers ensure that ML models, data pipelines, and environments run identically everywhere, independent of OS or dependencies.

👉 Containers are more lightweight and flexible than virtual machines. While VMs come with their own operating system and consume more resources, containers share the host operating system and start faster.

👉 There are three key steps when working with containers: Create a Dockerfile to define the environment, use docker build to create an image, and run it with docker run — optionally pushing it to a registry with docker push.

And then there’s Kubernetes.

A term that comes up a lot in this context: An orchestration tool that automates container management, ensuring scalability, load balancing and fault recovery. This is particularly useful for microservices and cloud applications.

Before Docker, VMs were the go-to solution (see more in ‘Virtualization & Containers for Data Science Newbiews’.) VMs offer strong isolation, but require more resources and start slower.

So, Docker was developed in 2013 by Solomon Hykes to solve this problem. Instead of virtualizing entire operating systems, containers run independently of the environment — whether on your laptop, a server or in the cloud. They contain all the necessary dependencies so that they work consistently everywhere.

I simplify tech for curious minds🚀 If you enjoy my tech insights on Python, data science, Data Engineering, machine learning and AI, consider subscribing to my substack.

Where Can You Continue Learning?

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Crooks are hijacking and reselling AI infrastructure: Report

Threat actors may not only be stealing AI access from fully developed applications, the researchers added. A developer trying to prototype an app, who, through carelessness, doesn’t secure a server, could be victimized through credential theft as well. Joseph Steinberg, a US-based AI and cybersecurity expert, said the report is

Read More »

Amazon confirms 16,000 job cuts, including to AWS

Amazon is cutting about 16,000 jobs across the company, SVP of People Experience and Technology Beth Galetti wrote in an email to employees Wednesday. The cuts were widely expected — and although Galetti’s email did not mention Amazon Web Services, the cuts came as no surprise to AWS staff, some

Read More »

Department of Energy Seeks Hosts for Nuclear Lifecycle Innovation Campuses

New Effort to Modernize America’s Nuclear Fuel Cycle and Support Advanced Reactor Deployment WASHINGTON—The U.S. Department of Energy (DOE) today issued a Request for Information (RFI) inviting states to express interest in hosting Nuclear Lifecycle Innovation Campuses, a new effort to modernize the nation’s full nuclear fuel cycle and strengthen America’s leadership in advanced nuclear energy. This action marks the first step towards potentially establishing voluntary Federal-State partnerships designed to advance regional economic growth, enhance national energy security, and build a coherent, end-to-end nuclear energy strategy for the country. “Unleashing the next American nuclear renaissance will drive innovation, fuel economic growth, and create good-paying American jobs while delivering the affordable, reliable and secure energy America needs to power its future,” said U.S. Energy Secretary Chris Wright. “Nuclear Lifecycle Innovation Campuses give us the opportunity to work directly with states on regional priorities that support President Trump’s vision to revitalize America’s nuclear base.” The proposed campuses could support activities across the full nuclear fuel lifecycle, including fuel fabrication, enrichment, reprocessing used nuclear fuel, and disposition of waste. Depending on state priorities and regional capabilities, the sites could also host advanced reactor deployment, power generation, advanced manufacturing, and co located data centers. DOE is inviting states to provide clear statements of interest and constructive feedback on the structure of the Innovation Campuses. Submissions should outline state priorities—such as workforce development, infrastructure investment, economic diversification, or technology leadership— and describe the scope of activities the state envisions hosting. States are also encouraged to identify the funding structures, risk sharing approaches, incentives and federal partnerships required to successfully establish and sustain a full-cycle Innovation Campus.  Responses to the RFI are requested no later than April 1, 2026. Interested parties can learn more and respond at SAM.gov. ###

Read More »

Oil Options on Longest Bullish Run Since 2024

Oil traders are paying a premium for bullish call options for the longest stretch in about 14 months as they huddle in the options market to protect against the risk of a new confrontation between the US and Iran.  The global Brent benchmark has registered a call skew for 14 consecutive sessions, while the equivalent US marker has seen such a pattern for the 13 most recent trading days. Those are the longest stretches since late 2024, when Israel launched attacks on Iranian military installations.  Thousands are estimated to have been killed in the recent wave of unrest to challenge Supreme Leader Ayatollah Ali Khamenei and his regime, sparking an international outcry, including warnings from US President Donald Trump of “strong action” if the killings did not stop. Trump said this week that the US has a “big armada” headed to the Middle East because of Iran, but added that he hoped the US won’t have to use it.  Options markets have been the main way traders have wagered on heightened geopolitical risk in the Middle East in recent years, in a period that started with Hamas’s attack on Israel in October 2023. When the US struck Iran last year, premiums for calls spiked and then collapsed after it became apparent that oil facilities had been spared.  “The focus on Iran continues,” said Arne Lohmann Rasmussen, chief analyst at A/S Global Risk Management. “The market will likely remain nervous over the coming days.” The uncertainty is leading to chunky additions of bullish options contracts. Open interest in Brent call options has accrued at the fastest pace this month in at least six years, according to Bloomberg calculations of ICE Futures Europe data. It follows the busiest ever day of Brent crude call options trading earlier this month. Hedge funds have also boosted net-bullish wagers on crude

Read More »

Trump Iran Threat Pushes Oil Higher

Oil rose to a fresh four-month high after US President Donald Trump threatened another attack on Iran, urging Tehran to negotiate a nuclear deal. “Hopefully Iran will quickly ‘Come to the Table’ and negotiate a fair and equitable deal,” Trump said in a post on his Truth Social network, adding that “the next attack will be far worse!” than one that took place last year. The potential risk to Iranian supplies has injected a premium into oil prices and led futures to start the year on a strong footing, up more than 10% this month, despite forecasts for a glut. That has also kept the cost of bullish options high relative to bearish ones. West Texas Intermediate futures settled above $63 a barrel after Trump’s post, the highest level since the end of September, extending a 2.9% jump in the previous session. Prices eased off of intra-day highs after Iran’s mission to the UN repeated in a post on X that it stands ready for dialogue based on mutual respect and interests, but said it will “defend itself and respond like never before,” to US aggression. Further capping gains, a gauge of the dollar rebounded after Treasury Secretary Scott Bessent said the US continues to have a “strong dollar” policy under Trump, and denied that the administration is intervening in FX markets, specifically to sell the dollar against the yen. The uptick in the dollar made commodities priced in the currency less attractive. Trump on Wednesday also said the fleet of US ships he’d ordered to the Middle East is larger than the one sent to Venezuela, where President Nicolas Maduro was outed by US forces earlier this year. There’s already been regional reaction to Trump’s signals over recent days. The Iranian and Qatari foreign ministers stressed the need to

Read More »

Greer Says India Must Do More to Wind Down Russian Oil Buys

India has more work to do in order to satisfy US concerns about its purchases of Russian oil and secure tariff relief, President Donald Trump’s trade representative said. While New Delhi has “made a lot of progress” on curbing buys of Russian crude, “it’s hard for them” to completely wean off the supplies because “they like the discount that you get from Russian oil,” US Trade Representative Jamieson Greer said Tuesday in a Fox Business interview.    “I am in frequent contact with my counterpart in India. I have a great working relationship with him, but they still have a ways to go on this point,” Greer said. The comments signal a deal to lower duties on Indian goods is still a ways off. US and Indian officials have been in talks for months over an agreement to lower Trump’s 50% tariff. The president imposed the rate last year, arguing that India’s oil purchases were fueling Russia’s war effort in Ukraine.  Discounted Russian crude has continued to make up a significant portion of Indian imports, a dynamic that analysts say may persist well into 2026. In the meantime, India and the European Union finalized a free-trade pact that was two decades in the making. The agreement was seen as a countermeasure to Trump’s aggressive trade policies.  “I think India comes out on top on this. Frankly, they have more market access into Europe. It sounds like they have some additional immigration rights,” Greer said Tuesday. “India is going to have a heyday with this. They have low-cost labor. And it looks like the EU is doubling down on globalization when we’re trying to fix some of the problems with globalization here in the US.” WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone.

Read More »

EIA Sees USA Diesel Price Dropping in 2026

In its latest short term energy outlook (STEO), which was released on January 13, the U.S. Energy Information Administration (EIA) projected that the U.S. on-highway diesel fuel retail price will drop in 2026. According to its latest STEO, the EIA sees the diesel price averaging $3.43 per gallon in 2026. In 2025, the U.S. on-highway diesel fuel retail price came in at $3.66 per gallon, the STEO showed. A quarterly breakdown included in the STEO projected that the U.S. diesel price will average $3.50 per gallon in the first quarter of 2026, $3.40 per gallon in the second quarter, and $3.41 per gallon across the third and fourth quarters of this year. The STEO showed that, in 2025, the U.S. diesel price came in at $3.63 per gallon in the first quarter, $3.55 per gallon in the second quarter, $3.76 per gallon in the third quarter, and $3.70 per MMBtu in the fourth quarter. In its latest diesel fuel update, which was released on January 27, the EIA showed a rising trend in the average U.S. on highway diesel fuel price. According to this fuel update, the U.S. on-highway diesel fuel price averaged $3.459 per gallon on January 12, $3.530 per gallon on January 19, and $3.624 per gallon on January 26. The January 26 price was $0.035 per gallon lower than the year ago price, however, the EIA fuel update showed. Of the five Petroleum Administration for Defense District (PADD) regions highlighted in the EIA’s latest fuel update, the West Coast was shown to have the highest U.S. on-highway diesel fuel price as of January 26, at $4.301 per gallon. The Gulf Coast was shown in the update to have the lowest U.S. on-highway diesel fuel price as of January 26, at $3.325 per gallon. A glossary section of the

Read More »

Chevron Adds GIP’s Thomas Horton to Board

Chevron Corp said Tuesday it has appointed Thomas Horton, a partner at Global Infrastructure Partners (GIP) of global asset manager BlackRock Inc, as an independent director, expanding its board to 13 members. Horton, 64, has joined Chevron’s Board Audit Committee, the Houston, Texas-based energy giant said in an online statement. “Horton previously held senior roles as chairman of American Airlines Group Inc, and chairman, CEO and president at American Airlines Inc and AMR Corp, where he successfully built American Airlines’ network both organically and through its combination with USAirways in 2015”, Chevron noted. Horton was also senior adviser at private equity investor Warburg Pincus. “In addition to executive management roles, Horton has served as a director with some of the Fortune 500’s top brands, including current seats on the boards of Walmart and General Electric (operating as GE Aerospace). He previously served on the boards of Qualcomm and Enlink Midstream”, Chevron added. Chair and chief executive Mike Wirth said of Horton, “His proven leadership, diverse board experience and thoughtful approach to governance will be invaluable as we continue to drive growth and create long-term value”. Chevron’s board now has 11 independent directors, according to the list of members on its website. Besides Wirth, the other non-independent is John Hess, who became a Chevron director July 2025 after Chevron acquired Hess Corp. In an earlier appointment, Chevron said November 3, 2025 that its assistant controller Amit Ghai will replace Alana Knowles as controller effective March 1, 2026. Knowles is expected to retire after 38 years with Chevron. “Ghai will lead Chevron’s accounting policy, corporate and external financial reporting, internal controls, global business services and digital finance teams”, Chevron said. Recently Wirth said he was in discussion with the board about his retirement. The 65-year-old has been chair of the board and CEO

Read More »

Mplify launches AI-focused Carrier Ethernet certifications

“We didn’t want to just put a different sticker on it,” Vachon said. “We wanted to give the opportunity for operators to recertify their infrastructure so at least you’ve now got this very competitive infrastructure.” Testing occurs on live production networks. The automated testing platform can be completed in days once technical preparation is finished. Organizations pay once per certification with predictable annual maintenance fees required to keep certifications active. Optional retesting can refresh certification test records. Carrier Ethernet for AI The Carrier Ethernet for AI certification takes the business certification baseline and adds a performance layer specifically designed for AI workloads. Rather than creating a separate track, the AI certification requires providers to first complete the Carrier Ethernet for Business validation, then demonstrate they can meet additional stringent requirements. “What we identified was that there was another tier that we could produce a standard around for AI,” Vachon explained. “With extensive technical discussions with our membership, our CTO, and our director of certification, they identified the critical performance and functionality parameters.” The additional validation focuses on three key performance parameters: frame delay, inter-frame delay variation, and frame loss ratio aligned with AI workload requirements. Testing uses MEF 91 test requirements with AI-specific traffic profiles and performance objectives that go beyond standard business service thresholds. The program targets three primary use cases: connecting subscriber premises running AI applications to AI edge sites, interconnecting AI edge sites to AI data centers, and AI data center to data center interconnections.

Read More »

Gauging the real impact of AI agents

That creates the primary network issue for AI agents, which is dealing with implicit and creeping data. There’s a singular important difference between an AI agent component and an ordinary software component. Software is explicit in its use of data. The programming includes data identification. AI is implicit in its data use; the model was trained on data, and there may well be some API linkage to databases that aren’t obvious to the user of the model. It’s also often true that when an agentic component is used, it’s determined that additional data resources are needed. Are all these resources in the same place? Probably not. The enterprises with the most experience with AI agents say it would be smart to expect some data center network upgrades to link agents to databases, and if the agents are distributed away from the data center, it may be necessary to improve the agent sites’ connection to the corporate VPN. As agents evolve into real-time applications, this requires they also be proximate to the real-time system they support (a factory or warehouse), so the data center, the users, and any real-time process pieces all pull at the source of hosting to optimize latency. Obviously, they can’t all be moved into one place, so the network has to make a broad and efficient set of connections. That efficiency demands QoS guarantees on latency as well as on availability. It’s in the area of availability, with a secondary focus on QoS attributes like latency, that the most agent-experienced enterprises see potential new service opportunities. Right now, these tend to exist within a fairly small circle—a plant, a campus, perhaps a city or town—but over time, key enterprises say that their new-service interest could span a metro area. They point out that the real-time edge applications

Read More »

Photonic chip vendor snags Gates investment

“Moore’s Law is slowing, but AI can’t afford to wait. Our breakthrough in photonics unlocks an entirely new dimension of scaling, by packing massive optical parallelism on a single chip,” said Patrick Bowen, CEO of Neurophos. “This physics-level shift means both efficiency and raw speed improve as we scale up, breaking free from the power walls that constrain traditional GPUs.” The new funding includes investments from Microsoft’s investment fund M12 that will help speed up delivery of Neurophos’ first integrated photonic compute system, including datacenter-ready OPU modules. Neurophos is not the only company exploring this field. Last April, Lightmatter announced the launch of photonic chips to tackle data center bottlenecks, And in 2024, IBM said its researchers were exploring optical chips and developing a prototype in this area.

Read More »

Intel wrestling with CPU supply shortage

“We have important customers in the data center side. We have important OEM customers on both data center and client and that needs to be our priority to get the limited supply we have to those customers,” he added. CEO Lip-Bu Tan added that the continuing proliferation and diversification of AI workloads is placing significant capacity constraints on traditional and new hardware infrastructure, reinforcing the growing and essential role CPUs play in the AI era. Because of this, Intel decided to simplify its server road map, focusing resources on the 16-channel Diamond Rapids product and accelerate the introduction of Coral Rapids. Intel had removed multithreading from diamond Rapids, presumably to get rid of the performance bottlenecks. With each core running two threads, they often competed for resources. That’s why, for example, Ampere does not use threading but instead applies many more cores per CPU. With Coral Rapids, Intel is not only reintroducing multi-threading back into our data center road map but working closely with Nvidia to build a custom Xeon fully integrated with their NVLink technology to Build the tighter connection between Intel Xeon processors and Nvidia GPUs. Another aspect impacting supply has been yields or the new 18A process node. Tan said he was disappointed that the company could not fully meet the demand of the markets, and that while yields are in line with internal plans, “they’re still below where I want them to be,” Tan said.  Tan said yields for 18A are improving month-over-month and Intel is targeting a 7% to 8% improvement each month.

Read More »

Intel’s AI pivot could make lower-end PCs scarce in 2026

However, he noted, “CPUs are not being cannibalized by GPUs. Instead, they have become ‘chokepoints’ in AI infrastructure.” For instance, CPUs such as Granite Rapids are essential in GPU clusters, and for handling agentic AI workloads and orchestrating distributed inference. How pricing might increase for enterprises Ultimately, rapid demand for higher-end offerings resulted in foundry shortages of Intel 10/7 nodes, Bickley noted, which represent the bulk of the company’s production volume. He pointed out that it can take up to three quarters for new server wafers to move through the fab process, so Intel will be “under the gun” until at least Q2 2026, when it projects an increase in chip production. Meanwhile, manufacturing capacity for Xeon is currently sold out for 2026, with varying lead times by distributor, while custom silicon programs are seeing lead times of 6 to 8 months, with some orders rolling into 2027, Bickley said. In the data center, memory is the key bottleneck, with expected price increases of more than 65% year over year in 2026 and up to 25% for NAND Flash, he noted. Some specific products have already seen price inflation of over 1,000% since 2025, and new greenfield capacity for memory is not expected until 2027 or 2028. Moor’s Sag was a little more optimistic, forecasting that, on the client side, “memory prices will probably stabilize this year until more capacity comes online in 2027.” How enterprises can prepare Supplier diversification is the best solution for enterprises right now, Sag noted. While it might make things more complex, it also allows data center operators to better absorb price shocks because they can rebalance against suppliers who have either planned better or have more resilient supply chains.

Read More »

Reports of SATA’s demise are overblown, but the technology is aging fast

The SATA 1.0 interface made its debut in 2003. It was developed by a consortium consisting of Intel, Dell, and storage vendors like Seagate and Maxtor. It quickly advanced to SATA III in 2009, but there never was a SATA IV. There was just nibbling around the edges with incremental updates as momentum and emphasis shifted to PCI Express and NVMe. So is there any life to be had in the venerable SATA interface? Surprisingly, yes, say the analysts. “At a high level, yes, SATA for consumer is pretty much a dead end, although if you’re storing TB of photos and videos, it is still the least expensive option,” said Bob O’Donnell, president and chief analyst with TECHnalysis Research. Similarly for enterprise, for massive storage demands, the 20 and 30 TB SATA drives from companies like Seagate and WD are apparently still in wide use in cloud data centers for things like cold storage. “In fact, both of those companies are seeing recording revenues based, in part, on the demand for these huge, high-capacity low-cost drives,” he said. “SATA doesn’t make much sense anymore. It underperforms NVMe significantly,” said Rob Enderle, principal analyst with The Enderle Group. “It really doesn’t make much sense to continue make it given Samsung allegedly makes three to four times more margin on NVMe.” And like O’Donnell, Enderle sees continued life for SATA-based high-capacity hard drives. “There will likely be legacy makers doing SATA for some time. IT doesn’t flip technology quickly and SATA drives do wear out, so there will likely be those producing legacy SATA products for some time,” he said.

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »