Stay Ahead, Stay ONMINE

Nine Rules for SIMD Acceleration of Your Rust Code (Part 1)

Thanks to Ben Lichtman (B3NNY) at the Seattle Rust Meetup for pointing me in the right direction on SIMD. SIMD (Single Instruction, Multiple Data) operations have been a feature of Intel/AMD and ARM CPUs since the early 2000s. These operations enable you to, for example, add an array of eight i32 to another array of eight i32 with just one CPU operation on a single core. Using SIMD operations greatly speeds up certain tasks. If you’re not using SIMD, you may not be fully using your CPU’s capabilities. Is this “Yet Another Rust and SIMD” article? Yes and no. Yes, I did apply SIMD to a programming problem and then feel compelled to write an article about it. No, I hope that this article also goes into enough depth that it can guide you through your project. It explains the newly available SIMD capabilities and settings in Rust nightly. It includes a Rust SIMD cheatsheet. It shows how to make your SIMD code generic without leaving safe Rust. It gets you started with tools such as Godbolt and Criterion. Finally, it introduces new cargo commands that make the process easier. The range-set-blaze crate uses its RangeSetBlaze::from_iter method to ingest potentially long sequences of integers. When the integers are “clumpy”, it can do this 30 times faster than Rust’s standard HashSet::from_iter. Can we do even better if we use Simd operations? Yes! See this documentation for the definition of “clumpy”. Also, what happens if the integers are not clumpy? RangeSetBlaze is 2 to 3 times slower than HashSet. On clumpy integers, RangeSetBlaze::from_slice — a new method based on SIMD operations — is 7 times faster than RangeSetBlaze::from_iter. That makes it more than 200 times faster than HashSet::from_iter. (When the integers are not clumpy, it is still 2 to 3 times slower than HashSet.) Over the course of implementing this speed up, I learned nine rules that can help you accelerate your projects with SIMD operations. The rules are: Use nightly Rust and core::simd, Rust’s experimental standard SIMD module. CCC: Check, Control, and Choose your computer’s SIMD capabilities. Learn core::simd, but selectively. Brainstorm candidate algorithms. Use Godbolt and AI to understand your code’s assembly, even if you don’t know assembly language. Generalize to all types and LANES with in-lined generics, (and when that doesn’t work) macros, and (when that doesn’t work) traits. See Part 2 for these rules: 7. Use Criterion benchmarking to pick an algorithm and to discover that LANES should (almost) always be 32 or 64. 8. Integrate your best SIMD algorithm into your project with as_simd, special code for i128/u128, and additional in-context benchmarking. 9. Extricate your best SIMD algorithm from your project (for now) with an optional cargo feature. Aside: To avoid wishy-washiness, I call these “rules”, but they are, of course, just suggestions. Rule 1: Use nightly Rust and core::simd, Rust’s experimental standard SIMD module. Rust can access SIMD operations either via the stable core::arch module or via nighty’s core::simd module. Let’s compare them: core::arch core::simd Nightly Delightfully easy and portable. Limits downstream users to nightly. I decided to go with “easy”. If you decide to take the harder road, starting first with the easier path may still be worthwhile. In either case, before we try to use SIMD operations in a larger project, let’s make sure we can get them working at all. Here are the steps: First, create a project called simd_hello: cargo new simd_hello cd simd_hello Edit src/main.rs to contain (Rust playground): // Tell nightly Rust to enable ‘portable_simd’ #![feature(portable_simd)] use core::simd::prelude::*; // constant Simd structs const LANES: usize = 32; const THIRTEENS: Simd = Simd::::from_array([13; LANES]); const TWENTYSIXS: Simd = Simd::::from_array([26; LANES]); const ZEES: Simd = Simd::::from_array([b’Z’; LANES]); fn main() { // create a Simd struct from a slice of LANES bytes let mut data = Simd::::from_slice(b”URYYBJBEYQVQBUBCRVGFNYYTBVATJRYY”); data += THIRTEENS; // add 13 to each byte // compare each byte to ‘Z’, where the byte is greater than ‘Z’, subtract 26 let mask = data.simd_gt(ZEES); // compare each byte to ‘Z’ data = mask.select(data – TWENTYSIXS, data); let output = String::from_utf8_lossy(data.as_array()); assert_eq!(output, “HELLOWORLDIDOHOPEITSALLGOINGWELL”); println!(“{}”, output); } Next — full SIMD capabilities require the nightly version of Rust. Assuming you have Rust installed, install nightly (rustup install nightly). Make sure you have the latest nightly version (rustup update nightly). Finally, set this project to use nightly (rustup override set nightly). You can now run the program with cargo run. The program applies ROT13 decryption to 32 bytes of upper-case letters. With SIMD, the program can decrypt all 32 bytes simultaneously. Let’s look at each section of the program to see how it works. It starts with: #![feature(portable_simd)] use core::simd::prelude::*; Rust nightly offers its extra capabilities (or “features”) only on request. The #![feature(portable_simd)] statement requests that Rust nightly make available the new experimental core::simd module. The use statement then imports the module’s most important types and traits. In the code’s next section, we define useful constants: const LANES: usize = 32; const THIRTEENS: Simd = Simd::::from_array([13; LANES]); const TWENTYSIXS: Simd = Simd::::from_array([26; LANES]); const ZEES: Simd = Simd::::from_array([b’Z’; LANES]); The Simd struct is a special kind of Rust array. (It is, for example, always memory aligned.) The constant LANES tells the length of the Simd array. The from_array constructor copies a regular Rust array to create a Simd. In this case, because we want const Simd’s, the arrays we construct from must also be const. The next two lines copy our encrypted text into data and then adds 13 to each letter. let mut data = Simd::::from_slice(b”URYYBJBEYQVQBUBCRVGFNYYTBVATJRYY”); data += THIRTEENS; What if you make an error and your encrypted text isn’t exactly length LANES (32)? Sadly, the compiler won’t tell you. Instead, when you run the program, from_slice will panic. What if the encrypted text contains non-upper-case letters? In this example program, we’ll ignore that possibility. The += operator does element-wise addition between the Simd data and Simd THIRTEENS. It puts the result in data. Recall that debug builds of regular Rust addition check for overflows. Not so with SIMD. Rust defines SIMD arithmetic operators to always wrap. Values of type u8 wrap after 255. Coincidentally, Rot13 decryption also requires wrapping, but after ‘Z’ rather than after 255. Here is one approach to coding the needed Rot13 wrapping. It subtracts 26 from any values on beyond ‘Z’. let mask = data.simd_gt(ZEES); data = mask.select(data – TWENTYSIXS, data); This says to find the element-wise places beyond ‘Z’. Then, subtract 26 from all values. At the places of interest, use the subtracted values. At the other places, use the original values. Does subtracting from all values and then using only some seem wasteful? With SIMD, this takes no extra computer time and avoids jumps. This strategy is, thus, efficient and common. The program ends like so: let output = String::from_utf8_lossy(data.as_array()); assert_eq!(output, “HELLOWORLDIDOHOPEITSALLGOINGWELL”); println!(“{}”, output); Notice the .as_array() method. It safely transmutes a Simd struct into a regular Rust array without copying. Surprisingly to me, this program runs fine on computers without SIMD extensions. Rust nightly compiles the code to regular (non-SIMD) instructions. But we don’t just want to run “fine”, we want to run faster. That requires us to turn on our computer’s SIMD power. Rule 2: CCC: Check, Control, and Choose your computer’s SIMD capabilities. To make SIMD programs run faster on your machine, you must first discover which SIMD extensions your machine supports. If you have an Intel/AMD machine, you can use my simd-detect cargo command. Run with: rustup override set nightly cargo install cargo-simd-detect –force cargo simd-detect On my machine, it outputs: extension width available enabled sse2 128-bit/16-bytes true true avx2 256-bit/32-bytes true false avx512f 512-bit/64-bytes true false This says that my machine supports the sse2, avx2, and avx512f SIMD extensions. Of those, by default, Rust enables the ubiquitous twenty-year-old sse2 extension. The SIMD extensions form a hierarchy with avx512f above avx2 above sse2. Enabling a higher-level extension also enables the lower-level extensions. Most Intel/AMD computers also support the ten-year-old avx2 extension. You enable it by setting an environment variable: # For Windows Command Prompt set RUSTFLAGS=-C target-feature=+avx2 # For Unix-like shells (like Bash) export RUSTFLAGS=”-C target-feature=+avx2″ “Force install” and run simd-detect again and you should see that avx2 is enabled. # Force install every time to see changes to ‘enabled’ cargo install cargo-simd-detect –force cargo simd-detect extension width available enabled sse2 128-bit/16-bytes true true avx2 256-bit/32-bytes true true avx512f 512-bit/64-bytes true false Alternatively, you can turn on every SIMD extension that your machine supports: # For Windows Command Prompt set RUSTFLAGS=-C target-cpu=native # For Unix-like shells (like Bash) export RUSTFLAGS=”-C target-cpu=native” On my machine this enables avx512f, a newer SIMD extension supported by some Intel computers and a few AMD computers. You can set SIMD extensions back to their default (sse2 on Intel/AMD) with: # For Windows Command Prompt set RUSTFLAGS= # For Unix-like shells (like Bash) unset RUSTFLAGS You may wonder why target-cpu=native isn’t Rust’s default. The problem is that binaries created using avx2 or avx512f won’t run on computers missing those SIMD extensions. So, if you are compiling only for your own use, use target-cpu=native. If, however, you are compiling for others, choose your SIMD extensions thoughtfully and let people know which SIMD extension level you are assuming. Happily, whatever level of SIMD extension you pick, Rust’s SIMD support is so flexible you can easily change your decision later. Let’s next learn details of programming with SIMD in Rust. Rule 3: Learn core::simd, but selectively. To build with Rust’s new core::simd module you should learn selected building blocks. Here is a cheatsheet with the structs, methods, etc., that I’ve found most useful. Each item includes a link to its documentation. Structs Simd – a special, aligned, fixed-length array of SimdElement. We refer to a position in the array and the element stored at that position as a “lane”. By default, we copy Simd structs rather than reference them. Mask – a special Boolean array showing inclusion/exclusion on a per-lane basis. SimdElements Floating-Point Types: f32, f64 Integer Types: i8, u8, i16, u16, i32, u32, i64, u64, isize, usize — but not i128, u128 Simd constructors Simd::from_array – creates a Simd struct by copying a fixed-length array. Simd::from_slice – creates a Simd struct by copying the first LANE elements of a slice. Simd::splat – replicates a single value across all lanes of a Simd struct. slice::as_simd – without copying, safely transmutes a regular slice into an aligned slice of Simd (plus unaligned leftovers). Simd conversion Simd::as_array – without copying, safely transmutes an Simd struct into a regular array reference. Simd methods and operators simd[i] – extract a value from a lane of a Simd. simd + simd – performs element-wise addition of two Simd structs. Also, supported -, *, /, %, remainder, bitwise-and, -or, xor, -not, -shift. simd += simd – adds another Simd struct to the current one, in place. Other operators supported, too. Simd::simd_gt – compares two Simd structs, returning a Mask indicating which elements of the first are greater than those of the second. Also, supported simd_lt, simd_le, simd_ge, simd_lt, simd_eq, simd_ne. Simd::rotate_elements_left – rotates the elements of a Simd struct to the left by a specified amount. Also, rotate_elements_right. simd_swizzle!(simd, indexes) – rearranges the elements of a Simd struct based on the specified const indexes. simd == simd – checks for equality between two Simd structs, returning a regular bool result. Simd::reduce_and – performs a bitwise AND reduction across all lanes of a Simd struct. Also, supported: reduce_or, reduce_xor, reduce_max, reduce_min, reduce_sum (but noreduce_eq). Mask methods and operators Mask::select – selects elements from two Simd struct based on a mask. Mask::all – tells if the mask is all true. Mask::any – tells if the mask contains any true. All about lanes Simd::LANES – a constant indicating the number of elements (lanes) in a Simd struct. SupportedLaneCount – tells the allowed values of LANES. Use by generics. simd.lanes – const method that tells a Simd struct’s number of lanes. Low-level alignment, offsets, etc. When possible, use to_simd instead. More, perhaps of interest With these building blocks at hand, it’s time to build something. Rule 4: Brainstorm candidate algorithms. What do you want to speed up? You won’t know ahead of time which SIMD approach (of any) will work best. You should, therefore, create many algorithms that you can then analyze (Rule 5) and benchmark (Rule 7). I wanted to speed up range-set-blaze, a crate for manipulating sets of “clumpy” integers. I hoped that creating is_consecutive, a function to detect blocks of consecutive integers, would be useful. Background: Crate range-set-blaze works on “clumpy” integers. “Clumpy”, here, means that the number of ranges needed to represent the data is small compared to the number of input integers. For example, these 1002 input integers 100, 101, …, 489, 499, 501, 502, …, 998, 999, 999, 100, 0 Ultimately become three Rust ranges: 0..=0, 100..=499, 501..=999. (Internally, the RangeSetBlaze struct represents a set of integers as a sorted list of disjoint ranges stored in a cache efficient BTreeMap.) Although the input integers are allowed to be unsorted and redundant, we expect them to often be “nice”. RangeSetBlaze’s from_iter constructor already exploits this expectation by grouping up adjacent integers. For example, from_iter first turns the 1002 input integers into four ranges 100..=499, 501..=999, 100..=100, 0..=0. with minimal, constant memory usage, independent of input size. It then sorts and merges these reduced ranges. I wondered if a new from_slice method could speed construction from array-like inputs by quickly finding (some) consecutive integers. For example, could it— with minimal, constant memory — turn the 1002 inputs integers into five Rust ranges: 100..=499, 501..=999, 999..=999, 100..=100, 0..=0. If so, from_iter could then quickly finish the processing. Let’s start by writing is_consecutive with regular Rust: pub const LANES: usize = 16; pub fn is_consecutive_regular(chunk: &[u32; LANES]) – > bool { for i in 1..LANES { if chunk[i – 1].checked_add(1) != Some(chunk[i]) { return false; } } true } The algorithm just loops through the array sequentially, checking that each value is one more than its predecessor. It also avoids overflow. Looping over the items seemed so easy, I wasn’t sure if SIMD could do any better. Here was my first attempt: Splat0 use std::simd::prelude::*; const COMPARISON_VALUE_SPLAT0: Simd = Simd::from_array([15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0]); pub fn is_consecutive_splat0(chunk: Simd) – > bool { if chunk[0].overflowing_add(LANES as u32 – 1) != (chunk[LANES – 1], false) { return false; } let added = chunk + COMPARISON_VALUE_SPLAT0; Simd::splat(added[0]) == added } Here is an outline of its calculations: Source: This and all following images by author. It first (needlessly) checks that the first and last items are 15 apart. It then creates added by adding 15 to the 0th item, 14 to the next, etc. Finally, to see if all items in added are the same, it creates a new Simd based on added’s 0th item and then compares. Recall that splat creates a Simd struct from one value. Splat1 & Splat2 When I mentioned the is_consecutive problem to Ben Lichtman, he independently came up with this, Splat1: const COMPARISON_VALUE_SPLAT1: Simd = Simd::from_array([0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15]); pub fn is_consecutive_splat1(chunk: Simd) – > bool { let subtracted = chunk – COMPARISON_VALUE_SPLAT1; Simd::splat(chunk[0]) == subtracted } Splat1 subtracts the comparison value from chunk and checks if the result is the same as the first element of chunk, splatted. He also came up with a variation called Splat2 that splats the first element of subtracted rather than chunk. That would seemingly avoid one memory access. I’m sure you are wondering which of these is best, but before we discuss that let’s look at two more candidates. Swizzle Swizzle is like Splat2 but uses simd_swizzle! instead of splat. Macro simd_swizzle! creates a new Simd by rearranging the lanes of an old Simd according to an array of indexes. pub fn is_consecutive_sizzle(chunk: Simd) – > bool { let subtracted = chunk – COMPARISON_VALUE_SPLAT1; simd_swizzle!(subtracted, [0; LANES]) == subtracted } Rotate This one is different. I had high hopes for it. const COMPARISON_VALUE_ROTATE: Simd = Simd::from_array([4294967281, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]); pub fn is_consecutive_rotate(chunk: Simd) – > bool { let rotated = chunk.rotate_elements_right::(); chunk – rotated == COMPARISON_VALUE_ROTATE } The idea is to rotate all the elements one to the right. We then subtract the original chunk from rotated. If the input is consecutive, the result should be “-15” followed by all 1’s. (Using wrapped subtraction, -15 is 4294967281u32.) Now that we have candidates, let’s start to evaluate them. Rule 5: Use Godbolt and AI to understand your code’s assembly, even if you don’t know assembly language. We’ll evaluate the candidates in two ways. First, in this rule, we’ll look at the assembly language generated from our code. Second, in Rule 7, we’ll benchmark the code’s speed. Don’t worry if you don’t know assembly language, you can still get something out of looking at it. The easiest way to see the generated assembly language is with the Compiler Explorer, AKA Godbolt. It works best on short bits of code that don’t use outside crates. It looks like this: Referring to the numbers in the figure above, follow these steps to use Godbolt: Open godbolt.org with your web browser. Add a new source editor. Select Rust as your language. Paste in the code of interest. Make the functions of interest public (pub fn). Do not include a main or unneeded functions. The tool doesn’t support external crates. Add a new compiler. Set the compiler version to nightly. Set options (for now) to -C opt-level=3 -C target-feature=+avx512f. If there are errors, look at the output. If you want to share or save the state of the tool, click “Share” From the image above, you can see that Splat2 and Sizzle are exactly the same, so we can remove Sizzle from consideration. If you open up a copy of my Godbolt session, you’ll also see that most of the functions compile to about the same number of assembly operations. The exceptions are Regular — which is much longer — and Splat0 — which includes the early check. In the assembly, 512-bit registers start with ZMM. 256-bit registers start YMM. 128-bit registers start with XMM. If you want to better understand the generated assembly, use AI tools to generate annotations. For example, here I ask Bing Chat about Splat2: Try different compiler settings, including -C target-feature=+avx2 and then leaving target-feature completely off. Fewer assembly operations don’t necessarily mean faster speed. Looking at the assembly does, however, give us a sanity check that the compiler is at least trying to use SIMD operations, inlining const references, etc. Also, as with Splat1 and Swizzle, it can sometimes let us know when two candidates are the same. You may need disassembly features beyond what Godbolt offers, for example, the ability to work with code the uses external crates. B3NNY recommended the cargo tool cargo-show-asm to me. I tried it and found it reasonably easy to use. The range-set-blaze crate must handle integer types beyond u32. Moreover, we must pick a number of LANES, but we have no reason to think that 16 LANES is always best. To address these needs, in the next rule we’ll generalize the code. Rule 6: Generalize to all types and LANES with in-lined generics, (and when that doesn’t work) macros, and (when that doesn’t work) traits. Let’s first generalize Splat1 with generics. #[inline] pub fn is_consecutive_splat1_gen( chunk: Simd, comparison_value: Simd, ) – > bool where T: SimdElement + PartialEq, Simd: Sub, LaneCount: SupportedLaneCount, { let subtracted = chunk – comparison_value; Simd::splat(chunk[0]) == subtracted } First, note the #[inline] attribute. It’s important for efficiency and we’ll use it on pretty much every one of these small functions. The function defined above, is_consecutive_splat1_gen, looks great except that it needs a second input, called comparison_value, that we have yet to define. If you don’t need a generic const comparison_value, I envy you. You can skip to the next rule if you like. Likewise, if you are reading this in the future and creating a generic const comparison_value is as effortless as having your personal robot do your household chores, then I doubly envy you. We can try to create a comparison_value_splat_gen that is generic and const. Sadly, neither From nor alternative T::One are const, so this doesn’t work: // DOESN’T WORK BECAUSE From is not const pub const fn comparison_value_splat_gen() – > Simd where T: SimdElement + Default + From + AddAssign, LaneCount: SupportedLaneCount, { let mut arr: [T; N] = [T::from(0usize); N]; let mut i_usize = 0; while i_usize { #[inline] pub fn $function(chunk: Simd) – > bool where LaneCount: SupportedLaneCount, { define_comparison_value_splat!(comparison_value_splat, $type); let subtracted = chunk – comparison_value_splat(); Simd::splat(chunk[0]) == subtracted } }; } #[macro_export] macro_rules! define_comparison_value_splat { ($function:ident, $type:ty) = > { pub const fn $function() – > Simd where LaneCount: SupportedLaneCount, { let mut arr: [$type; N] = [0; N]; let mut i = 0; while i bool where Self: SimdElement, Simd: Sub, LaneCount: SupportedLaneCount; } macro_rules! impl_is_consecutive { ($type:ty) = > { impl IsConsecutive for $type { #[inline] // very important fn is_consecutive(chunk: Simd) – > bool where Self: SimdElement, Simd: Sub, LaneCount: SupportedLaneCount, { define_is_consecutive_splat1!(is_consecutive_splat1, $type); is_consecutive_splat1(chunk) } } }; } impl_is_consecutive!(i8); impl_is_consecutive!(i16); impl_is_consecutive!(i32); impl_is_consecutive!(i64); impl_is_consecutive!(isize); impl_is_consecutive!(u8); impl_is_consecutive!(u16); impl_is_consecutive!(u32); impl_is_consecutive!(u64); impl_is_consecutive!(usize); We can now call fully generic code (Rust Playground): // Works on i32 and 16 lanes let a: Simd = black_box(Simd::from_array(array::from_fn(|i| 100 + i as i32))); let ninety_nines: Simd = black_box(Simd::from_array([99; 16])); assert!(IsConsecutive::is_consecutive(a)); assert!(!IsConsecutive::is_consecutive(ninety_nines)); // Works on i8 and 64 lanes let a: Simd = black_box(Simd::from_array(array::from_fn(|i| 10 + i as i8))); let ninety_nines: Simd = black_box(Simd::from_array([99; 64])); assert!(IsConsecutive::is_consecutive(a)); assert!(!IsConsecutive::is_consecutive(ninety_nines)); With this technique, we can create multiple candidate algorithms that are fully generic over type and LANES. Next, it is time to benchmark and see which algorithms are fastest. Those are the first six rules for adding SIMD code to Rust. In Part 2, we look at rules 7 to 9. These rules will cover how to pick an algorithm and set LANES. Also, how to integrate SIMD operations into your existing code and (importantly) how to make it optional. Part 2 concludes with a discussion of when/if you should use SIMD and ideas for improving Rust’s SIMD experience. I hope to see you there. Please follow Carl on Medium. I write on scientific programming in Rust and Python, machine learning, and statistics. I tend to write about one article per month.

Thanks to Ben Lichtman (B3NNY) at the Seattle Rust Meetup for pointing me in the right direction on SIMD.

SIMD (Single Instruction, Multiple Data) operations have been a feature of Intel/AMD and ARM CPUs since the early 2000s. These operations enable you to, for example, add an array of eight i32 to another array of eight i32 with just one CPU operation on a single core. Using SIMD operations greatly speeds up certain tasks. If you’re not using SIMD, you may not be fully using your CPU’s capabilities.

Is this “Yet Another Rust and SIMD” article? Yes and no. Yes, I did apply SIMD to a programming problem and then feel compelled to write an article about it. No, I hope that this article also goes into enough depth that it can guide you through your project. It explains the newly available SIMD capabilities and settings in Rust nightly. It includes a Rust SIMD cheatsheet. It shows how to make your SIMD code generic without leaving safe Rust. It gets you started with tools such as Godbolt and Criterion. Finally, it introduces new cargo commands that make the process easier.


The range-set-blaze crate uses its RangeSetBlaze::from_iter method to ingest potentially long sequences of integers. When the integers are “clumpy”, it can do this 30 times faster than Rust’s standard HashSet::from_iter. Can we do even better if we use Simd operations? Yes!

See this documentation for the definition of “clumpy”. Also, what happens if the integers are not clumpy? RangeSetBlaze is 2 to 3 times slower than HashSet.

On clumpy integers, RangeSetBlaze::from_slice — a new method based on SIMD operations — is 7 times faster than RangeSetBlaze::from_iter. That makes it more than 200 times faster than HashSet::from_iter. (When the integers are not clumpy, it is still 2 to 3 times slower than HashSet.)

Over the course of implementing this speed up, I learned nine rules that can help you accelerate your projects with SIMD operations.

The rules are:

  1. Use nightly Rust and core::simd, Rust’s experimental standard SIMD module.
  2. CCC: Check, Control, and Choose your computer’s SIMD capabilities.
  3. Learn core::simd, but selectively.
  4. Brainstorm candidate algorithms.
  5. Use Godbolt and AI to understand your code’s assembly, even if you don’t know assembly language.
  6. Generalize to all types and LANES with in-lined generics, (and when that doesn’t work) macros, and (when that doesn’t work) traits.

See Part 2 for these rules:

7. Use Criterion benchmarking to pick an algorithm and to discover that LANES should (almost) always be 32 or 64.

8. Integrate your best SIMD algorithm into your project with as_simd, special code for i128/u128, and additional in-context benchmarking.

9. Extricate your best SIMD algorithm from your project (for now) with an optional cargo feature.

Aside: To avoid wishy-washiness, I call these “rules”, but they are, of course, just suggestions.

Rule 1: Use nightly Rust and core::simd, Rust’s experimental standard SIMD module.

Rust can access SIMD operations either via the stable core::arch module or via nighty’s core::simd module. Let’s compare them:

core::arch

core::simd

  • Nightly
  • Delightfully easy and portable.
  • Limits downstream users to nightly.

I decided to go with “easy”. If you decide to take the harder road, starting first with the easier path may still be worthwhile.


In either case, before we try to use SIMD operations in a larger project, let’s make sure we can get them working at all. Here are the steps:

First, create a project called simd_hello:

cargo new simd_hello
cd simd_hello

Edit src/main.rs to contain (Rust playground):

// Tell nightly Rust to enable 'portable_simd'
#![feature(portable_simd)]
use core::simd::prelude::*;

// constant Simd structs
const LANES: usize = 32;
const THIRTEENS: Simd = Simd::::from_array([13; LANES]);
const TWENTYSIXS: Simd = Simd::::from_array([26; LANES]);
const ZEES: Simd = Simd::::from_array([b'Z'; LANES]);

fn main() {
    // create a Simd struct from a slice of LANES bytes
    let mut data = Simd::::from_slice(b"URYYBJBEYQVQBUBCRVGFNYYTBVATJRYY");

    data += THIRTEENS; // add 13 to each byte

    // compare each byte to 'Z', where the byte is greater than 'Z', subtract 26
    let mask = data.simd_gt(ZEES); // compare each byte to 'Z'
    data = mask.select(data - TWENTYSIXS, data);

    let output = String::from_utf8_lossy(data.as_array());
    assert_eq!(output, "HELLOWORLDIDOHOPEITSALLGOINGWELL");
    println!("{}", output);
}

Next — full SIMD capabilities require the nightly version of Rust. Assuming you have Rust installed, install nightly (rustup install nightly). Make sure you have the latest nightly version (rustup update nightly). Finally, set this project to use nightly (rustup override set nightly).

You can now run the program with cargo run. The program applies ROT13 decryption to 32 bytes of upper-case letters. With SIMD, the program can decrypt all 32 bytes simultaneously.

Let’s look at each section of the program to see how it works. It starts with:

#![feature(portable_simd)]
use core::simd::prelude::*;

Rust nightly offers its extra capabilities (or “features”) only on request. The #![feature(portable_simd)] statement requests that Rust nightly make available the new experimental core::simd module. The use statement then imports the module’s most important types and traits.

In the code’s next section, we define useful constants:

const LANES: usize = 32;
const THIRTEENS: Simd = Simd::::from_array([13; LANES]);
const TWENTYSIXS: Simd = Simd::::from_array([26; LANES]);
const ZEES: Simd = Simd::::from_array([b'Z'; LANES]);

The Simd struct is a special kind of Rust array. (It is, for example, always memory aligned.) The constant LANES tells the length of the Simd array. The from_array constructor copies a regular Rust array to create a Simd. In this case, because we want const Simd’s, the arrays we construct from must also be const.

The next two lines copy our encrypted text into data and then adds 13 to each letter.

let mut data = Simd::::from_slice(b"URYYBJBEYQVQBUBCRVGFNYYTBVATJRYY");
data += THIRTEENS;

What if you make an error and your encrypted text isn’t exactly length LANES (32)? Sadly, the compiler won’t tell you. Instead, when you run the program, from_slice will panic. What if the encrypted text contains non-upper-case letters? In this example program, we’ll ignore that possibility.

The += operator does element-wise addition between the Simd data and Simd THIRTEENS. It puts the result in data. Recall that debug builds of regular Rust addition check for overflows. Not so with SIMD. Rust defines SIMD arithmetic operators to always wrap. Values of type u8 wrap after 255.

Coincidentally, Rot13 decryption also requires wrapping, but after ‘Z’ rather than after 255. Here is one approach to coding the needed Rot13 wrapping. It subtracts 26 from any values on beyond ‘Z’.

let mask = data.simd_gt(ZEES);
data = mask.select(data - TWENTYSIXS, data);

This says to find the element-wise places beyond ‘Z’. Then, subtract 26 from all values. At the places of interest, use the subtracted values. At the other places, use the original values. Does subtracting from all values and then using only some seem wasteful? With SIMD, this takes no extra computer time and avoids jumps. This strategy is, thus, efficient and common.

The program ends like so:

let output = String::from_utf8_lossy(data.as_array());
assert_eq!(output, "HELLOWORLDIDOHOPEITSALLGOINGWELL");
println!("{}", output);

Notice the .as_array() method. It safely transmutes a Simd struct into a regular Rust array without copying.

Surprisingly to me, this program runs fine on computers without SIMD extensions. Rust nightly compiles the code to regular (non-SIMD) instructions. But we don’t just want to run “fine”, we want to run faster. That requires us to turn on our computer’s SIMD power.

Rule 2: CCC: Check, Control, and Choose your computer’s SIMD capabilities.

To make SIMD programs run faster on your machine, you must first discover which SIMD extensions your machine supports. If you have an Intel/AMD machine, you can use my simd-detect cargo command.

Run with:

rustup override set nightly
cargo install cargo-simd-detect --force
cargo simd-detect

On my machine, it outputs:

extension       width                   available       enabled
sse2            128-bit/16-bytes        true            true
avx2            256-bit/32-bytes        true            false
avx512f         512-bit/64-bytes        true            false

This says that my machine supports the sse2avx2, and avx512f SIMD extensions. Of those, by default, Rust enables the ubiquitous twenty-year-old sse2 extension.

The SIMD extensions form a hierarchy with avx512f above avx2 above sse2. Enabling a higher-level extension also enables the lower-level extensions.

Most Intel/AMD computers also support the ten-year-old avx2 extension. You enable it by setting an environment variable:

# For Windows Command Prompt
set RUSTFLAGS=-C target-feature=+avx2

# For Unix-like shells (like Bash)
export RUSTFLAGS="-C target-feature=+avx2"

“Force install” and run simd-detect again and you should see that avx2 is enabled.

# Force install every time to see changes to 'enabled'
cargo install cargo-simd-detect --force
cargo simd-detect
extension         width                   available       enabled
sse2            128-bit/16-bytes        true            true
avx2            256-bit/32-bytes        true            true
avx512f         512-bit/64-bytes        true            false

Alternatively, you can turn on every SIMD extension that your machine supports:

# For Windows Command Prompt
set RUSTFLAGS=-C target-cpu=native

# For Unix-like shells (like Bash)
export RUSTFLAGS="-C target-cpu=native"

On my machine this enables avx512f, a newer SIMD extension supported by some Intel computers and a few AMD computers.

You can set SIMD extensions back to their default (sse2 on Intel/AMD) with:

# For Windows Command Prompt
set RUSTFLAGS=

# For Unix-like shells (like Bash)
unset RUSTFLAGS

You may wonder why target-cpu=native isn’t Rust’s default. The problem is that binaries created using avx2 or avx512f won’t run on computers missing those SIMD extensions. So, if you are compiling only for your own use, use target-cpu=native. If, however, you are compiling for others, choose your SIMD extensions thoughtfully and let people know which SIMD extension level you are assuming.

Happily, whatever level of SIMD extension you pick, Rust’s SIMD support is so flexible you can easily change your decision later. Let’s next learn details of programming with SIMD in Rust.

Rule 3: Learn core::simd, but selectively.

To build with Rust’s new core::simd module you should learn selected building blocks. Here is a cheatsheet with the structs, methods, etc., that I’ve found most useful. Each item includes a link to its documentation.

Structs

  • Simd – a special, aligned, fixed-length array of SimdElement. We refer to a position in the array and the element stored at that position as a “lane”. By default, we copy Simd structs rather than reference them.
  • Mask – a special Boolean array showing inclusion/exclusion on a per-lane basis.

SimdElements

  • Floating-Point Types: f32f64
  • Integer Types: i8u8i16u16i32u32i64u64isizeusize
  • — but not i128u128

Simd constructors

  • Simd::from_array – creates a Simd struct by copying a fixed-length array.
  • Simd::from_slice – creates a Simd struct by copying the first LANE elements of a slice.
  • Simd::splat – replicates a single value across all lanes of a Simd struct.
  • slice::as_simd – without copying, safely transmutes a regular slice into an aligned slice of Simd (plus unaligned leftovers).

Simd conversion

  • Simd::as_array – without copying, safely transmutes an Simd struct into a regular array reference.

Simd methods and operators

  • simd[i] – extract a value from a lane of a Simd.
  • simd + simd – performs element-wise addition of two Simd structs. Also, supported -*/%, remainder, bitwise-and, -or, xor, -not, -shift.
  • simd += simd – adds another Simd struct to the current one, in place. Other operators supported, too.
  • Simd::simd_gt – compares two Simd structs, returning a Mask indicating which elements of the first are greater than those of the second. Also, supported simd_ltsimd_lesimd_gesimd_ltsimd_eqsimd_ne.
  • Simd::rotate_elements_left – rotates the elements of a Simd struct to the left by a specified amount. Also, rotate_elements_right.
  • simd_swizzle!(simd, indexes) – rearranges the elements of a Simd struct based on the specified const indexes.
  • simd == simd – checks for equality between two Simd structs, returning a regular bool result.
  • Simd::reduce_and – performs a bitwise AND reduction across all lanes of a Simd struct. Also, supported: reduce_orreduce_xorreduce_maxreduce_minreduce_sum (but noreduce_eq).

Mask methods and operators

  • Mask::select – selects elements from two Simd struct based on a mask.
  • Mask::all – tells if the mask is all true.
  • Mask::any – tells if the mask contains any true.

All about lanes

  • Simd::LANES – a constant indicating the number of elements (lanes) in a Simd struct.
  • SupportedLaneCount – tells the allowed values of LANES. Use by generics.
  • simd.lanes – const method that tells a Simd struct’s number of lanes.

Low-level alignment, offsets, etc.

When possible, use to_simd instead.

More, perhaps of interest

With these building blocks at hand, it’s time to build something.

Rule 4: Brainstorm candidate algorithms.

What do you want to speed up? You won’t know ahead of time which SIMD approach (of any) will work best. You should, therefore, create many algorithms that you can then analyze (Rule 5) and benchmark (Rule 7).

I wanted to speed up range-set-blaze, a crate for manipulating sets of “clumpy” integers. I hoped that creating is_consecutive, a function to detect blocks of consecutive integers, would be useful.

Background: Crate range-set-blaze works on “clumpy” integers. “Clumpy”, here, means that the number of ranges needed to represent the data is small compared to the number of input integers. For example, these 1002 input integers

100, 101, …, 489, 499, 501, 502, …, 998, 999, 999, 100, 0

Ultimately become three Rust ranges:

0..=0, 100..=499, 501..=999.

(Internally, the RangeSetBlaze struct represents a set of integers as a sorted list of disjoint ranges stored in a cache efficient BTreeMap.)

Although the input integers are allowed to be unsorted and redundant, we expect them to often be “nice”. RangeSetBlaze’s from_iter constructor already exploits this expectation by grouping up adjacent integers. For example, from_iter first turns the 1002 input integers into four ranges

100..=499, 501..=999, 100..=100, 0..=0.

with minimal, constant memory usage, independent of input size. It then sorts and merges these reduced ranges.

I wondered if a new from_slice method could speed construction from array-like inputs by quickly finding (some) consecutive integers. For example, could it— with minimal, constant memory — turn the 1002 inputs integers into five Rust ranges:

100..=499, 501..=999, 999..=999, 100..=100, 0..=0.

If so, from_iter could then quickly finish the processing.

Let’s start by writing is_consecutive with regular Rust:

pub const LANES: usize = 16;
pub fn is_consecutive_regular(chunk: &[u32; LANES]) -> bool {
    for i in 1..LANES {
        if chunk[i - 1].checked_add(1) != Some(chunk[i]) {
            return false;
        }
    }
    true
}

The algorithm just loops through the array sequentially, checking that each value is one more than its predecessor. It also avoids overflow.

Looping over the items seemed so easy, I wasn’t sure if SIMD could do any better. Here was my first attempt:

Splat0

use std::simd::prelude::*;

const COMPARISON_VALUE_SPLAT0: Simd =
    Simd::from_array([15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0]);

pub fn is_consecutive_splat0(chunk: Simd) -> bool {
    if chunk[0].overflowing_add(LANES as u32 - 1) != (chunk[LANES - 1], false) {
        return false;
    }
    let added = chunk + COMPARISON_VALUE_SPLAT0;
    Simd::splat(added[0]) == added
}

Here is an outline of its calculations:

Source: This and all following images by author.

It first (needlessly) checks that the first and last items are 15 apart. It then creates added by adding 15 to the 0th item, 14 to the next, etc. Finally, to see if all items in added are the same, it creates a new Simd based on added’s 0th item and then compares. Recall that splat creates a Simd struct from one value.

Splat1 & Splat2

When I mentioned the is_consecutive problem to Ben Lichtman, he independently came up with this, Splat1:

const COMPARISON_VALUE_SPLAT1: Simd =
    Simd::from_array([0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15]);

pub fn is_consecutive_splat1(chunk: Simd) -> bool {
    let subtracted = chunk - COMPARISON_VALUE_SPLAT1;
    Simd::splat(chunk[0]) == subtracted
}

Splat1 subtracts the comparison value from chunk and checks if the result is the same as the first element of chunk, splatted.

He also came up with a variation called Splat2 that splats the first element of subtracted rather than chunk. That would seemingly avoid one memory access.

I’m sure you are wondering which of these is best, but before we discuss that let’s look at two more candidates.

Swizzle

Swizzle is like Splat2 but uses simd_swizzle! instead of splat. Macro simd_swizzle! creates a new Simd by rearranging the lanes of an old Simd according to an array of indexes.

pub fn is_consecutive_sizzle(chunk: Simd) -> bool {
    let subtracted = chunk - COMPARISON_VALUE_SPLAT1;
    simd_swizzle!(subtracted, [0; LANES]) == subtracted
}

Rotate

This one is different. I had high hopes for it.

const COMPARISON_VALUE_ROTATE: Simd =
    Simd::from_array([4294967281, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]);

pub fn is_consecutive_rotate(chunk: Simd) -> bool {
    let rotated = chunk.rotate_elements_right::();
    chunk - rotated == COMPARISON_VALUE_ROTATE
}

The idea is to rotate all the elements one to the right. We then subtract the original chunk from rotated. If the input is consecutive, the result should be “-15” followed by all 1’s. (Using wrapped subtraction, -15 is 4294967281u32.)

Now that we have candidates, let’s start to evaluate them.

Rule 5: Use Godbolt and AI to understand your code’s assembly, even if you don’t know assembly language.

We’ll evaluate the candidates in two ways. First, in this rule, we’ll look at the assembly language generated from our code. Second, in Rule 7, we’ll benchmark the code’s speed.

Don’t worry if you don’t know assembly language, you can still get something out of looking at it.

The easiest way to see the generated assembly language is with the Compiler Explorer, AKA Godbolt. It works best on short bits of code that don’t use outside crates. It looks like this:

Referring to the numbers in the figure above, follow these steps to use Godbolt:

  1. Open godbolt.org with your web browser.
  2. Add a new source editor.
  3. Select Rust as your language.
  4. Paste in the code of interest. Make the functions of interest public (pub fn). Do not include a main or unneeded functions. The tool doesn’t support external crates.
  5. Add a new compiler.
  6. Set the compiler version to nightly.
  7. Set options (for now) to -C opt-level=3 -C target-feature=+avx512f.
  8. If there are errors, look at the output.
  9. If you want to share or save the state of the tool, click “Share”

From the image above, you can see that Splat2 and Sizzle are exactly the same, so we can remove Sizzle from consideration. If you open up a copy of my Godbolt session, you’ll also see that most of the functions compile to about the same number of assembly operations. The exceptions are Regular — which is much longer — and Splat0 — which includes the early check.

In the assembly, 512-bit registers start with ZMM. 256-bit registers start YMM. 128-bit registers start with XMM. If you want to better understand the generated assembly, use AI tools to generate annotations. For example, here I ask Bing Chat about Splat2:

Try different compiler settings, including -C target-feature=+avx2 and then leaving target-feature completely off.

Fewer assembly operations don’t necessarily mean faster speed. Looking at the assembly does, however, give us a sanity check that the compiler is at least trying to use SIMD operations, inlining const references, etc. Also, as with Splat1 and Swizzle, it can sometimes let us know when two candidates are the same.

You may need disassembly features beyond what Godbolt offers, for example, the ability to work with code the uses external crates. B3NNY recommended the cargo tool cargo-show-asm to me. I tried it and found it reasonably easy to use.

The range-set-blaze crate must handle integer types beyond u32. Moreover, we must pick a number of LANES, but we have no reason to think that 16 LANES is always best. To address these needs, in the next rule we’ll generalize the code.

Rule 6: Generalize to all types and LANES with in-lined generics, (and when that doesn’t work) macros, and (when that doesn’t work) traits.

Let’s first generalize Splat1 with generics.

#[inline]
pub fn is_consecutive_splat1_gen(
    chunk: Simd,
    comparison_value: Simd,
) -> bool
where
    T: SimdElement + PartialEq,
    Simd: Sub, Output = Simd>,
    LaneCount: SupportedLaneCount,
{
    let subtracted = chunk - comparison_value;
    Simd::splat(chunk[0]) == subtracted
}

First, note the #[inline] attribute. It’s important for efficiency and we’ll use it on pretty much every one of these small functions.

The function defined above, is_consecutive_splat1_gen, looks great except that it needs a second input, called comparison_value, that we have yet to define.

If you don’t need a generic const comparison_value, I envy you. You can skip to the next rule if you like. Likewise, if you are reading this in the future and creating a generic const comparison_value is as effortless as having your personal robot do your household chores, then I doubly envy you.

We can try to create a comparison_value_splat_gen that is generic and const. Sadly, neither From nor alternative T::One are const, so this doesn’t work:

// DOESN'T WORK BECAUSE From is not const
pub const fn comparison_value_splat_gen() -> Simd
where
    T: SimdElement + Default + From + AddAssign,
    LaneCount: SupportedLaneCount,
{
    let mut arr: [T; N] = [T::from(0usize); N];
    let mut i_usize = 0;
    while i_usize < N {
        arr[i_usize] = T::from(i_usize);
        i_usize += 1;
    }
    Simd::from_array(arr)
}

Macros are the last refuge of scoundrels. So, let’s use macros:

#[macro_export]
macro_rules! define_is_consecutive_splat1 {
    ($function:ident, $type:ty) => {
        #[inline]
        pub fn $function(chunk: Simd) -> bool
        where
            LaneCount: SupportedLaneCount,
        {
            define_comparison_value_splat!(comparison_value_splat, $type);

            let subtracted = chunk - comparison_value_splat();
            Simd::splat(chunk[0]) == subtracted
        }
    };
}
#[macro_export]
macro_rules! define_comparison_value_splat {
    ($function:ident, $type:ty) => {
        pub const fn $function() -> Simd
        where
            LaneCount: SupportedLaneCount,
        {
            let mut arr: [$type; N] = [0; N];
            let mut i = 0;
            while i < N {
                arr[i] = i as $type;
                i += 1;
            }
            Simd::from_array(arr)
        }
    };
}

This lets us run on any particular element type and all number of LANES (Rust Playground):

define_is_consecutive_splat1!(is_consecutive_splat1_i32, i32);

let a: Simd = black_box(Simd::from_array(array::from_fn(|i| 100 + i as i32)));
let ninety_nines: Simd = black_box(Simd::from_array([99; 16]));
assert!(is_consecutive_splat1_i32(a));
assert!(!is_consecutive_splat1_i32(ninety_nines));

Sadly, this still isn’t enough for range-set-blaze. It needs to run on all element types (not just one) and (ideally) all LANES (not just one).

Happily, there’s a workaround, that again depends on macros. It also exploits the fact that we only need to support a finite list of types, namely: i8i16i32i64isizeu8u16u32u64, and usize. If you need to also (or instead) support f32 and f64, that’s fine.

If, on the other hand, you need to support i128 and u128, you may be out of luck. The core::simd module doesn’t support them. We’ll see in Rule 8 how range-set-blaze gets around that at a performance cost.

The workaround defines a new trait, here called IsConsecutive. We then use a macro (that calls a macro, that calls a macro) to implement the trait on the 10 types of interest.

pub trait IsConsecutive {
    fn is_consecutive(chunk: Simd) -> bool
    where
        Self: SimdElement,
        Simd: Sub, Output = Simd>,
        LaneCount: SupportedLaneCount;
}

macro_rules! impl_is_consecutive {
    ($type:ty) => {
        impl IsConsecutive for $type {
            #[inline] // very important
            fn is_consecutive(chunk: Simd) -> bool
            where
                Self: SimdElement,
                Simd: Sub, Output = Simd>,
                LaneCount: SupportedLaneCount,
            {
                define_is_consecutive_splat1!(is_consecutive_splat1, $type);
                is_consecutive_splat1(chunk)
            }
        }
    };
}

impl_is_consecutive!(i8);
impl_is_consecutive!(i16);
impl_is_consecutive!(i32);
impl_is_consecutive!(i64);
impl_is_consecutive!(isize);
impl_is_consecutive!(u8);
impl_is_consecutive!(u16);
impl_is_consecutive!(u32);
impl_is_consecutive!(u64);
impl_is_consecutive!(usize);

We can now call fully generic code (Rust Playground):

// Works on i32 and 16 lanes
let a: Simd = black_box(Simd::from_array(array::from_fn(|i| 100 + i as i32)));
let ninety_nines: Simd = black_box(Simd::from_array([99; 16]));

assert!(IsConsecutive::is_consecutive(a));
assert!(!IsConsecutive::is_consecutive(ninety_nines));

// Works on i8 and 64 lanes
let a: Simd = black_box(Simd::from_array(array::from_fn(|i| 10 + i as i8)));
let ninety_nines: Simd = black_box(Simd::from_array([99; 64]));

assert!(IsConsecutive::is_consecutive(a));
assert!(!IsConsecutive::is_consecutive(ninety_nines));

With this technique, we can create multiple candidate algorithms that are fully generic over type and LANES. Next, it is time to benchmark and see which algorithms are fastest.


Those are the first six rules for adding SIMD code to Rust. In Part 2, we look at rules 7 to 9. These rules will cover how to pick an algorithm and set LANES. Also, how to integrate SIMD operations into your existing code and (importantly) how to make it optional. Part 2 concludes with a discussion of when/if you should use SIMD and ideas for improving Rust’s SIMD experience. I hope to see you there.

Please follow Carl on Medium. I write on scientific programming in Rust and Python, machine learning, and statistics. I tend to write about one article per month.

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

AI agent traffic drives first profitable year for Fastly

Fetcher bots, which retrieve content in real time when users make queries to AI assistants, show different concentration patterns. OpenAI’s ChatGPT and related bots generated 68% of fetcher bot requests. In some cases, fetcher bot request volumes exceeded 39,000 requests per minute to individual sites. AI agents check multiple websites

Read More »

Naftogaz Seeks USA Funds to Renovate Destroyed Plants

Ukraine’s state-run oil and gas company Naftogaz Group is seeking funds to restore and renovate its facilities after the destruction caused by constant Russian attacks, said its top executive. At least €3 billion ($3.5 billion) of damage has been done to the country’s facilities, with equipment needs exceeding €900 million, according to the company.  Naftogaz is particularly interested in Ukraine’s ongoing talks with partners such as the US Exim Bank and the US International Development Finance Corp., Chief Executive Officer Sergii Koretskyi told Bloomberg News in an interview at his office in Kyiv. He also stressed the importance of European assistance. Some $250 million in unspent Ukraine assistance funds remain with the US State Department, he said — part of which could be used to purchase US-made gas compressor units to allow Kyiv to repair production facilities. Their use would also be a boon to American companies, he added.  “Now we need funding for imports, investments and technologies. This is definitely a win-win situation for all parties — we’re not saying ‘help us’ but offering mutually beneficial cooperation,” said Koretskyi. Naftogaz, which provides gas to 12.5 million households, is a key element of Ukraine’s energy sector. Its infrastructure, as well as that of other power companies, has come under intense Russian bombardment in recent weeks, depriving many civilians of heating amid freezing temperatures. Since the start of this year, Naftogaz infrastructure has already faced 20 strikes, damaging oil and gas production and transportation systems, Koretskyi said.  He said that last year was the most destructive for Ukraine’s energy sector since Russian President Vladimir Putin began his full-scale invasion nearly four years ago, with hundreds of missiles and drones hitting facilities. Last February and October were the hardest months for Naftogaz specifically, the CEO added. The company’s biggest challenge is the unpredictable consequences

Read More »

VEN Plans to Grant More Oil Blocks to Chevron and Repsol

Venezuela plans to grant more oil-production land to Chevron Corp. and Spain’s Repsol SA as the Trump administration pushes for private companies to rebuild the nation’s energy sector, according to people with knowledge of the matter. Officials in Caracas are poised to award the exploration and production blocks as soon as this week, the people said. Giving US and European companies more access to Venezuela’s oil-rich territory is a key piece of US President Donald Trump’s push to revive the nation’s dilapidated energy sector while eroding China and Russia’s local influence.  On Thursday, US Energy Secretary Chris Wright toured a project operated by Chevron in Venezuela’s Orinoco oil belt and told reporters that the opportunity for cooperation between the US and the South American nation is immense following the capture of former Venezuela President Nicolás Maduro.  In an interview with Bloomberg TV, Wright said the US would release additional licenses “soon,” with companies like Chevron seeing benefits from an increase of as much as a 30% in production in the next 18 to 24 months.  “Chevron is being enabled to massively grow their business here. They’re the largest producer in Venezuela today, and they’re going to be able to both expand the reserves they have and expand their operations,” Wright said. “They’re just one of many, but they’re going to be a big one,” he added. Repsol declined to comment. Chevron didn’t immediately respond to a request for comment. The Trump administration is expected to issue general license to allow international oil companies to explore and produce in Venezuela without violating US sanctions, Bloomberg reported earlier this month. It would be the latest is part of a string of authorizations from the Treasury Department to open up the nation’s oil sector since US forces captured Venezuela’s former President Nicolás Maduro on Jan.

Read More »

Oil Posts Second Straight Weekly Drop

Oil notched its first back-to-back weekly drop this year as traders weighed the prospect of expanded OPEC+ supplies against US-Iran nuclear talks and recent weakness in wider markets. West Texas Intermediate fell 1% for the week and ended the day little changed on Friday. President Donald Trump said the US deployed an additional aircraft carrier to the Middle East in case a nuclear deal is not reached with Iran. “If we don’t have a deal, we’ll need it,” Trump said at the White House. He added he thinks negotiations will ultimately be successful. Traders have been watching for any uptick in tensions between Washington and Tehran that could pose a threat to supply from the Middle East. The commodity was down earlier as OPEC+ members see scope for output increases to resume in April, believing concerns about a glut are overblown, delegates said. The group has not yet committed to any course of action or begun formal discussions for a March 1 meeting, they added. A second weekly decline in the futures market stands to snap a long run of gains for early 2026, when recurrent bouts of geopolitical tension including the US stand-off with Iran supported oil prices. At an energy conference in London this week, attendees flagged that they expect worldwide supplies to top demand this year, potentially feeding into higher inventories in the Atlantic basin, the region where global prices are set. Still, a pile-up of sanctioned oil coupled with supply disruptions in various nations has limited the impact thus far. Trading may be thinner ahead of the Presidents’ Day holiday in the US, contributing to exaggerated price swings. Oil Prices WTI for March delivery settled up 0.1% at $62.89 a barrel in New York. Brent for April settlement edged 0.3% higher to $67.75 a barrel. What

Read More »

Reliance Gets USA License to Directly Buy VEN Crude

Indian refiner Reliance Industries Ltd. has received a general license from the US government that will allow it to purchase Venezuelan oil directly, according to a person familiar with the matter.  Reliance, owned by billionaire Mukesh Ambani, applied for the permit last month and received it from the Treasury Department a few days ago, the person said, asking not to be named as the matter is not public. The move comes immediately on the heels of a trade deal with the US that slashes punitive tariffs for Indian exports but demands that the country stop importing discounted Russian oil. The Indian government has asked state-owned refiners to consider buying more Venezuelan crude, as well as oil from the US.  Venezuela is unlikely to produce large volumes of crude anytime soon, but even limited supplies provide a fallback option for India’s largest refiner. The US — which has stepped up involvement in Venezuela’s oil sector after capturing the country’s president last month — has been considering general licenses to permit purchases, trading and investment in a sprawling but threadbare industry. Reliance is the first Indian refiner to receive clearance in the current push.  Reliance has historically been an important consumer of the country’s heavy crude, having struck a term deal to secure as much as 400,000 barrels a day from Petroleos de Venezuela SA in 2012. It is among only a handful of refiners in India that have the capacity to process the high-viscosity, sour oil, which is difficult to extract and refine without diluent.  The Indian refining giant took about 25% of Venezuela’s exports in 2019, before its term deal got suspended in 2019 due to US sanctions. It last received a general license in 2024 and took crude until that expired last year, and was not renewed. Reuters first reported the issuance of

Read More »

Baker Hughes Explores $1.5B Sale of Waygate Unit

Baker Hughes Co. is exploring a potential sale of its Waygate Technologies unit, which provides industrial testing and inspection equipment, people with knowledge of the matter said.  The world’s second-biggest oilfield contractor is working with advisers to study a possible divestment of the Waygate business, which could fetch around $1.5 billion, according to the people. A sale process could kick off in the next few months and attract interest from private equity firms, the people said, asking not to be identified because the information is private.  Deliberations are ongoing and there’s no certainty they will lead to a transaction, the people said. A representative for Baker Hughes declined to comment.  Waygate, based in Hürth, Germany, makes radiographic testing systems, industrial CT scanners, remote visual inspection machines and ultrasonic testing devices. It operates in more than 80 countries and is known for brands including Krautkrämer, phoenix|x-ray, Seifert, Everest and Agfa NDT.  The company was started in 2004 as GE Inspection Technologies. It’s been under the current ownership since 2017, when General Electric Co. combined its oil and gas division with Baker Hughes in a $32 billion deal.  Baker Hughes is selling the non-core asset after agreeing last year to buy industrial equipment maker Chart Industries Inc. for about $9.6 billion in one of its biggest-ever acquisitions. Chief Executive Officer Lorenzo Simonelli said in October last year that Baker Hughes is undertaking a “comprehensive evaluation” of its capital allocation focus following the Chart deal in order to boost shareholder value.  The pending sale would join other sizeable corporate divestments in Europe. Volkswagen AG has launched the sale of a majority stake in its heavy diesel engine maker Everllence, while Continental AG is selling its Contitech business. WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions

Read More »

EIA Raises 2026 WTI Forecast, Lowers 2027 Projection

The U.S. Energy Information Administration (EIA) increased its 2026 West Texas Intermediate (WTI) crude oil average spot price forecast, and lowered its 2027 projection, in its latest short term energy outlook (STEO). According to the EIA’s February STEO, which was released on February 10, the EIA now sees the WTI spot price averaging $53.42 per barrel this year and $49.34 per barrel next year. In its previous STEO, which was released in January, the EIA projected that the WTI spot price would average $52.21 per barrel in 2026 and $50.36 per barrel in 2027. A quarterly breakdown included in the EIA’s latest STEO projected that the WTI average spot price will come in at $58.62 per barrel in the first quarter of this year, $53.65 per barrel in the second quarter, $51.69 per barrel in the third quarter, $50.00 per barrel in the fourth quarter, $49.00 per barrel in the first quarter of next year, $49.66 per barrel in the second quarter, $49.68 per barrel in the third quarter, and $49.00 per barrel in the fourth quarter of 2027. In its previous STEO, the EIA forecast that the WTI spot price would average $54.93 per barrel in the first quarter of this year, $52.67 per barrel in the second quarter, $52.03 per barrel in the third quarter, $49.34 per barrel in the fourth quarter, $49.00 per barrel in the first quarter of next year, $50.66 per barrel in the second quarter, $50.68 per barrel in the third quarter, and $51.00 per barrel in the fourth quarter of 2027. In a BMI report sent to Rigzone by the Fitch Group on Friday, BMI projected that the front month WTI crude price will average $64.00 per barrel in 2026 and $68.00 per barrel in 2027. Standard Chartered sees the NYMEX WTI nearby

Read More »

Arista laments ‘horrendous’ memory situation

Digging in on campus Arista has been clear about its plans to grow its presence campus networking environments. Last Fall, Ullal said she expects Arista’s campus and WAN business would grow from the current $750 million-$800 million run rate to $1.25 billion, representing a 60% growth opportunity for the company. “We are committed to our aggressive goal of $1.25 billion for ’26 for the cognitive campus and branch. We have also successfully deployed in many routing edge, core spine and peering use cases,” Ullal said. “In Q4 2025, Arista launched our flagship 7800 R4 spine for many routing use cases, including DCI, AI spines with that massive 460 terabits of capacity to meet the demanding needs of multiservice routing, AI workloads and switching use cases. The combined campus and routing adjacencies together contribute approximately 18% of revenue.” Ethernet leads the way “In terms of annual 2025 product lines, our core cloud, AI and data center products built upon our highly differentiated Arista EOS stack is successfully deployed across 10 gig to 800 gigabit Ethernet speeds with 1.6 terabit migration imminent,” Ullal said. “This includes our portfolio of EtherLink AI and our 7000 series platforms for best-in-class performance, power efficiency, high availability, automation, agility for both the front and back-end compute, storage and all of the interconnect zones.” Ullal said she expects Ethernet will get even more of a boost later this year when the multivendor Ethernet for Scale-Up Networking (ESUN) specification is released.  “We have consistently described that today’s configurations are mostly a combination of scale out and scale up were largely based on 800G and smaller ratings. Now that the ESUN specification is well underway, we need a good solid spec. Otherwise, we’ll be shipping proprietary products like some people in the world do today. And so we will tie our

Read More »

From NIMBY to YIMBY: A Playbook for Data Center Community Acceptance

Across many conversations at the start of this year, at PTC and other conferences alike, the word on everyone’s lips seems to be “community.” For the data center industry, that single word now captures a turning point from just a few short years ago: we are no longer a niche, back‑of‑house utility, but a front‑page presence in local politics, school board budgets, and town hall debates. That visibility is forcing a choice in how we tell our story—either accept a permanent NIMBY-reactive framework, or actively build a YIMBY narrative that portrays the real value digital infrastructure brings to the markets and surrounding communities that host it. Speaking regularly with Ilissa Miller, CEO of iMiller Public Relations about this topic, there is work to be done across the ecosystem to build communications. Miller recently reflected: “What we’re seeing in communities isn’t a rejection of digital infrastructure, it’s a rejection of uncertainty driven by anxiety and fear. Most local leaders have never been given a framework to evaluate digital infrastructure developments the way they evaluate roads, water systems, or industrial parks. When there’s no shared planning language, ‘no’ becomes the safest answer.” A Brief History of “No” Community pushback against data centers is no longer episodic; it has become organized, media‑savvy, and politically influential in key markets. In Northern Virginia, resident groups and environmental organizations have mobilized against large‑scale campuses, pressing counties like Loudoun and Prince William to tighten zoning, question incentives, and delay or reshape projects.1 Loudoun County’s move in 2025 to end by‑right approvals for new facilities, requiring public hearings and board votes, marked a watershed moment as the world’s densest data center market signaled that communities now expect more say over where and how these campuses are built. Prince William County’s decision to sharply increase its tax rate on

Read More »

Nomads at the Frontier: PTC 2026 Signals the Digital Infrastructure Industry’s Moment of Execution

Each January, the Pacific Telecommunications Council conference serves as a barometer for where digital infrastructure is headed next. And according to Nomad Futurist founders Nabeel Mahmood and Phillip Koblence, the message from PTC 2026 was unmistakable: The industry has moved beyond hype. The hard work has begun. In the latest episode of The DCF Show Podcast, part of our ongoing ‘Nomads at the Frontier’ series, Mahmood and Koblence joined Data Center Frontier to unpack the tone shift emerging across the AI and data center ecosystem. Attendance continues to grow year over year. Conversations remain energetic. But the character of those conversations has changed. As Mahmood put it: “The hype that the market started to see is actually resulting a bit more into actions now, and those conversations are resulting into some good progress.” The difference from prior years? Less speculation. More execution. From Data Center Cowboys to Real Deployments Koblence offered perhaps the sharpest contrast between PTC conversations in 2024 and those in 2026. Two years ago, many projects felt speculative. Today, developers are arriving with secured power, customers, and construction underway. “If 2024’s PTC was data center cowboys — sites that in someone’s mind could be a data center — this year was: show me the money, show me the power, give me accurate timelines.” In other words, the market is no longer rewarding hypothetical capacity. It is demanding delivered capacity. Operators now speak in terms of deployments already underway, not aspirational campuses still waiting on permits and power commitments. And behind nearly every conversation sits the same gating factor. Power. Power Has Become the Industry’s Defining Constraint Whether discussions centered on AI factories, investment capital, or campus expansion, Mahmood and Koblence noted that every conversation eventually returned to energy availability. “All of those questions are power,” Koblence said.

Read More »

Cooling Consolidation Hits AI Scale: LiquidStack, Submer, and the Future of Data Center Thermal Strategy

As AI infrastructure scales toward ever-higher rack densities and gigawatt-class campuses, cooling has moved from a technical subsystem to a defining strategic issue for the data center industry. A trio of announcements in early February highlights how rapidly the cooling and AI infrastructure stack is consolidating and evolving: Trane Technologies’ acquisition of LiquidStack; Submer’s acquisition of Radian Arc, extending its reach from core data centers into telco edge environments; and Submer’s partnership with Anant Raj to accelerate sovereign AI infrastructure deployment across India. Layered atop these developments is fresh guidance from Oracle Cloud Infrastructure explaining why closed-loop, direct-to-chip cooling is becoming central to next-generation facility design, particularly in regions where water use has become a flashpoint in community discussions around data center growth. Taken together, these developments show how the industry is moving beyond point solutions toward integrated, scalable AI infrastructure ecosystems, where cooling, compute, and deployment models must work together across hyperscale campuses and distributed edge environments alike. Trane Moves to Own the Cooling Stack The most consequential development comes from Trane Technologies, which on February 10 announced it has entered into a definitive agreement to acquire LiquidStack, one of the pioneers and leading innovators in data center liquid cooling. The acquisition significantly strengthens Trane’s ambition to become a full-service thermal partner for data center operators, extending its reach from plant-level systems all the way down to the chip itself. LiquidStack, headquartered in Carrollton, Texas, built its reputation on immersion cooling and advanced direct-to-chip liquid solutions supporting high-density deployments across hyperscale, enterprise, colocation, edge, and blockchain environments. Under Trane, those technologies will now be scaled globally and integrated into a broader thermal portfolio. In practical terms, Trane is positioning itself to deliver cooling across the full thermal chain, including: • Central plant equipment and chillers.• Heat rejection and controls

Read More »

Infrastructure Maturity Defines the Next Phase of AI Deployment

The State of Data Infrastructure Global Report 2025 from Hitachi Vantara arrives at a moment when the data center industry is undergoing one of the most profound structural shifts in its history. The transition from enterprise IT to AI-first infrastructure has moved from aspiration to inevitability, forcing operators, developers, and investors to confront uncomfortable truths about readiness, resilience, and risk. Although framed around “AI readiness,” the report ultimately tells an infrastructure story: one that maps directly onto how data centers are designed, operated, secured, and justified economically. Drawing on a global survey of more than 1,200 IT leaders, the report introduces a proprietary maturity model that evaluates organizations across six dimensions: scalability, reliability, security, governance, sovereignty, and sustainability. Respondents are then grouped into three categories—Emerging, Defined, and Optimized—revealing a stark conclusion: most organizations are not constrained by access to AI models or capital, but by the fragility of the infrastructure supporting their data pipelines. For the data center industry, the implications are immediate, shaping everything from availability design and automation strategies to sustainability planning and evolving customer expectations. In short, extracting value from AI now depends less on experimentation and more on the strength and resilience of the underlying infrastructure. The Focus of the Survey: Infrastructure, Not Algorithms Although the report is positioned as a study of AI readiness, its primary focus is not models, training approaches, or application development, but rather the infrastructure foundations required to operate AI reliably at scale. Drawing on responses from more than 1,200 organizations, Hitachi Vantara evaluates how enterprises are positioned to support production AI workloads across six dimensions as stated above: scalability, reliability, security, governance, sovereignty, and sustainability. These factors closely reflect the operational realities shaping modern data center design and management. The survey’s central argument is that AI success is no longer

Read More »

AI’s New Land Grab: Meta’s Indiana Megaproject and the Rise of Europe’s Neocloud Challengers

While Meta’s Indiana campus anchors hyperscale expansion in the United States, Europe recorded its own major infrastructure milestone this week as Amsterdam-based AI infrastructure provider Nebius unveiled plans for a 240-megawatt data center campus in Béthune, France, near Lille in the country’s northern industrial corridor. When completed, the campus will rank among Europe’s largest AI-focused data center facilities and positions northern France as a growing node in the continent’s expanding AI infrastructure map. The development repurposes a former Bridgestone tire manufacturing site, reflecting a broader trend across Europe in which legacy industrial properties, already equipped with heavy power access, transport links, and industrial zoning, are being converted into large-scale digital infrastructure hubs. Located within reach of connectivity and enterprise corridors linking Paris, Brussels, London, and Amsterdam, the site allows Nebius to serve major European markets while avoiding the congestion and power constraints increasingly shaping Tier 1 data center hubs. Industrial Infrastructure Becomes Digital Infrastructure Developers increasingly view former industrial sites as ideal for AI campuses because they often provide: • Existing grid interconnection capacity built for heavy industry• Transport and logistics infrastructure already in place• Industrial zoning that reduces permitting friction• Large contiguous parcels suited to phased campus expansion For regions like Hauts-de-France, redevelopment projects also offer economic transition opportunities, replacing legacy manufacturing capacity with next-generation digital infrastructure investment. Local officials have positioned the project as part of broader efforts to reposition northern France as a logistics and technology hub within Europe. The Neocloud Model Gains Ground Beyond the site itself, Nebius’ expansion illustrates the rapid emergence of neocloud infrastructure providers, companies building GPU-intensive AI capacity without operating full hyperscale cloud ecosystems. These firms increasingly occupy a strategic middle ground: supplying AI compute capacity to enterprises, startups, and even hyperscalers facing short-term infrastructure constraints. Nebius’ rise over the past year

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »