Stay Ahead, Stay ONMINE

Industry Body Looks at March Texas Upstream Employment

According to the Texas Independent Producers and Royalty Owners Association’s (TIPRO) analysis, direct Texas upstream employment for March totaled 204,400. That’s what TIPRO said in a statement sent to Rigzone by the TIPRO team recently, which cited the latest Current Employment Statistics (CES) report from the U.S. Bureau of Labor Statistics (BLS). In the statement, TIPRO highlighted […]

According to the Texas Independent Producers and Royalty Owners Association’s (TIPRO) analysis, direct Texas upstream employment for March totaled 204,400.

That’s what TIPRO said in a statement sent to Rigzone by the TIPRO team recently, which cited the latest Current Employment Statistics (CES) report from the U.S. Bureau of Labor Statistics (BLS). In the statement, TIPRO highlighted that the March figure was “a decrease of 700 industry positions from February employment numbers, subject to revisions”. TIPRO noted in the statement that this represented a decline of 900 jobs in the services sector and an increase of 200 jobs in oil and gas extraction.

“TIPRO’s new workforce data still indicated strong job postings for the Texas oil and natural gas industry,” the organization said in its statement.

“According to the association, there were 10,120 active unique jobs postings for the Texas oil and natural gas industry last month, including 3,458 new postings,” it added.

“In comparison, the state of California had 2,777 unique job postings in March, followed by New York (2,892), Florida (1,781), and Colorado (1,438). TIPRO reported a total of 53,285 unique job postings nationwide last month within the oil and natural gas sector,” it continued.

In its statement, TIPRO noted that, among the 19 specific industry sectors it uses to define the Texas oil and natural gas industry, “Gasoline Stations with Convenience Stores led in the ranking for unique job listings in March with 2,806 postings, followed by Support Activities for Oil and Gas Operations (2,247), and Petroleum Refineries (820)”.

The leading three cities by total unique oil and natural gas job postings were Houston, with 2,212 postings, Midland, with 635 postings, and Odessa, with 412 postings, TIPRO highlighted in its statement. The top three companies ranked by unique job postings in March were Cefco, with 1,200, Love’s, with 726, and Energy Transfer, with 307, according to TIPRO.

“Of the top ten companies listed by unique job postings last month, five companies were in the services sector, two in the gasoline stations with convenience stores category, two midstream companies, and one oil and gas operator,” TIPRO said in its statement.

“Top posted industry occupations for March included first-line supervisors of retail sales workers (620), heavy and tractor-trailer truck drivers (434), and maintenance and repair workers (306),” it added.

“The top posted job titles for March included assistant store managers (257), customer service representatives (251), and maintenance people (159),” it went on to state.

TIPRO revealed in its statement that top qualifications for unique job postings included a valid driver’s license, with 1,661 postings, a commercial driver’s license, with 331 postings, and hazmat endorsement, with 196 postings.

“TIPRO reports that 44 percent of unique job postings had no education requirement listed, 28 percent required a bachelor’s degree, and 28 percent required a high school diploma or GED,” the industry body noted.

In its statement, TIPRO said there were 1,776 advertised salary observations with a median salary of $60,000 and added that the highest percentage of advertised salaries were in the $90,000 to $500,000 range.

“As a result of recent commodity price movement and significant market volatility, there are high uncertainties in outlooks for future energy supply, demand, and prices,” TIPRO warned in its statement.

“Based on current market conditions as of early April, analysts at the Energy Information Administration (EIA) are still projecting production growth this year and next in oil and natural gas output,” it added.

The EIA’s April STEO, which was released on April 10, projects that U.S. crude oil production, including lease condensate, will average 13.51 million barrels per day overall in 2025 and 13.56 million barrels per day overall in 2026.

The EIA expects the Lower 48 states, excluding the Gulf of America, to provide 11.28 million barrels per day of the 2025 total and 11.29 million barrels per day of the 2026 total, the STEO shows.

According to the STEO, the Permian region provides the largest slice of Lower 48 states, excluding the Gulf of America, production in both 2025 and 2026. This output is projected to come in at 6.59 million barrels per day in 2025 and 6.76 million barrels per day in 2026, the STEO outlined.

The Federal Gulf of America is expected to provide 1.80 million barrels of the 2025 total and 1.83 million barrels of the 2026 total, and Alaska is expected to provide 0.42 million barrels of this year’s total and 0.44 million barrels of next year’s total, the STEO highlighted.

TIPRO describes itself as a trade association representing the interests of nearly 3,000 independent oil and natural gas producers and royalty owners throughout Texas. 

To contact the author, email [email protected]

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Linkerd 2.18 advances cloud-native service mesh

The project’s focus has evolved significantly over the years. While early adoption centered on mutual TLS between pods, today’s enterprises are tackling much larger challenges. “For a long time, the most common pattern was simply, ‘I want to get mutual TLS between all my pods, which gives me encryption, and

Read More »

18 essential commands for new Linux users

[jdoe@fedora ~]$ ls -ld /home/jdoedrwx——. 1 jdoe jdoe 106 Apr 3 14:39 /home/jdoe As you may have suspected, “r” stands for read, “w” means write and “x” is for execute. Note that no permissions are available for other group members and anyone else on the system. Each user will be

Read More »

Grangemouth a ‘good example of transition done badly’ – Shanks

The closure of Scotland’s last oil refinery at Grangemouth is a “really good example of a transition done badly”, the UK energy minister has admitted. Speaking at an industry conference in London, Michael Shanks said he was “acutely aware that there is uncertainty and there is unease in the industry”. He said the situation at Grangemouth, which will see hundreds of jobs lost when refinery owner PetroIneos shutters the facility, was a problem his government “inherited”. Labour politicians in Scotland and Westminster have come under fire for failing to fulfil pledges to save jobs at the site. The UK and Scottish Governments have jointly drawn up a plan called “Project Willow” aimed at delivering a long-term industrial future for Grangemouth through investment in a number of energy and recycling schemes. Speaking to the audience on the second day of the North Sea Decarbonisation Conference, Shanks said the “problems” should have been addressed years earlier. “There is a kind of truth in government, that you sort of wish you could have dealt with some of the problems you inherit on day one, many, many years before,” he said. “The most acute example for me is Grangemouth, which is a really good example of a transition done badly. “You wish you could have tackled these things five, six years ago, when they first emerged. “You don’t get to choose your entry as a government minister, and that’s just the reality of it. But what we are seeking to do is grapple with the uncertainty and the challenge.” Shanks urged attendees at the conference to engage with the government’s current consultation on the North Sea, which is set to close 30 April. The minister said energy supply chain responses to its “Building the North Sea’s Energy Future: Consultation,” consultation will help the government

Read More »

USA Crude Oil Inventories Rise Week on Week

U.S. commercial crude oil inventories, excluding those in the Strategic Petroleum Reserve (SPR), increased by 0.2 million barrels from the week ending April 11 to the week ending April 18, the U.S. Energy Information Administration (EIA) highlighted in its latest weekly petroleum status report. This report was released on April 23 and included data for the week ending April 18. It showed that crude oil stocks, not including the SPR, stood at 443.1 million barrels on April 18, 442.9 million barrels on April 11, and 453.6 million barrels on April 19, 2024. Crude oil in the SPR stood at 397.5 million barrels on April 18, 397.0 million barrels on April 11, and 365.7 million barrels on April 19, 2024, the report outlined. Total petroleum stocks – including crude oil, total motor gasoline, fuel ethanol, kerosene type jet fuel, distillate fuel oil, residual fuel oil, propane/propylene, and other oils – stood at 1.605 billion barrels on April 18, the report showed. Total petroleum stocks were down 0.3 million barrels week on week and up 5.9 million barrels year on year, the report revealed. “At 443.1 million barrels, U.S. crude oil inventories are about five percent below the five year average for this time of year,” the EIA said in its latest weekly petroleum status report. “Total motor gasoline inventories decreased by 4.5 million barrels from last week and are about three percent below the five year average for this time of year. Finished gasoline inventories increased and blending components inventories decreased last week,” it added. “Distillate fuel inventories decreased by 2.4 million barrels last week and are about 13 percent below the five year average for this time of year. Propane/ propylene inventories increased by 2.3 million barrels from last week and are seven percent below the five year average for

Read More »

Lightshift to build 11-MW storage project for Virginia municipal utility

Danville Utilities, a municipal utility based in Danville, Virginia, is contracting for an 11-MW/44-MWh battery storage project to be built and owned by Lightshift Energy, the battery storage company said Wednesday. Danville Utilities plans to use the storage facility to reduce its peak demand, which it expects will lower its demand and transmission charges by about $30 million over the 20-year life of the project. About 40% of Danville Utilities’ power supply costs are demand-related and they are increasing by 10% to 15% a year, according to a Nov. 7 Danville City Council memo on the project, which Lightshift expects will be operating in the second quarter next year. The memo states that PJM Interconnection’s capacity rate for Danville Utilities is jumping to $8.29/kW-month starting June 1 from 96 cents/kW-month. An existing 10.6-MW storage project built by Lightshift has helped reduce the Danville Utilities’ transmission, capacity and congestion charges since it started operating in 2022, according to the memo from Lightshift, formerly Delorean Power. Lightshift expects the project will save the utility $40 million over its life. Lightshift will use the new battery system to sell ancillary services into the PJM market and arbitrage wholesale energy prices, according to the memo.  “The project will help lower transmission and capacity costs and allow the city’s electric rates to be competitive to other neighboring electric utilities when competing for economic development projects,” Jason Grey, Danville director of utilities, said in the press release. Danville Utilities has about 42,000 customers. The project in February received $1.5 million from the Virginia Tobacco Region Revitalization Commission’s Energy Ingenuity Fund. Municipal utilities and rural cooperatives are focused on cost, and energy storage can save them money while adding resilience to their systems, according to Laura Coriell, head of market development for Lightshift, which is based in

Read More »

Carbon capture technology is ready. Permitting needs to catch up.

Anna Littlefield is the program manager for low carbon energy technologies for the Payne Institute at the Colorado School of Mines. It’s no secret that carbon capture is gaining momentum — and the United States is leading the pack. U.S. capacity alone for carbon capture, utilization and storage, or CCUS, is expected to increase sevenfold by 2035. In 2023, announced carbon capture capacity by 2030 increased by 35%. In the U.S., there are already 14 operational commercial-scale facilities that can capture and store roughly 21.4 million metric tons of CO2 per year, equivalent to the annual emissions of 5 million passenger vehicles. And new projects continue to be announced. Since 2018 when Congress enhanced the 45Q tax credit, which supports CCUS, there have been more than 120 CCUS projects announced in the states. In 2023, Chevron announced it was expanding its Bayou Bend CCS project, positioning it to be one of the largest carbon storage projects in the country. Despite all this investment and decades of proven operations, critics continue to insist the technology doesn’t work — apparently because it hasn’t scaled fast enough. A Scientific American headline from 2017 asked, “Will Carbon Capture and Storage Ever Work?” One environmental group asserted the technology “won’t work,” while others call it a “pipe dream.” This narrative allows other critics to suggest any federal support for the technology is misallocated. Two U.S. lawmakers recently introduced the 45Q Repeal Act, a bill aimed at repealing the 45Q tax credit. Rep. Scott Perry, R-Pa., one of the sponsors of the bill, called CCUS an “inefficient and market-distorting technology.” Opponents of CCUS cheered the legislation. What these critics overlook is that the primary challenge facing CCUS is the regulatory uncertainty created by current federal policies. Fifteen years ago, the Task Force on Carbon Capture and Storage

Read More »

Miliband touts low-carbon power as ‘solution’ to energy security

Low-carbon power can play a “critical role” in delivering energy security around the world, according to UK energy secretary Ed Miliband. Speaking at the first global Future of Energy Summit held at Lancaster House, London on Thursday, Miliband touted renewable and low-carbon energy sources such as nuclear power as the primary “solution” to international energy crises, and a pathway to “abundance”. “For the UK, just to talk about us for a moment, there is an exciting vision of energy security and abundance from cheap, homegrown low-carbon power,” Miliband said. “Following Russia’s invasion of Ukraine, we saw family finances, business finances and public finances wrecked as fossil fuel prices rocketed on the global markets, and therefore, here in Britain.” Renewable energy presents a solution to energy security that was not possible a decade and a half ago, according to Miliband. He attributed this to declines in the cost of solar power, which has plunged globally by 90% since 2010, and offshore wind costs that have declined by over 60%. Last year, $2 trillion was invested in clean energy, double the total amount invested in fossil fuels, data from the International Energy Agency’s latest inventory showed. Meanwhile, 92.5% of new electricity demand was met by renewable and nuclear power globally, according to a separate report by the International Renewable Energy Agency (IRENA). Renewables in 2024: 5 Key Facts Behind a Record-Breaking Year The golden rule Miliband’s words were echoed by IEA executive director Fatih Birol who described homegrown energy as “the best friend of energy security”. Birol warned that, to manufacture clean technologies, the industry needs “critical minerals”. He said diversification is a golden rule, not just of energy sources but also of the countries one imports from, to achieve energy security. Birol also warned that unpredictable economies lead investors to “fly

Read More »

Eni Logs Lower Q1 Profit

Eni SpA on Thursday reported EUR 1.41 billion ($1.6 billion) in adjusted net income attributable to shareholders, down 11 percent from the same three-month period a year ago as oil prices weakened and oil and gas output fell. Before adjustment for nonrecurring or extraordinary items, net earnings landed at EUR 1.17 billion, down 3 percent year-on-year. The Italian state-backed integrated energy major saw liquid production in the January-March 2025 period drop 1 percent year-over-year to 786,000 barrels per day. Natural gas production declined 9 percent to 4.5 billion cubic feet a day. “The asset divestments closed in 2024 in Nigeria, Alaska, Congo and the mature fields declines were substantially offset by rampups at organic projects in Côte d’Ivoire, Congo, Mexico, and Italy”, Eni said in a report on its website. Realized liquid prices averaged $69.72 a barrel, down 6 percent against Q1 2024. Gas realizations rose 8 percent to $7.57 per thousand cubic feet. Upstream turnover totaled EUR 5.41 billion, down 4 percent year-on-year. Refining throughput decreased 8 percent year-on-year to 5.86 million metric tons. Sales of chemical products totaled 800,000 metric tons, down 7 percent. The refining and chemicals segment logged EUR 4.93 billion in sales, down 13 percent. “The refining business reported a proforma adjusted loss of EUR 91 mln, lower both y-o-y and sequentially due to a continuing deterioration in products crack spreads”, Eni said. “The chemicals business reported a loss of EUR 0.24 bln amidst a prolonged downturn of the European sector due to lower demand and margin pressure from cost-advantaged players”. Eni sold 2.8 billion cubic meters (98.88 billion cubic feet) of liquefied natural gas in Q1 2025, up 4 percent year-on-year. Total gas sales landed at 12.12 Bcm, down 22 percent. Eni’s “Global Gas and LNG Portfolio” segment generated EUR 5.59 billion in sales,

Read More »

Slowdown in AWS data center leasing plans poses little threat to CIOs

Oracle, according to Westfall, is committed to investing $10 billion in 2025 to build 100 new data centers and expand 66 existing ones, aiming to double its capacity this year. Likewise, Google is investing $75 billion in 2025 for data center construction, focusing on AI and cloud infrastructure, with projects such as a $600 million facility in Mesa, Arizona, and a $2 billion data center in Fort Wayne and Indiana underway, Westfall said. Meta, too, plans to spend up to $65 billion in 2025, a sizable bump up from $40 billion in 2024, primarily for data center expansion to support AI (Llama models, Meta AI) and metaverse workloads, Westfall added. However, these expansion plans will not result in the relatively smaller players catching up with AWS and Microsoft. “For smaller players like Google and Oracle, catching up with AWS and Microsoft would require historically large capital investments that likely aren’t justified by their current growth rates,” Alletto said.

Read More »

TSMC targets AI acceleration with A14 process and ‘System on Wafer-X’

Nvidia’s flagship GPUs currently integrate two chips, while its forthcoming Rubin Ultra platform will connect four. “The SoW-X delivers wafer-scale compute performance and significantly boosts speed by integrating multiple advanced compute SoC dies, stacked HBM memory, and optical interconnects into a single package,” said Neil Shah, partner and co-founder at Counterpoint Research. “This approach reduces latency, improves power efficiency, and enhances scalability compared to traditional multi-chip setups — giving enterprises and hyperscalers AI servers capable of handling future workloads faster, more efficiently, and in a smaller footprint.” This not only boosts capex savings in the long run but also opex savings in terms of energy and space. “Wafer-X technology isn’t just about bigger chips — it’s a signal that the future of AI infrastructure is being redesigned at the silicon level,” said Abhivyakti Sengar, practice director at Everest Group. “By tightly integrating compute, memory, and optical interconnects within a single wafer-scale package, TSMC targets the core constraints of AI: bandwidth and energy. For hyperscale data centers and frontier model training, this could be a game-changer.” Priorities for enterprise customers For enterprises investing in custom AI silicon, choosing the right foundry partner goes beyond performance benchmarks. It’s about finding a balance between cutting-edge capabilities, flexibility, and cost. “First, enterprise buyers need to assess manufacturing process technologies (such as TSMC’s 3nm, 2nm, or Intel’s 18A) to determine if they meet AI chip performance and power requirements, along with customization capabilities,” said Galen Zeng, senior research manager for semiconductor research at IDC Asia Pacific. “Second, buyers should evaluate advanced packaging abilities; TSMC leads in 3D packaging and customized packaging solutions, suitable for highly integrated AI chips, while Intel has advantages in x86 architecture. Finally, buyers should assess pricing structures.”

Read More »

Cloudbrink pushes SASE boundaries with 300 Gbps data center throughput

Those core components are functionally table stakes and don’t really serve to differentiate Cloudbrink against its myriad competitors in the SASE market. Where Cloudbrink looks to differentiate is at a technical level through a series of innovations including: Distributed edge architecture: The company has decoupled software from hardware, allowing their platform to run across 800 data centers by leveraging public clouds, telco networks and edge computing infrastructure. This approach reduces network latency from 300 milliseconds to between 7 and 20 milliseconds, the company says. This density dramatically improves TCP performance and responsiveness. Protocol optimization: Cloudbrink developed its own algorithms for SD-WAN optimization that bring enterprise-grade reliability to last mile links. These algorithms significantly improve efficiency on consumer broadband connections, enabling enterprise-grade performance over standard internet links. Integrated security stack: “We’ve been able to produce secure speeds at line rate on our platform by bringing security to the networking stack itself,” Mana noted. Rather than treating security as a separate overlay that degrades performance, Cloudbrink integrates security functions directly into the networking stack. The solution consists of three core components: client software for user devices, a cloud management plane, and optional data center connectors for accessing internal applications. The client intelligently connects to multiple edge nodes simultaneously, providing redundancy and application-specific routing optimization. Cloudbrink expands global reach Beyond its efforts to increase throughput, Cloudbrink is also growing its global footprint. Cloudbrink today announced a global expansion through new channel agreements and the opening of a Brazil office to serve emerging markets in Latin America, Korea and Africa. The expansion includes exclusive partnerships with WITHX in Korea, BAMM Technologies for Latin America distribution and OneTic for African markets. The company’s software-defined FAST (Flexible, Autonomous, Smart and Temporary) Edges technology enables rapid deployment of points of presence by leveraging existing infrastructure from multiple

Read More »

CIOs could improve sustainability with data center purchasing decisions — but don’t

CIOs can drive change Even though it’s difficult to calculate an organization’s carbon footprint, CIOs and IT purchasing leaders trying to reduce their environmental impact can influence data center operators, experts say. “Customers have a very large voice,” Seagate’s Feist says. “Don’t underestimate how powerful that CIO feedback loop is. The large cloud accounts are customer-obsessed organizations, so they listen, and they react.” While DataBank began using renewable energy years ago, customer demand can push more data center operators to follow suit, Gerson says. “For sure, if there is a requirement to purchase renewable power, we are going to purchase renewable power,” she adds.

Read More »

Copper-to-optics technology eyed for next-gen AI networking gear

Broadcom’s demonstration and a follow-up session explored the benefits of further developing CPC, such as reduced signal integrity penalties and extended reach, through channel modeling and simulations, Broadcom wrote in a blog about the DesignCon event. “Experimental results showed successful implementation of CPC, demonstrating its potential to address bandwidth and signal integrity challenges in data centers, which is crucial for AI applications,” Broadcom stated. In addition to the demo, Broadcom and Samtec also authored a white paper on CPC that stated: “Co-packaged connectivity (CPC) provides the opportunity to omit loss and reflection penalties from the [printed circuit board (PCB)] and the package. When high speed I/O is cabled from the top of the package advanced PCB materials are not necessary. Losses from package vertical paths and PCB routing can be transferred to the longer reach of cables,” the authors stated. “As highly complex systems are challenged to scale the number of I/O and their reach, co- packaged connectivity presents opportunity. As we approach 224G-PAM4 [which uses optical techniques to support 224 Gigabits per second data rates per optical lane] and above, system loss and dominating noise sources necessitate the need to re-consider that which has been restricted in the back of the system architect’s mind for years: What if we attached to the package?” At OFC, Samtec demonstrated its Si-FlyHD co-packaged cable assemblies and Samtec FlyoverOctal Small Form-factor Pluggable (OSFP) over the Samtec Eye Speed Hyper Low Skew twinax copper cable. Flyover is Samtec’s proprietary way of addressing signal integrity and reach limitations of routing high-speed signals through traditional printed circuit boards (PCBs). “This evaluation platform incorporates Broadcom’s industry-leading 200G SerDes technology and Samtec’s co-packaged Flyover technology. Si-Fly HD CPC offers the industry’s highest footprint density and robust interconnect which enables 102.4T (512 lanes at 200G) in a 95 x

Read More »

The Rise of AI Factories: Transforming Intelligence at Scale

AI Factories Redefine Infrastructure The architecture of AI factories reflects a paradigm shift that mirrors the evolution of the industrial age itself—from manual processes to automation, and now to autonomous intelligence. Nvidia’s framing of these systems as “factories” isn’t just branding; it’s a conceptual leap that positions AI infrastructure as the new production line. GPUs are the engines, data is the raw material, and the output isn’t a physical product, but predictive power at unprecedented scale. In this vision, compute capacity becomes a strategic asset, and the ability to iterate faster on AI models becomes a competitive differentiator, not just a technical milestone. This evolution also introduces a new calculus for data center investment. The cost-per-token of inference—how efficiently a system can produce usable AI output—emerges as a critical KPI, replacing traditional metrics like PUE or rack density as primary indicators of performance. That changes the game for developers, operators, and regulators alike. Just as cloud computing shifted the industry’s center of gravity over the past decade, the rise of AI factories is likely to redraw the map again—favoring locations with not only robust power and cooling, but with access to clean energy, proximity to data-rich ecosystems, and incentives that align with national digital strategies. The Economics of AI: Scaling Laws and Compute Demand At the heart of the AI factory model is a requirement for a deep understanding of the scaling laws that govern AI economics. Initially, the emphasis in AI revolved around pretraining large models, requiring massive amounts of compute, expert labor, and curated data. Over five years, pretraining compute needs have increased by a factor of 50 million. However, once a foundational model is trained, the downstream potential multiplies exponentially, while the compute required to utilize a fully trained model for standard inference is significantly less than

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »