dotlah! dotlah!
  • Cities
  • Technology
  • Business
  • Politics
  • Society
  • Science
  • About
Social Links
  • zedreviews.com
  • citi.io
  • aster.cloud
  • liwaiwai.com
  • guzz.co.uk
  • atinatin.com
0 Likes
0 Followers
0 Subscribers
dotlah!
  • Cities
  • Technology
  • Business
  • Politics
  • Society
  • Science
  • About
  • Artificial Intelligence
  • Data
  • Research
  • Science

Generative AI Could Offer A Faster Way To Test Theories Of How The Universe Works

  • March 17, 2024
Total
0
Shares
0
0
0

IBM and CERN are using time-series transformers to model particle collisions at the world’s largest collider; it’s part of a larger goal to turn raw observational data into actionable insights.

Ordinary matter gives shape to everything we know of the world, but it represents just 5% of the universe. The rest is made up of mysterious particles that scientists have termed dark energy and dark matter.

An ongoing search for this missing stuff is taking place in tunnels below the Alps, at CERN’s Large Hadron Collider (LHC). Charged particles are smashed together at near light speed to uncover what the universe is made of and how it works. New particles are created in each collision, and as they interact with the LHC’s detectors, sub-particles are formed and measured.

The Higgs boson, involved in giving all other particles their mass, was discovered this way in 2012. But before collecting the physical proof, scientists ran extensive simulations to design their experiments, interpret the results, and test their hypotheses by comparing simulated outcomes against real-life observations.

CERN generates as much synthetic data from simulations as it does real-life experiments. But simulations are expensive, and only getting more so as the LHC and its detectors are upgraded to learn more about Higgs and improve the odds of finding new particles. Once finished, CERN’s real and surrogate experiments will generate substantially more data and consume substantially more computing power.

To help ease the crunch, IBM recently began working with CERN to apply its time series foundation models to modeling particle collisions. What large language models (LLMs) did for text analysis, IBM is hoping time-series transformers can do for prediction tasks based on real or synthetic data. If an AI model can learn a physical process from a sequence of measurements, rather than statistical calculations, it could pave the way for faster, more powerful predictions in just about any field.

Theoretical physics is IBM’s focus with CERN, but the project has broad relevance to any organization modeling the behavior of complex physical systems. Foundation models have the power to transform raw, high-resolution sensor data into a digital representation of reality that enterprises can mine for new ways of improving their products or operations.

“How can I adjust the manufacturing process to produce more paper or sugar,” said Jayant Kalagnanam, director of AI applications at IBM Research. “How often should I service a machine to prolong its lifespan? And what’s the best way to design an experiment to get the desired results? A foundation model that can ‘learn’ a physical process from observational data lets you ask the operational questions that previously were out of reach.”

Modeling physical processes from observations instead of statistical probabilities

Many organizations are drowning in sensor data. Billions to trillions of highly detailed observations could be turned into valuable insights if only we could see the forest for the trees.

Foundation models are now bringing this clarity to modeling the physical world as they did for modeling natural language and images. IBM was the first to apply transformers to raw time-series data with multiple variables in 2021. The work inspired a wave of similar models but that initial enthusiasm waned after a team from Hong Kong showed that a simple regression model could do better on a range of tasks.

The main flaw of those early time-series transformers was to treat measurements in time the same way as words, the team showed. When each time step is tokenized and fed to the transformer, local context gets lost. The resulting errors are compounded as the number of variables, and risk of meaningless correlations, grows.

This problem had earlier been addressed in adapting transformers to vision tasks like object classification and scene recognition. IBM researchers saw that those same ideas could be applied to time series forecasting.

IBM’s first innovation was to group consecutive time points into one token, just as neighboring pixels in an image are aggregated into “patches” to make the data easier for the transformer to digest. Segmenting and consolidating time steps in this way preserves more local context. It also cuts computation costs, speeding up training and freeing up memory to process more historical data.

IBM’s next breakthrough was to narrow the transformer’s sweeping attention mechanism. Rather than model the interaction of all variables through time, it was enough to just compute their interactions at each time step. These time-synchronized correlations could then be organized into a matrix, where the most meaningful relationships could be extracted.

“Think of it as a graph,” said IBM researcher Vijay Ekambaram. “It’s a map of how the variables are related and influence each other over time.” IBM’s efficient PatchTSMixer time-series model has been shown to outperform other forecasting models by up to 60% while using two to three times less memory.

Learning how sub-atomic particles behave

Predicting how high-energy particles will multiply in a collider is no easy task. Statistical calculations are needed to model the particles produced in the initial collision, and the shower of thousands to millions of secondary particles created in the collider’s detectors later. The fate of each particle must be calculated one by one.

The signal they leave in the detectors are used to infer a target particle’s identity, momentum, and other properties. Simulation software currently provides high-resolution estimates, but researchers hope that foundation models might simplify the computation involved and provide comparable results at least 100 times faster.

IBM’s work with CERN is focused on the part of the detector known as the calorimeter, which measures the position and energy of shower particles. Using synthetic data from past simulations, IBM trained its PatchTSMixer model to understand how a typical shower unfolds. Values plugged into the model included the angle and energy of the incoming particle, and the amount of energy deposited in the calorimeter by the shower of secondary particles.

When asked to replicate a shower with a set of desired parameters, PatchTSMixer can quickly give estimates of the final energy values. “Each shower event is different, with a randomly determined outcome,” said IBM researcher Kyongmin Yeo. “We’ve adapted our time series foundation model to simulate these random events by learning the probability distribution using a generative method.”

To be useful, the estimates must be fast as well as accurate. “The results so far look very promising,” said Anna Zaborowska, a physicist at CERN. “If we can speed up the simulation of single showers by 100 times that would be brilliant.”

What’s next

CERN has proposed building a $17 billion accelerator three times the size of the LHC and is currently using simulations to design it. If IBM’s fast simulations are successful, and the Future Circular Collider (FCC) is approved, IBM’s models could be used to design experiments in the new collider and interpret the results.

Fast simulations could also play a role in shrinking the LHC’s carbon footprint. “In principle, it could lead to substantial energy savings,” said Zaborowska.

Beyond CERN’s work to unlock the mysteries of the universe, time series foundation models hold the potential to optimize industrial processes in myriad ways to grow revenue and cut costs. Sensors now give enterprises a detailed view of the business, from each step on the assembly line to each power station on a multi-state grid.

Until transformers came along, mining this firehose of sensor data for insights was either impractical or impossible. IBM is now working with several companies in disparate industries to take their sensor data and build proxy models of the manufacturing process that can be interrogated for ways of increasing throughput or reducing energy use.

“These learned proxy models are also a good way to monitor the behavior of a machine or process by comparing observed outcomes against predictions to flag any anomalies,” said Kalagnanam.

In addition to working with enterprises to optimize their operation with time-series transformers, IBM has open-sourced its PatchTST and Patch TSMixer models on Hugging Face, where they’ve been downloaded several thousand times in the last two months.

By: Kim Martineau
Originally published at: IBM

Total
0
Shares
Share
Tweet
Share
Share
Related Topics
  • CERN
  • Generative AI
  • IBM
  • Physics
  • Science
  • Theoretical Physics
dotlah.com

Previous Article
Cloud
  • Engineering
  • People
  • Public Cloud

Why We Need Both Cloud Engineers And Cloud Architects

  • March 11, 2024
View Post
Next Article
  • Artificial Intelligence
  • People
  • Technology

AI vs. elections: 4 essential reads about the threat of high-tech deception in politics

  • March 18, 2024
View Post
You May Also Like
View Post
  • Cities
  • Climate Change
  • Science

New research may help scientists predict when a humid heat wave will break

  • dotlah.com
  • January 6, 2026
View Post
  • Artificial Intelligence
  • Technology

How AI can accelerate the energy transition, rather than compete with it

  • dotlah.com
  • November 19, 2025
View Post
  • Artificial Intelligence
  • Technology

Thoughts on America’s AI Action Plan

  • Dean Marc
  • July 24, 2025
View Post
  • Artificial Intelligence
  • Machine Learning
  • Technology

Apple supercharges its tools and technologies for developers to foster creativity, innovation, and design

  • Dean Marc
  • June 11, 2025
oracle-ibm
View Post
  • Artificial Intelligence
  • Technology

IBM and Oracle Expand Partnership to Advance Agentic AI and Hybrid Cloud

  • Dean Marc
  • May 6, 2025
View Post
  • Cities
  • Research

Mathematicians uncover the logic behind how people walk in crowds

  • dotlah.com
  • April 3, 2025
View Post
  • Artificial Intelligence
  • Technology

Tokyo Electron and IBM Renew Collaboration for Advanced Semiconductor Technology

  • Dean Marc
  • April 2, 2025
View Post
  • Artificial Intelligence
  • Technology

IBM contributes key open-source projects to Linux Foundation to advance AI community participation

  • dotlah.com
  • March 22, 2025


Trending
  • 1
    • Lah!
    • Technology
    Singapore And Apple Partner On National Health Initiative Using Apple Watch
    • September 16, 2020
  • 2
    • Technology
    Pratt And Whitney Announces First-in-MRO Application of 3D Printing For Aero-Engine Component
    • February 14, 2020
  • Valentine's Day 3
    • Gears
    Best Valentine’s Day Gifts That Will ‘Heart’ly Disappoint You.
    • January 30, 2025
  • Fashion. Fashion walk. 4
    • Cities
    • Features
    French Chic. An intro to Parisian Fashion.
    • May 10, 2024
  • USA flag 5
    • Featured
    • People
    Stars, Stripes, And Service. Exploring The Hierarchies And Heroes Of The U.S. Military
    • July 3, 2023
  • 6
    • Lah!
    • Society
    ​NTU Singapore Disburses S$1.5 Million In Financial Aid To Students Impacted By COVID-19
    • August 27, 2020
  • 7
    • Cities
    • Lah!
    A Fresh Take On Singapore’s Train System Map
    • May 4, 2019
  • nobel-prize-popular-physics-prize-2024-figure1 8
    • Artificial Intelligence
    • Featured
    • Features
    • Machine Learning
    • Software
    • Technology
    They Used Physics To Find Patterns In Information
    • October 9, 2024
  • 9
    • Cities
    • Society
    Our Home, Above All – RSAF50
    • April 26, 2018
  • times square new york 10
    • Economy
    • People
    The US Economy Is Reliant On Consumer Spending – Can It Survive A Pandemic?
    • September 29, 2020
  • 11
    • Technology
    Most Innovation Originates From Customers, Not Companies
    • September 27, 2019
  • 12
    • People
    • World Events
    How The Trans-Atlantic Slave Trade Uprooted Millions
    • June 29, 2020
Trending
  • 1
    New research may help scientists predict when a humid heat wave will break
    • January 6, 2026
  • 2
    This is what the new frontier of AI-powered financial inclusion looks like
    • January 2, 2026
  • 3
    How bus stops and bike lanes can make or break your festive city trip
    • December 29, 2025
  • 4
    Skills development is critical to bridging the global digital talent gap
    • December 22, 2025
  • Tech Not To Miss 5
    Zed Approves | 12 Cool Tech You’ll Regret Missing
    • December 21, 2025
  • zedreviews-12-gaming-holiday-deals-202512 6
    Zed Approves | 12 Gaming Upgrades You Actually Need This Holiday Season
    • December 17, 2025
  • zedreviews-amazon-uk-50-christmas-deals 7
    Zed Approves | The Amazon 50+ Holiday Gift Deals Worth Buying – UK Edition
    • December 14, 2025
  • Watches 8
    Zed Approves | 12 Watch Gifts for the Holiday Season
    • December 14, 2025
  • 6 Bags You Might Be Missing for Your Next Trip 9
    Zed Approves | 6 Bags You Might Be Missing for Your Next Trip
    • December 2, 2025
  • Zed Approves | 48 Highly Rated Black Friday Deals in 2025 10
    Zed Approves | 48 Highly Rated Black Friday Deals in 2025
    • November 28, 2025
Social Links
dotlah! dotlah!
  • Cities
  • Technology
  • Business
  • Politics
  • Society
  • Science
  • About
Connecting Dots Across Asia's Tech and Urban Landscape

Input your search keywords and press Enter.