idw – Informationsdienst Wissenschaft

Nachrichten, Termine, Experten

Grafik: idw-Logo
Grafik: idw-Logo

idw - Informationsdienst
Wissenschaft

Science Video Project
idw-Abo

idw-News App:

AppStore

Google Play Store



Instanz:
Teilen: 
06.03.2025 15:34

Training artificial intelligence more efficiently

Julia Rinner Corporate Communications Center
Technische Universität München

    The enormous computing resources needed to train neural networks for artificial intelligence (AI) result in massive power consumption. Researchers at the Technical University of Munich (TUM) have developed a method that is 100 times faster and therefore much more energy efficient. Instead of taking an iterative approach, the parameters are computed directly based on probabilities. The results so far are comparable in quality to existing iterative methods.

    AI applications such as large language models (LLMs) have become an integral part of our everyday lives. The required computing, storage and transmission capacities are provided by data centers that consume vast amounts of energy. In Germany alone this amounted to around 16 billion kWh in 2020, or around 1% of the country’s total energy consumption. For 2025, this figure is expected to increase to 22 billion kWh.

    New method is 100 times faster with comparable accuracy

    The arrival of more complex AI applications in the coming years will substantially increase the demands on data center capacity. These applications will use up huge amounts of energy for the training of neural networks. To counteract this trend, researchers have developed a training method that is 100 times faster while achieving accuracy comparable to existing procedures. This will significantly reduce the energy consumption for training.

    The functioning of neural networks, which are used in AI for such tasks as image recognition or language processing, is inspired by way the human brain works. These networks consist of interconnected nodes called artificial neurons. The input signals are weighted with certain parameters and then summed up. If a defined threshold is exceeded, the signal is passed on to the next node. To train the network, the initial selection of parameter values is usually randomized, for example using a normal distribution. The values are then incrementally adjusted to gradually improve the network predictions. Because of the many iterations required, this training is extremely demanding and consumes a lot of electricity.

    Parameters selected according to probabilities

    Felix Dietrich, a professor of Physics-enhanced Machine Learning, and his team have developed a new method. Instead of iteratively determining the parameters between the nodes, their approach uses probabilities. Their probabilistic method is based on the targeted use of values at critical locations in the training data where large and rapid changes in values are taking place. The objective of the current study is to use this approach to acquire energy-conserving dynamic systems from the data. Such systems change over the course of time in accordance with certain rules and are found in climate models and in financial markets, for example.

    “Our method makes it possible to determine the required parameters with minimal computing power. This can make the training of neural networks much faster and, as a result, more energy efficient,” says Felix Dietrich. “In addition, we have seen that the accuracy of the new method is comparable to that of iteratively trained networks.”


    Wissenschaftliche Ansprechpartner:

    Prof. Dr. Felix Dietrich
    Physics-enhanced Machine Learning
    Technical University of Munich
    felix.dietrich@tum.de
    www.tum.de


    Originalpublikation:

    Rahma, Atamert, Chinmay Datar, and Felix Dietrich, “Training Hamiltonian Neural Networks without Backpropagation”, 2024. Machine Learning and the Physical Sciences
    Workshop at the 38th conference on Neural Information Processing Systems (NeurIPS)
    https://neurips.cc/virtual/2024/99994


    Weitere Informationen:

    https://www.tum.de/en/news-and-events/all-news/press-releases/details/new-method...


    Bilder

    Felix Dietrich, Professor of Physics-enhanced Machine Learning
    Felix Dietrich, Professor of Physics-enhanced Machine Learning
    Andreas Heddergott / TUM
    Andreas Heddergott / TUM; Free for use in reporting on TUM, with the copyright noted


    Merkmale dieser Pressemitteilung:
    Journalisten
    Informationstechnik
    überregional
    Forschungsergebnisse, Wissenschaftliche Publikationen
    Englisch


     

    Hilfe

    Die Suche / Erweiterte Suche im idw-Archiv
    Verknüpfungen

    Sie können Suchbegriffe mit und, oder und / oder nicht verknüpfen, z. B. Philo nicht logie.

    Klammern

    Verknüpfungen können Sie mit Klammern voneinander trennen, z. B. (Philo nicht logie) oder (Psycho und logie).

    Wortgruppen

    Zusammenhängende Worte werden als Wortgruppe gesucht, wenn Sie sie in Anführungsstriche setzen, z. B. „Bundesrepublik Deutschland“.

    Auswahlkriterien

    Die Erweiterte Suche können Sie auch nutzen, ohne Suchbegriffe einzugeben. Sie orientiert sich dann an den Kriterien, die Sie ausgewählt haben (z. B. nach dem Land oder dem Sachgebiet).

    Haben Sie in einer Kategorie kein Kriterium ausgewählt, wird die gesamte Kategorie durchsucht (z.B. alle Sachgebiete oder alle Länder).