Data-Driven Simulations

AI is on the cusp of disrupting simulations at industry-scale

My years in Amsterdam – first at the Amsterdam Machine Learning Lab and then 2 years at in the AI4Science lab at Microsoft Research – have shaped my research vision. I am firmly convinced that AI is on the cusp of disrupting simulations at industry-scale. Therefore, I have started a new group at JKU Linz which has strong computer vision, numerical simulation, and engineering components. We want to advance data-driven simulations at industry-scale, and place the Austrian industry engine Linz as a center for doing that.

About my research

Working in high energy physics means to make the invisible visible. When we published for example the first direct observation of the Higgs boson decaying into pairs of tau leptons non of the processes, i.e., the Higgs boson signal itself nor the many ‘‘background’’ processes, was measured directly. Everything needed to be reconstructed from terabytes of detector data, and – even more fascinating to me – by using sophisticated physical simulations.

My first contact with simulations in the context of Deep Learning was through the work on Boundary graph neural networks for 3D simulations, which was financed by the Austrian Research Promotion Agency (FFG) after winning a research grant together with DCS computing.

During my years in Amsterdam I fully pivoted towards neural modeling of partial differential equations (PDEs). For example:

How to disrupt simulations at industry-scale

Recent years have shown that scalable ideas, improving the datasets, and clever engineering are the ingredients for ever better Deep Learning systems. LLMs and large-scale computer vision models have given birth to the heavily overused term “foundation model”. After language and computer vision, foundation models have spread over to various disciplines, with AlphaFold as very prominent example.

I want to go one step further, and bring foundation models to the every-day engineering world, and in doing so, disrupt how our industry works. Every day thousands and thousands of compute hours are spent on turbulence modeling, simulations of fluid or air flows, heat transfer in materials, traffic flows, and many more. Many of these processes follow similar underlying patterns, but yet need different and extremely specialized softward to simulate. Even worse, for different parameter settings the costly simulations need to be run at full length from scratch.

This is what I want to change! Deep Learning techniques are ready to develope models which run simulations in seconds instead of days or even weeks. The hardware is ready to digest high-resolutional industry-scale inputs, e.g., 3D meshes or images, and subsequently sets the stage to train Deep Learning models at scale.

But what about the data you might ask? Don’t we need high-quality ground truth data, which usually must come from the very simulations we are going to replace, presenting us with a proverbial chicken-and-egg problem? Luckily, many of the above mentioned processes have common underlying dynamics – similar to how different languages share structure and grammar. Simulation data is abundant, we just need to use the right ones, and many of that.

Personally, I have experienced modern computer vision as one of the driving tools for developing data-driven simulations. But it will take much more. It will need expertise in numerical simulations, geometric and physics-informed deep learning, and engineering at scale. We have already put together a very strong interdisciplinary team with different expertise, and keep hiring. The journey is going to be exciting.

Stay stuned, first works will appear soon