It was that computing energy trickled down from hulking supercomputers to the chips in our pockets.
Over the previous 15 years, innovation has modified course: GPUs, born from gaming and scaled via accelerated computing, have surged upstream to remake supercomputing and carry the AI revolution to scientific computing’s most rarefied methods.

JUPITER at Forschungszentrum Jülich is the logo of this new period.
Not solely is it among the many best supercomputers — producing 63.3 gigaflops per watt — however it’s additionally a powerhouse for AI, delivering 116 AI exaflops, up from 92 at ISC Excessive Efficiency 2025.
That is the “flip” in motion. In 2019, almost 70% of the TOP100 high-performance computing methods have been CPU-only. Immediately, that quantity has plunged beneath 15%, with 88 of the TOP100 methods accelerated — and 80% of these powered by NVIDIA GPUs.
Throughout the broader TOP500, 388 methods, 78%, now use NVIDIA know-how, together with 218 GPU-accelerated methods (up 34 methods yr over yr) and 362 methods related by high-performance NVIDIA networking. The pattern is unmistakable: accelerated computing has change into the usual.
However the true revolution is in AI efficiency. With architectures like NVIDIA Hopper and Blackwell and methods like JUPITER, researchers now have entry to orders of magnitude extra AI compute than ever.
AI FLOPS have change into the brand new yardstick, enabling breakthroughs in local weather modeling, drug discovery and quantum simulation — issues that demand each scale and effectivity.
At SC16, years earlier than right this moment’s generative AI wave, NVIDIA founder and CEO Jensen Huang noticed what was coming. He predicted that AI would quickly reshape the world’s strongest computing methods.
“A number of years in the past, deep studying got here alongside, like Thor’s hammer falling from the sky, and gave us an extremely highly effective software to unravel a number of the most tough issues on the planet,” Huang declared.

The mathematics behind computing energy consumption had already made the shift to GPUs inevitable.
Nevertheless it was the AI revolution, ignited by the NVIDIA CUDA-X computing platform constructed on these GPUs, that prolonged the capabilities of those machines dramatically.
All of a sudden, supercomputers might ship significant science at double precision (FP64) in addition to at combined precision (FP32, FP16) and even at ultra-efficient codecs like INT8 and past — the spine of recent AI.
This flexibility allowed researchers to stretch energy budgets additional than ever to run bigger, extra advanced simulations and prepare deeper neural networks, all whereas maximizing efficiency per watt.
However even earlier than AI took maintain, the uncooked numbers had already compelled the problem. Energy budgets don’t negotiate. Supercomputer researchers — inside NVIDIA and throughout the group — have been coming to grips with the street forward, and it was paved with GPUs.
To achieve exascale with no Hoover Dam‑sized electrical invoice, researchers wanted acceleration. GPUs delivered much more operations per watt than CPUs. That was the pre‑AI inform of what was to come back, and that’s why when the AI increase hit, large-scale GPU methods already had momentum.
The seeds have been planted with Titan in 2012 on the Oak Ridge Nationwide Laboratory, one of many first main U.S. methods to pair CPUs with GPUs at unprecedented scale — exhibiting how hierarchical parallelism might unlock big utility positive aspects.
In Europe in 2013, Piz Daint set a brand new bar for each efficiency and effectivity, then proved the purpose the place it issues: actual functions like COSMO forecasting for climate prediction.
By 2017, the inflection was plain. Summit at Oak Ridge Nationwide Laboratory and Sierra at Lawrence Livermore Laboratory ushered in a brand new customary for management‑class methods: acceleration first. They didn’t simply run sooner; they modified the questions science might ask for local weather modeling, genomics, supplies and extra.
These methods are capable of do far more with a lot much less. On the Green500 listing of probably the most environment friendly methods, the highest eight are NVIDIA‑accelerated, with NVIDIA Quantum InfiniBand connecting 7 of the High 10.
However the story behind these headline numbers is how AI capabilities have change into the yardstick: JUPITER delivers 116 AI exaflops alongside 1 EF FP64 — a transparent sign of how science now blends simulation and AI.
Energy effectivity didn’t simply make exascale attainable; it made AI at exascale sensible. And as soon as science had AI at scale, the curve bent sharply upward.
What It Means Subsequent
This isn’t nearly benchmarks. It’s about actual science:
- Quicker, extra correct climate and local weather fashions
- Breakthroughs in drug discovery and genomics
- Simulations of fusion reactors and quantum methods
- New frontiers in AI-driven analysis throughout each self-discipline
The shift began as a power-efficiency crucial, grew to become an architectural benefit and has matured right into a scientific superpower: simulation and AI, collectively, at unprecedented scale.
It begins with scientific computing. Now, the remainder of computing will comply with.
