Technology
Visualizing the Power of the World’s Supercomputers
Visualizing the Power of the World’s Supercomputers
A supercomputer is a machine that is built to handle billions, if not trillions of calculations at once. Each supercomputer is actually made up of many individual computers (known as nodes) that work together in parallel.
A common metric for measuring the performance of these machines is flops, or floating point operations per second.
In this visualization, we’ve used November 2021 data from TOP500 to visualize the computing power of the world’s top five supercomputers. For added context, a number of modern consumer devices were included in the comparison.
Ranking by Teraflops
Because supercomputers can achieve over one quadrillion flops, and consumer devices are much less powerful, we’ve used teraflops as our comparison metric.
1 teraflop = 1,000,000,000,000 (1 trillion) flops.
Rank | Name | Type | Teraflops |
---|---|---|---|
#1 | 🇯🇵 Supercomputer Fugaku | Supercomputer | 537,212 |
#2 | 🇺🇸 Summit | Supercomputer | 200,795 |
#3 | 🇺🇸 Sierra | Supercomputer | 125,712 |
#4 | 🇨🇳 Sunway Taihulight | Supercomputer | 125,436 |
#5 | 🇺🇸 Perlmutter | Supercomputer | 93,750 |
n/a | Nvidia Titan RTX | Consumer device | 130 |
n/a | Nvidia GeForce RTX 3090 | Consumer device | 36 |
n/a | Xbox Series X | Consumer device | 12 |
n/a | Tesla Model S (2021) | Consumer device | 10 |
Supercomputer Fugaku was completed in March 2021, and is officially the world’s most powerful supercomputer. It’s used for various applications, including weather simulations and innovative drug discovery.
Sunway Taihulight is officially China’s top supercomputer and fourth most powerful in the world. That said, some experts believe that the country is already operating two much more powerful systems, based on data from anonymous sources.
As you can see, the most advanced consumer devices do not come close to supercomputing power. For example, it would take the combined power of 4,000 Nvidia Titan RTX graphics cards (the most powerful consumer card available) to measure up to the Fugaku.
Upcoming Supercomputers
One of China’s unrevealed supercomputers is supposedly named Oceanlite, and is a successor to Sunway Taihulight. It’s believed to have reached 1.3 exaflops, or 1.3 quintillion flops. The following table makes it easier to follow all of these big numbers.
Name | Notation | Exponent | Prefix |
---|---|---|---|
Quintillion | 1,000,000,000,000,000,000 | 10^18 | Exa |
Quadrillion | 1,000,000,000,000,000 | 10^15 | Peta |
Trillion | 1,000,000,000,000 | 10^12 | Tera |
Billion | 1,000,000,000 | 10^9 | Giga |
Million | 1,000,000 | 10^6 | Mega |
In the U.S., rival chipmakers AMD and Intel have both won contracts from the U.S. Department of Energy to build exascale supercomputers. On the AMD side, there’s Frontier and El Capitan, while on the Intel side, there’s Aurora.
Also involved in the EL Capitan project is Hewlett Packard Enterprise (HPE), which claims the supercomputer will be able to reach 2 exaflops upon its completion in 2023. All of this power will be used to support several exciting endeavors:
- Enable advanced simulation and modeling to support the U.S. nuclear stockpile and ensure its reliability and security.
- Accelerate cancer drug discovery from six years to one year through a partnership with pharmaceutical company, GlaxoSmithKline
- Understand the dynamic and mutations of RAS proteins that are linked to 30% of human cancers
Altogether, exascale computing represents the ability to conduct complex analysis in a matter of seconds, rather than hours. This could unlock an even faster pace of innovation.
Technology
Charted: The Exponential Growth in AI Computation
In eight decades, artificial intelligence has moved from purview of science fiction to reality. Here’s a quick history of AI computation.

Charted: The Exponential Growth in AI Computation
Electronic computers had barely been around for a decade in the 1940s, before experiments with AI began. Now we have AI models that can write poetry and generate images from textual prompts. But what’s led to such exponential growth in such a short time?
This chart from Our World in Data tracks the history of AI through the amount of computation power used to train an AI model, using data from Epoch AI.
The Three Eras of AI Computation
In the 1950s, American mathematician Claude Shannon trained a robotic mouse called Theseus to navigate a maze and remember its course—the first apparent artificial learning of any kind.
Theseus was built on 40 floating point operations (FLOPs), a unit of measurement used to count the number of basic arithmetic operations (addition, subtraction, multiplication, or division) that a computer or processor can perform in one second.
Computation power, availability of training data, and algorithms are the three main ingredients to AI progress. And for the first few decades of AI advances, compute, which is the computational power needed to train an AI model, grew according to Moore’s Law.
Period | Era | Compute Doubling |
---|---|---|
1950–2010 | Pre-Deep Learning | 18–24 months |
2010–2016 | Deep Learning | 5–7 months |
2016–2022 | Large-scale models | 11 months |
Source: “Compute Trends Across Three Eras of Machine Learning” by Sevilla et. al, 2022.
However, at the start of the Deep Learning Era, heralded by AlexNet (an image recognition AI) in 2012, that doubling timeframe shortened considerably to six months, as researchers invested more in computation and processors.
With the emergence of AlphaGo in 2015—a computer program that beat a human professional Go player—researchers have identified a third era: that of the large-scale AI models whose computation needs dwarf all previous AI systems.
Predicting AI Computation Progress
Looking back at the only the last decade itself, compute has grown so tremendously it’s difficult to comprehend.
For example, the compute used to train Minerva, an AI which can solve complex math problems, is nearly 6 million times that which was used to train AlexNet 10 years ago.
Here’s a list of important AI models through history and the amount of compute used to train them.
AI | Year | FLOPs |
---|---|---|
Theseus | 1950 | 40 |
Perceptron Mark I | 1957–58 | 695,000 |
Neocognitron | 1980 | 228 million |
NetTalk | 1987 | 81 billion |
TD-Gammon | 1992 | 18 trillion |
NPLM | 2003 | 1.1 petaFLOPs |
AlexNet | 2012 | 470 petaFLOPs |
AlphaGo | 2016 | 1.9 million petaFLOPs |
GPT-3 | 2020 | 314 million petaFLOPs |
Minerva | 2022 | 2.7 billion petaFLOPs |
Note: One petaFLOP = one quadrillion FLOPs. Source: “Compute Trends Across Three Eras of Machine Learning” by Sevilla et. al, 2022.
The result of this growth in computation, along with the availability of massive data sets and better algorithms, has yielded a lot of AI progress in seemingly very little time. Now AI doesn’t just match, but also beats human performance in many areas.
It’s difficult to say if the same pace of computation growth will be maintained. Large-scale models require increasingly more compute power to train, and if computation doesn’t continue to ramp up it could slow down progress. Exhausting all the data currently available for training AI models could also impede the development and implementation of new models.
However with all the funding poured into AI recently, perhaps more breakthroughs are around the corner—like matching the computation power of the human brain.
Where Does This Data Come From?
Source: “Compute Trends Across Three Eras of Machine Learning” by Sevilla et. al, 2022.
Note: The time estimated to for computation to double can vary depending on different research attempts, including Amodei and Hernandez (2018) and Lyzhov (2021). This article is based on our source’s findings. Please see their full paper for further details. Furthermore, the authors are cognizant of the framing concerns with deeming an AI model “regular-sized” or “large-sized” and said further research is needed in the area.
Methodology: The authors of the paper used two methods to determine the amount of compute used to train AI Models: counting the number of operations and tracking GPU time. Both approaches have drawbacks, namely: a lack of transparency with training processes and severe complexity as ML models grow.
-
Misc3 weeks ago
Ranked: The World’s Largest Stadiums
-
Maps1 week ago
The Incredible Historical Map That Changed Cartography
-
China2 weeks ago
Charted: Six Red Flags Pointing to China’s Economy Slowing Down
-
VC+6 days ago
What’s New on VC+ in September
-
Maps4 weeks ago
Visualizing the BRICS Expansion in 4 Charts
-
Markets2 weeks ago
The 25 Best Stocks by Shareholder Wealth Creation (1926-2022)
-
Business3 days ago
Ranked: The 20 Best Franchises to Open in the U.S.
-
AI4 weeks ago
Nvidia vs. AMD vs. Intel: Comparing AI Chip Sales