The World’s 100 Most Valuable Brands in 2018
According to Forbes, the world’s 100 most valuable brands are worth a staggering $2.15 trillion.
While that singular number is impressive, the publication’s 2018 rankings of global brands can be further broken down in other ways that are also quite intriguing. Let’s take a look at brands by individual brand value, as well as sorted by relevant industry.
Ranking the Most Valuable Brands in 2018
Today’s infographic comes to us from HowMuch.net and it showcases the 100 most valuable brands in the world, according to recent Forbes rankings.
Here are the brands with the most assessed value, along with their one-year change and industry.
|Rank||Brand||Brand Value ($B)||1-Year Change||Industry|
Apple remains the world’s most valuable brand at $182.8 billion, but there are four other tech companies hot on the iPhone maker’s heels – and each of them is growing brand value at a rapid pace.
Google (+30%), Microsoft (+21%), Facebook (+29%), and Amazon (31%) are all gaining at double-digit clips. At this point, each has lapped Coca-Cola, the highest ranked non-tech brand in the Top 10 at $57.3 billion.
Brands by Industry
The aforementioned top five brands are all focused on technology, but it’s important to recognize that this is also a part of a much wider trend.
Over the last decade, tech brands have gained consumer prominence to make the industry dominant both in terms of quantity of brands (20%) and total brand value (41%) on the Forbes 100 Most Valuable Brand list.
|Industry||# of Brands||Total Brand Value ($B)|
Only a handful of brands in consumer-facing industries like media, apparel, alcohol, and restaurants make the rankings.
Meanwhile, sectors that traditionally rely on heavy amounts of advertising – like consumer packaged goods and retail – have just 20 brands on the list between them. The highest ranked brand in either of those categories is Walmart at #26th with a brand value of $24.9 billion, which is about 1/3 of the brand value of online competitor Amazon.
Charted: The Exponential Growth in AI Computation
In eight decades, artificial intelligence has moved from purview of science fiction to reality. Here’s a quick history of AI computation.
Charted: The Exponential Growth in AI Computation
Electronic computers had barely been around for a decade in the 1940s, before experiments with AI began. Now we have AI models that can write poetry and generate images from textual prompts. But what’s led to such exponential growth in such a short time?
The Three Eras of AI Computation
In the 1950s, American mathematician Claude Shannon trained a robotic mouse called Theseus to navigate a maze and remember its course—the first apparent artificial learning of any kind.
Theseus was built on 40 floating point operations (FLOPs), a unit of measurement used to count the number of basic arithmetic operations (addition, subtraction, multiplication, or division) that a computer or processor can perform in one second.
Computation power, availability of training data, and algorithms are the three main ingredients to AI progress. And for the first few decades of AI advances, compute, which is the computational power needed to train an AI model, grew according to Moore’s Law.
|1950–2010||Pre-Deep Learning||18–24 months|
|2010–2016||Deep Learning||5–7 months|
|2016–2022||Large-scale models||11 months|
Source: “Compute Trends Across Three Eras of Machine Learning” by Sevilla et. al, 2022.
However, at the start of the Deep Learning Era, heralded by AlexNet (an image recognition AI) in 2012, that doubling timeframe shortened considerably to six months, as researchers invested more in computation and processors.
With the emergence of AlphaGo in 2015—a computer program that beat a human professional Go player—researchers have identified a third era: that of the large-scale AI models whose computation needs dwarf all previous AI systems.
Predicting AI Computation Progress
Looking back at the only the last decade itself, compute has grown so tremendously it’s difficult to comprehend.
For example, the compute used to train Minerva, an AI which can solve complex math problems, is nearly 6 million times that which was used to train AlexNet 10 years ago.
Here’s a list of important AI models through history and the amount of compute used to train them.
|Perceptron Mark I||1957–58||695,000|
|AlphaGo||2016||1.9 million petaFLOPs|
|GPT-3||2020||314 million petaFLOPs|
|Minerva||2022||2.7 billion petaFLOPs|
Note: One petaFLOP = one quadrillion FLOPs. Source: “Compute Trends Across Three Eras of Machine Learning” by Sevilla et. al, 2022.
The result of this growth in computation, along with the availability of massive data sets and better algorithms, has yielded a lot of AI progress in seemingly very little time. Now AI doesn’t just match, but also beats human performance in many areas.
It’s difficult to say if the same pace of computation growth will be maintained. Large-scale models require increasingly more compute power to train, and if computation doesn’t continue to ramp up it could slow down progress. Exhausting all the data currently available for training AI models could also impede the development and implementation of new models.
However with all the funding poured into AI recently, perhaps more breakthroughs are around the corner—like matching the computation power of the human brain.
AI4 weeks ago
AI vs. Humans: Which Performs Certain Skills Better?
Maps2 weeks ago
Mapped: The Deadliest Earthquakes of the 21st Century
Oil and Gas20 hours ago
Charted: The World’s Biggest Oil Producers
Inequality3 weeks ago
Visualizing the World’s Growing Millionaire Population (2012-2022)
Energy2 weeks ago
What Electricity Sources Power the World?
Wealth3 weeks ago
Mapped: The Richest Billionaires in U.S. States
Markets2 weeks ago
The 25 Worst Stocks by Shareholder Wealth Losses (1926-2022)
Jobs3 weeks ago
Charted: Youth Unemployment in the OECD and China