Chipmakers Battling for Stake in High-Flying AI Chip Market

At the core of artificial intelligence (AI) is machine learning—a computer’s ability to use data (and lots of it) to learn and continuously improve its decision-making based on complex algorithms. In order for data centers, robots, drones, autonomous vehicles, and devices like digital assistants and smart phones to process these massive quantities of data, they require a key component: semiconductor chips. The result: the robotics and AI revolution has sparked a new battle for leadership among chip manufacturers that hope to claim a piece of the AI chip market that UBS has forecast to hit $35B by 2021, up from just $6B in 2016.[1] The race is on.

And yet, with so many players on the field, it’s increasingly clear that not just one semiconductor chip will dominate the computing landscape. While Nvidia has rapidly secured a dominant position in AI training applications, a multitude of companies from the US to China have entered the AI chip race, all pouring vast amounts of money into this transformative sector. US tech giants such as Google, Microsoft, Apple, Amazon, and Facebook are working furiously to increase processor speeds to support AI applications such as facial and speech recognition, search, and custom image recognition to improve personalization and robotic assistance.

Announced in 2016, Google’s Tensor Processing Units (TPUs) are its newest AI chips that deliver 15–30x higher performance and 30–80x higher performance-per-watt than standard CPUs and GPUs. TPUs improve the performance of Google’s Cloud-based services by enabling it to run state-of-the-art neural networks at scale—and at a much more affordable cost. These high-powered chips have also helped accelerate the development of some of Google’s most innovative applications, including Google Assistant, which recognizes voice commands on Androids, and Google Translate, which provides instant language translations.

Apple is hoping to set the stage with its new iPhone XS, featuring an A12 Bionic chip—the industry’s first ever 7-nanometer chip with a 6-core CPU and 4-core GPU, plus an updated neural engine. The iPhone’s apps will still run on the company’s machine learning framework, Core ML, but it will run 9x faster with the A12. Among traditional chipmakers, Nvidia currently dominates, followed by Intel, AMD, and Xilinx. China’s big three—Baidu, Alibaba, and Tencent—have all either released or are developing their own customized processors. Baidu recently announced its dedicated AI chip called Kunlun, while Alibaba and Tencent are deploying AI processing capabilities in their own cloud platforms. Meanwhile, start-ups such as Cerebras and Graphcore have been actively getting into the game, each having raised more than $100M in funding from leading VCs. Their mission: develop chips that can optimize and communicate with the rest of the system to enable AI applications in smart phones, autos, and other consumer devices.

Nvidia, a dominant player in GPUs with first-mover advantage, will likely sustain leadership in data center training and deep learning inference, especially with its new game-changing GPU platform. Based on the Turing architecture developed over the past 10 years, the company’s recently launched RTX platform combines tensor cores for AI inferencing with ray tracing capabilities to accelerate workloads. The new chip delivers 6x more performance than its predecessor Pascal. Nvidia claims that this new GPU architecture represents a fundamental shift in capabilities and could drive the entire industry towards a new mode of graphic rendering using ray tracing. The equity market has rewarded Nvidia’s market dominance and impressive growth with a nearly 10x rise in its share price in the past 3 years.

In the meantime, Intel has been left in the dust, with a declining share in the server market and conspicuous delays in reaching targets for next-generation chips. However, Intel is still in the AI race and a contender not to underestimate. The company currently offers compatible and comprehensive solutions ranging from CPUs, Nervana ASICs, FPGAs from its Altera acquisition, 3D-Xpoint, and Mobileye’s computer vision ASICs. These new offerings will allow the company to become more competitive against Nvidia and others as the industry shifts to inference and edge computing.

AMD, another strong contender in the space, is targeting Nvidia’s core market, datacenter AI training, with a new single solution set that includes the world’s first 7-nanometer GPU architecture based on the Vega platform. Its Radeon Instinct series is aiming to capture market share in image, video, and speech recognition, as well as natural language processing (NLP). In addition, AMD’s open-source software allows users to tap into its hardware while also being capable of using Nvidia processors.