Hi,welcome
86-755-88844016 +852 2632 9637 6*12 hours online call
Artificial intelligence chips are exploding! Three major mainstream AI chips, leading companies are fully sorted out
2023-08-22

AI chips are the core component of AI server computing power, and with the rapid growth of AI computing power, there will be a greater demand for AI chips. According to the survey data, it is expected that the market size of China's artificial intelligence chips will reach 85 billion yuan in 2022, 103.9 billion yuan in 2023, and 178 billion yuan in 2025.


未标题-1


Overview of AI Chip Industry


AI chips, also known as AI accelerators or computing cards, are specifically designed for the field of artificial intelligence. Their architecture and instruction set are optimized for various algorithms and applications in the field of artificial intelligence, and can efficiently support intelligent processing tasks such as vision, speech, natural language processing, and traditional machine learning.


From the perspective of technical architecture, Al chips are mainly divided into three categories: GPU (graphics processor), FPGA (field programmable logic gate array), and ASIC (specialized integrated circuit).


Among them, GPU is a relatively mature universal artificial intelligence chip, while FPGA and ASIC are semi customized and fully customized chips tailored to the characteristics of artificial intelligence needs.


未标题-2


GPU


GPU is the abbreviation for graphics processor, which is a hardware device specifically used for processing high-performance computing such as graphics, videos, games, etc.


Compared to traditional central processing units (CPUs), GPUs have more computing cores and faster memory bandwidth, which can significantly improve computing efficiency and graphics rendering speed.


At present, with the release of models such as NVIDIA A100 and H100, GPUs have significant advantages in computing power compared to other hardware, and their work has gradually shifted from graphic processing to computing.


GPU, due to its strongest computing power and deep learning capabilities, has become the most suitable hardware for supporting artificial intelligence training and learning, and is currently the preferred accelerator chip in servers.


According to its purpose and performance, GPUs can be divided into two categories: professional cards and consumer grade cards. Professional cards are usually used for high-performance computing and large-scale data processing in fields such as engineering, science, and medicine, with major manufacturers including NVIDIA, AMD, etc; Consumer grade cards are mainly used for ordinary households and game players, with major manufacturers including NVIDIA, AMD, Intel, and others.


The relationship between the number of GPUs and total computing power (GPU is NVIDIA A100):


未标题-3


On August 31, 2022, GPU products produced by NVIDIA and AMD were included in the restricted range by the United States. NVIDIA's restricted products include A100 and H100, while AMD's regulated GPU products include the MI100 and MI200 series. Haiguang DCU belongs to a type of GPGPU. In typical application scenarios, the Haiguang Deep Computing No.1 indicator has reached the level of high-end products of the same type internationally. Against the backdrop of stricter overseas supervision, domestic GPU manufacturers represented by Haiguang are entering a golden development period.


AI chips are mainly based on GPUs, accounting for about 90% of the total market size:


未标题-4


FPGA


FPGA has incomparable advantages in terms of flexibility compared to processors such as ASICs and GPUs.


FPGA can completely eliminate the need to read DRAM, and the entire algorithm is completed on chip. For example, Shenjian Technology used FPGA to create an ESE model and ran it on different processors (CPU/GPU/FPGA). It was found that the FPGA had the shortest training time and the lowest energy consumption.


In terms of energy consumption, CPU Dense consumes 11W, CPU Sparse consumes 38W, and GPU Dense consumes 202W, which is the highest energy consumption scenario. GPU Spare consumes 136W, compared to FPGA which only requires 41W; In terms of training delay, FPGA takes 82.7 hours μ s. Far smaller than 6017.3 of CPU μ s. It is only one-third of the GPU training duration.


未标题-5


Source: ESE


FPGA not only has software programmability and flexibility, but also has hardware parallelism and low latency, which has great advantages in terms of market cycle and cost.


Therefore, in fields such as artificial intelligence and 5G communication where iterative upgrades are frequent, FPGA has become a relatively ideal solution.


未标题-6


According to industry research database data, the FPGA industry chain is relatively complex, and FPGA manufacturers have strong bargaining power over upstream and downstream enterprises. The FPGA industry includes upstream enterprises such as low-level algorithm design companies, wafer foundries, specialized materials and equipment suppliers, midstream enterprises such as FPGA chip manufacturers and packaging manufacturers, as well as downstream enterprises such as visual industry manufacturers, automotive manufacturers, communication service providers, and cloud data centers. The entire industrial chain is relatively complex, with well-known domestic and foreign enterprises such as IBM, TSMC, Intel, Huawei, and Tencent participating. Among them, the leading FPGA companies in the middle reaches of the industry chain occupy a monopolistic position in the global market, and have strong bargaining power over upstream software and hardware suppliers and downstream customer enterprises.


未标题-7


From the perspective of market structure, the global FPGA market is mainly occupied by Celestine and Altera, with current market share of 52% and 35% respectively; Next are Lattice and Microsemi, both with 5% share. Among Chinese FPGA manufacturers, Ziguang Guowei, Fudan Microelectronics, and Anlu Technology have a market share of over 15% in 2021. Benefiting from the accelerated promotion of localization, Chinese FPGA manufacturers will have enormous growth potential.


未标题-8


ASIC


If ARM's counterattack on the X86 architecture in the chip industry is due to the rise of mobile terminals, then under the AI wave, AI chips, especially ASICs dedicated to deep learning, are used to achieve leapfrog development in a point-to-point manner, which is a good opportunity to overtake in a curve.


In terms of competitive space, traditional CPU fields include Intel and Qualcomm, GPU fields include NVIDIA, FPGA fields include Xilinx and Altera, and only the ASIC field that is most customized with AI computing has not yet had an absolute monopoly leader.


未标题-9

In application scenarios, smartphones, wearable devices, and security front-end devices may all become leading destinations for the deployment of ASIC chips.


AI chips, especially ASIC chips, are particularly suitable for smart phones, smart security cameras, smart homes, drones and other intelligent terminals with low power consumption and small space due to their low power consumption and high efficiency characteristics. These fields may become the first areas for ASIC chips to expand.


未标题-10


At present, the mainstream ASICs in the market include TPU chips, NPU chips, VPU chips, and BPU chips, which are designed and produced by Google, Cambrian, Intel, and Horizon respectively.


Due to the long development cycle of ASIC, only large factories have the funds and strength to conduct research and development. Meanwhile, ASIC is a fully customized chip that operates with the highest efficiency in certain specific scenarios. Therefore, when the downstream market space is large enough in certain scenarios, mass production of ASIC chips can achieve rich profits.


ASIC market pattern and representative manufacturers:


未标题-11


Source: Compilation of public information


AI chip market pattern


According to Liftr Insights, currently in the forefront of AI technology development in the North American data center AI chip market, Nvidia has a market share of over 80% and continues to lead in training and reasoning; In the data center AI acceleration market, NVIDIA's market share reached 82% in 2022, with AWS and Xilinx accounting for 8% and 4% respectively, and AMD, Intel, and Google accounting for 2%.


The domestic AI chips are represented by the Cambrian Siyuan series and Huawei Shengteng series. The performance of some of the Cambrian and Huawei Shengteng AI chip products has reached a high level, which is expected to accelerate the realization of domestic substitution and usher in a period of rapid development.


Local AI chip manufacturers in China:


未标题-12

Source: Guangfa Securities


Under the AI wave, artificial intelligence industries such as cloud computing, intelligent cars, and intelligent robots are rapidly developing, and the market demand for AI chips is constantly increasing. The size of the AI chip market will continue to grow.


Solemnly declare that the article only represents the author's views and does not represent the views of our company. The copyright of this article belongs to the original author, and the reprint of the article is only for the purpose of disseminating more information. If the author's information is marked incorrectly, please contact us immediately to modify or delete it. Thank you for your attention!

Hot news
AUO
TFT-LCD modules, TFT-LCD panels, energy storage/management systems, touch solutions, etc.
The working principle and classification of electromagnetic voltage transformers
Electromagnetic voltage transformers are commonly used in power systems to measure voltage on high-voltage transmission lines. They can also be used to monitor the voltage waveform and amplitude in the power system, in order to timely detect faults and problems in the power system. In this article, we will provide a detailed introduction to the working principle and classification of electromagnetic voltage transformers.
Differences between thermal relays and thermal overload relays
Thermal relays and thermal overload relays are common electrical protection devices, but their working principles and protection objects are different. In this article, we will provide a detailed introduction to the differences between thermal relays and thermal overload relays.
Types and Packaging of Tantalum Capacitors
Tantalum capacitors are electronic components that use tantalum metal as the electrode material. They are usually divided into two types: polarized and unpolarized, and come in various packaging forms. In this article, we will discuss in detail the types and packaging of tantalum capacitors.
The difference between thermal relays and fuses
Thermal relays and fuses are common electrical components that play a protective role in circuits. Although they can both interrupt the circuit, there are some differences between them. In this article, we will provide a detailed introduction to the differences between thermal relays and fuses.
FT2232 Development Board
A development board designed with FT2232 chip, which fully leads out the IO port, can be used to design an interface expansion board based on this.
AI high-performance computing - integrated storage and computing
Integrated storage and computing or in memory computing is the complete integration of storage and computing, directly utilizing memory for data processing or computation. Under the traditional von Neumann architecture, data storage and computation are separated. Due to the increasing performance gap between storage and computation, the speed at which the processor accesses stored data is much lower than the processor's computation speed. The energy consumption of data transportation between memory and main memory is also much higher than the energy consumed by the processor's computation.
AI High Performance Computing - Google TPU
Since Google launched the first generation self-developed artificial intelligence chip Tensor Processing Unit (TPU) in 2016, it has been upgraded to the fourth generation TPU v4 after several years of development (as of the end of 2022). The TPU architecture design also achieves efficient computation of network layers such as deep learning convolutional layer and fully connected layer by efficiently parallelizing a large number of multiplication and accumulation operations.
AI High Performance Computing - Cambrian NPU
The Cambrian period was one of the earliest AI chip companies in China to study. The design of their AI chip NPU (Neural Network Processing Unit) originated from a series of early AI chip architecture studies, mainly including DianNao, DaDianNao, PuDianNao, ShiDianNao, Cambricon-X, and other research achievements.
AI High Performance Computing - AI Chip Design
The simplest and most direct design approach for AI chips is to directly map neurons to hardware chips, as shown in the figure. The Full Hardware Implementation scheme maps each neuron to a logical computing unit and each synapse to a data storage unit. This architecture design can achieve a high-performance and low-power AI chip, such as an Intel ETANN chip. In the full hardware implementation scheme, the output data of the previous layer is multiplied by the weight, and the results of the multiplication are then added up, and then output to the next layer for calculation through an activation function. This architecture design tightly couples computing and storage, allowing the chip to avoid large-scale data access while performing high-speed computing, improving overall computing performance while also reducing power consumption.
User Info:
Phone number
+86
  • +86
  • +886
  • +852
Company Name
Email
Product model
Quantity
Comment message