Neural processing unit
A neural processing unit (NPU), also known as AI accelerator or deep learning processor, is a class of specialized hardware accelerator[1] or computer system[2][3] designed to accelerate artificial intelligence (AI) and machine learning applications, including artificial neural networks and computer vision. They can be used either to efficiently execute already trained AI models (inference) or for training AI models. Typical applications include algorithms for robotics, Internet of Things, and other data-intensive or sensor-driven tasks.[4] They are often manycore designs and generally focus on low-precision arithmetic, novel dataflow architectures or in-memory computing capability. As of 2024[update], a typical AI integrated circuit chip contains tens of billions of MOSFETs.[5] AI accelerators are used in mobile devices such as Apple iPhones and Huawei cellphones,[6] and personal computers such as Intel laptops,[7] AMD laptops[8] and Apple silicon Macs.[9] Accelerators are used in cloud computing servers, including tensor processing units (TPU) in Google Cloud Platform[10] and Trainium and Inferentia chips in Amazon Web Services.[11] Many vendor-specific terms exist for devices in this category, and it is an emerging technology without a dominant design. Graphics processing units designed by companies such as Nvidia and AMD often include AI-specific hardware, and are commonly used as AI accelerators, both for training and inference.[12] Neural Processing Units (NPU) are another more native approach. Since 2017, several CPUs and SoCs have on-die NPUs: for example, Intel Meteor Lake, Lunar Lake, and Apple A11. All models of Intel Meteor Lake processors have a Versatile Processor Unit (VPU) built-in for accelerating inference for computer vision and deep learning.[13] References
External links
|
Portal di Ensiklopedia Dunia