Registered users can unlock up to five pieces of premium content each month.
AI Accelerated Hardware Unlocks On-Device Generative AI |
NEWS |
Generative Artificial Intelligence (AI) inferencing was initially restricted to cloud environments due to the vast memory, compute, and commensurate power to run workloads from models containing hundreds of billions of parameters. However, Small Language Models (SLMs), enabled through advances in compression techniques (including optimization), coupled with improvements in hardware performance through the addition of AI accelerators within heterogenous systems, are enabling inference to move toward more memory- and resource-constrained form factors like Personal Computers (PCs) (and smartphones). The market for on-device generative AI has made huge strides over the last year, with all major Original Equipment Manufacturers (OEMs) and chip vendors announcing and bringing to market solutions capable of handling AI inference workloads on-device. Notable developments in PC AI include:
Interestingly, both Apple and Qualcomm SoCs use Arm CPUs, eschewing the more power-hungry x86 CPUs from Intel and AMD. Apple’s successful journey to Arm SoCs in its PCs started with its 2020 M1 series chipsets, initiating its transition from Intel x86 processors. Two generations later, it has demonstrated Arm’s viability in the PC form factor. Moreover, the inclusion of neural engines in M1 heterogeneous SoCs, continued through the proceeding M2 and M3 families, proves the success in powering general AI workloads, and later, generative AI with the M3.
If Apple Can, Why Can't Others? |
IMPACT |
An impact of this innovation cycle is the renewed interest in Arm architectures running Windows, which remains the most popular Operating System (OS) for PCs. It was reported that NVIDIA and AMD have begun designing CPUs with Arm Instruction Set Architecture (ISA), as part of Microsoft’s drive for chip vendors to implement an alternative architecture to x86 and capitalize on, for example, the improved energy efficiency of Arm SoCs. Add to that Microsoft’s October 2023 announcement that it is launching Arm Advisory Service to help Independent Software Vendors (ISVs) working on software for Windows running on Arm to increase customer adoption of Arm systems, which coincides neatly with the rumors surrounding AMD and NVIDIA’s interest in the alternative architecture.
The vast majority of Windows PCs run with legacy x86 architecture CPUs, around which a billion-dollar software and OEM ecosystem has been built (Apple has its own OS, MacOS). However, Windows has long supported Arm architectures, and if confirmed, NVIDIA and AMD would thereby join Qualcomm, which has been producing SoCs for laptops using Arm CPUs for several years (and will continue to do so with Snapdragon X Elite, set to ship mid-2024 in time for the next academic year, when many will replace existing devices). Moreover, forthcoming PCs running Windows will need on-device AI capabilities as the company rolls out Copilot and other generative AI features requiring heterogenous architectures across the newest PCs running Windows 11 (and eventually 12), as the company proclaims 2024 to be the “year of the AI PC.” These capabilities are supported by Qualcomm’s PC chipset, and would be joined by any Arm PC systems developed by AMD and NVIDIA.
Applications, Not Hardware, Will Drive Adoption |
RECOMMENDATIONS |
For 2024 to be the year of the AI PC, and for Arm to catch up, the ISV ecosystem will need to build out its applications, which must be powerful enough to convince customers to upgrade. To this end, all of the above vendors offer Software Development Kits (SDKs) and toolkits to spur the development of AI software that can run efficiently on their respective hardware platforms. But for Arm to become a leader and join x86-powered Window PCs, its relatively niche position will need to be elevated. The ecosystem of OEMs, chip vendors, and ISVs must address several items:
Custom Arm licenses to run Windows are expensive, which raises the question if a RISC-V solution, explored in ABI Research’s RISC-V for Edge AI Applications report—and an alternative to x86 and Arm—would appear more to benefit from the instruction set’s flexibility and lack of licensing fees. The interest in RISC-V in China also renders it a potentially commercially beneficial decision, able to be developed using home-grown chipsets likely favored by Chinese consumers.