Home About us

Microsoft MAIA 100 AI accelerator, the core hardware of Azure

Zhineng Automobile 2024/09/05 08:16

At Hot Chips 2024, Microsoft unveiled its latest custom AI accelerator, MAIA 100, which has attracted a lot of attention in the industry.

As one of the core hardware of Microsoft's Azure platform, the MAIA 100 demonstrates Microsoft's ambition in the field of AI, as well as its deep accumulation of hardware customization and performance optimization.

MAIA 100 is Microsoft's custom AI accelerator designed specifically for the Azure platform to run OpenAI models.

Microsoft chose TSMC's CoWoS-S process technology for hardware design and adopted the 5nm technology node.

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

Part 1

Microsoft MAIA 100 

technical specifications and design philosophy

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

MAIA 100 reflects Microsoft's consideration of balancing performance and power consumption.

The MAIA 100 is equipped with 64GB of HBM2E storage, which is more than enough to handle the memory needs of larger models while avoiding competition with NVIDIA and AMD in the high-end HBM supply chain.

The MAIA 100 has a large L1/L2 cache of 500MB, which is not common in AI accelerators, and wants to use a larger cache to accelerate data access and improve computing efficiency.

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

In terms of network connectivity, the MAIA 100 provides 12x 400GbE of network bandwidth to support the large-scale data transfers required for high-performance computing.

This makes it ideal for hyperscale data center applications, especially in deep learning tasks that require large amounts of data transfer, with a TDP of up to 700W compared to about 500W per accelerator in actual production, which also shows the trade-off in performance and power consumption.

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

The MAIA 100's architecture is designed to highlight its high performance and versatility. Each SoC contains 16 clusters with four tiles within each cluster, enabling efficient parallel computing capabilities.

In addition, it includes an image decoder and confidential computing capabilities, further enhancing its multitasking capabilities.

MAIA 100 supports a wide range of data types and can handle both 9-bit and 6-bit calculations. This diverse data support allows it to flexibly respond to different AI computing tasks, thereby improving overall computing efficiency. This flexibility is especially important for high-performance computing, especially when dealing with tasks such as image and natural language processing.

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

Part 2

Ethernet-based interconnect technology

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

When it comes to interconnect technology, Microsoft has opted for an Ethernet-based solution and developed a custom RoCE-like protocol.

This choice, unlike the traditional InfiniBand, demonstrates Microsoft's innovative thinking in interconnect technology. Ethernet-based interconnect technology not only reduces costs, but also leverages existing network infrastructure to further drive adoption and adoption of the Azure platform.

As a major enabler of the Super Ethernet Consortium (UEC), Microsoft's choice of Ethernet-based interconnection was also a reasonable strategic choice.

By pushing for Ethernet-based interconnect standards, Microsoft can have a greater say in the data center market.

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

The MAIA 100 is not only innovative in terms of hardware, but also in its software ecosystem.

Microsoft provides a dedicated Maia SDK and asynchronous programming model for MAIA 100, which supports programming via Triton or Maia APIs. This design gives developers more flexibility to use high-level programming models or more granular performance optimizations through low-level APIs.

Maia 100 has out-of-the-box support for PyTorch models, allowing developers to simply switch the backend from CUDA to Maia, which greatly reduces migration costs and learning curve.

At the same time, toolchains such as maia-smi are available in the Maia-SDK, drawing on the experience of NVIDIA and AMD, further simplifying the development and debugging process.

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

The launch of MAIA 100, a custom AI accelerator for the Azure platform, shows that Microsoft is looking for more hardware autonomy and market competitiveness.

NVIDIA and AMD have always dominated the AI compute acceleration space, but Microsoft can better optimize the performance and cost structure of its cloud platform by customizing hardware solutions, especially in the field of large models and high-performance computing.

The MAIA 100's HBM capacity is relatively small, but its efficient caching and interconnect design, as well as its mission-specific optimization, give it a clear advantage in specific scenarios.

In the current energy-tight global environment, Microsoft's strategy is clearly to win more market share by offering lower-cost GPUs.

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

ZINENG Auto, 2024 Hot Chips|Microsoft MAIA 100 AI accelerator, core hardware of Azure

brief summary

Microsoft's MAIA 100 represents another bold attempt in AI. By innovating in hardware and software, Microsoft hopes to challenge not only the existing market leaders in terms of performance, but also in overall cost and energy efficiency. 

This article is from Xinzhi self-media and does not represent the views and positions of Business Xinzhi.If there is any suspicion of infringement, please contact the administrator of the Business News Platform.Contact: system@shangyexinzhi.com