Be a part of our every day and weekly newsletters for the newest updates and unique content material on industry-leading AI protection. Study Extra
On the Ignite developer convention immediately, Microsoft unveiled two new chips designed for its knowledge middle infrastructure: the Azure Built-in HSM and the Azure Enhance DPU.
Scheduled for launch within the coming months, these custom-designed chips purpose to deal with safety and effectivity gaps confronted in present knowledge facilities, additional optimizing their servers for large-scale AI workloads. The announcement follows the launch of Microsoft’s Maia AI accelerators and Cobalt CPUs, marking one other main step within the firm’s complete technique to rethink and optimize each layer of its stack— from silicon to software program—to assist superior AI.
The Satya Nadella-led firm additionally detailed new approaches aimed toward managing energy utilization and warmth emissions of information facilities, as many proceed to lift alarms over the environmental influence of information facilities working AI.
Only recently, Goldman Sachs revealed analysis estimating that superior AI workloads are poised to drive a 160% improve in knowledge middle energy demand by 2030, with these amenities consuming 3-4% of world energy by the top of the last decade.
The brand new chips
Whereas persevering with to make use of industry-leading {hardware} from firms like Nvidia and AMD, Microsoft has been pushing the bar with its {custom} chips.
Final yr at Ignite, the corporate made headlines with Azure Maia AI accelerator, optimized for synthetic intelligence duties and generative AI, in addition to Azure Cobalt CPU, an Arm-based processor tailor-made to run general-purpose compute workloads on the Microsoft Cloud.
Now, as the subsequent step on this journey, it has expanded its {custom} silicon portfolio with a particular deal with safety and effectivity.
The brand new in-house safety chip, Azure Built-in HSM, comes with a devoted {hardware} safety module, designed to satisfy FIPS 140-3 Degree 3 safety requirements.
In line with Omar Khan, the vp for Azure Infrastructure advertising, the module primarily hardens key administration to ensure encryption and signing keys keep safe throughout the bounds of the chip, with out compromising efficiency or growing latency.
To realize this, Azure Built-in HSM leverages specialised {hardware} cryptographic accelerators that allow safe, high-performance cryptographic operations instantly throughout the chip’s bodily remoted atmosphere. Not like conventional HSM architectures that require community round-trips or key extraction, the chip performs encryption, decryption, signing, and verification operations completely inside its devoted {hardware} boundary.
Whereas Built-in HSM paves the best way for enhanced knowledge safety, Azure Enhance DPU (knowledge processing unit) optimizes knowledge facilities for extremely multiplexed knowledge streams akin to thousands and thousands of community connections, with a deal with energy effectivity.
The providing, first within the class from Microsoft, enhances CPUs and GPUs by absorbing a number of elements of a conventional server right into a single piece of silicon — proper from high-speed Ethernet and PCIe interfaces to community and storage engines, knowledge accelerators and safety features.
It really works with a classy hardware-software co-design, the place a {custom}, light-weight data-flow working system allows greater efficiency, decrease energy consumption and enhanced effectivity in comparison with conventional implementations.
Microsoft expects the chip will simply run cloud storage workloads at 3 times much less energy and 4 instances the efficiency in comparison with present CPU-based servers.
New approaches to cooling, energy optimization
Along with the brand new chips, Microsoft additionally shared developments made in direction of enhancing knowledge middle cooling and optimizing their energy consumption.
For cooling, the corporate introduced a complicated model of its warmth exchanger unit – a liquid cooling ‘sidekick’ rack. It didn’t share the particular positive factors promised by the tech however famous that it may be retrofitted into Azure knowledge facilities to handle warmth emissions from large-scale AI techniques utilizing AI accelerators and power-hungry GPUs equivalent to these from Nvidia.
On the vitality administration entrance, the corporate mentioned it has collaborated with Meta on a brand new disaggregated energy rack, aimed toward enhancing flexibility and scalability.
“Each disaggregated power rack will feature 400-volt DC power that enables up to 35% more AI accelerators in each server rack, enabling dynamic power adjustments to meet the different demands of AI workloads,” Khan wrote within the weblog.
Microsoft is open-sourcing the cooling and energy rack specs for the {industry} by way of the Open Compute Challenge. As for the brand new chips, the corporate mentioned it plans to put in Azure Built-in HSMs in each new knowledge middle server beginning subsequent yr. The timeline for the DPU roll-out, nevertheless, stays unclear at this stage.
Microsoft Ignite runs from November 19-22, 2024