Site icon Aragon Research

Microsoft Silicon—2 New Chips to Power Azure Cloud and AI

By: Craig Kennedy

 

Microsoft Silicon—2 New Chips to Power Azure Cloud and AI

Earlier this month at Microsoft’s annual Ignite conference, Microsoft CEO Satya Nadella announced two new custom silicon chips designed and built by Microsoft to power its Azure cloud infrastructure. The Maia 100 and Cobalt 100 chips are being piloted today and are scheduled to be available in Microsoft Azure data centers beginning early in 2024.

Azure Maia AI Chip: Designed for AI Workloads

Microsoft unveiled the Azure Maia AI Accelerator, a novel AI chip, in response to the growing demand for AI technology. This chip codenamed “Athena,” allows Microsoft to diversify its supply chain and provides its customers a choice in infrastructure.

The Azure Maia AI chip is designed to address GPU shortages and reduce reliance on NVIDIA designed GPUs. The chip is designed to run large language models (LLMs) and Microsoft worked in tandem with OpenAI to help refine the chip for efficiently running their models.

Azure Cobalt CPU: Arm Based Power for Cloud Computing

Microsoft launched the Azure Cobalt 100 CPU, a revolutionary processor based on Arm Neoverse CSS. It is a 64-bit, 128 core chip that boasts a 40% boost in performance over the generation of Arm chips currently in Azure datacenters. The Azure Cobalt CPU is engineered to efficiently run scale-out, cloud-native workloads at a significantly reduced power.

Fast Network Performance is Key

The introduction of the Azure Maia AI chip and the Azure Cobalt CPU represents a significant shift in the cloud computing and AI landscape. Both chips share some impressive network technology, including hollow core fiber, providing 12.5 GBs throughput and 650K IOPs for remote storage and 200 GBs in network performance. These technologies not only enhance the performance of cloud-native workloads but also promise better overall total cost of ownership for Microsoft.

Keeping Your Cool in the Datacenter

The Maia 100 chip requires liquid cooling to dissipate heat generated by these chips when running intensive AI tasks. Standard server racks couldn’t be used to house the Maia server boards, so Microsoft designed a new server rack that would house the new Maia server boards and provide ample room for the liquid cooling, as well as network cabling required for remote storage and networking between servers. 

Bottom Line

Microsoft is joining the other hyperscalers in designing and running its in-house silicon in its datacenters. Google has its TPU chip, IBM has its AIU chip, AWS has its Graviton, Trainium, and Inferentia chips, and now Microsoft has its Cobalt and Maia chips. By developing its own AI chip and Arm processors, Microsoft is easing some of its supply chain issues with the insatiable demand for leading AI chips while cutting costs to run its AI workloads.


Get Ready for 2024 with Aragon’s 2023 Q1 Research Agenda!

Wednesday, January 17th, 2024 at 10 AM PT | 1 PM ET

 

Aragon Research’s 2024 Q1 Agenda

Aragon Research provides the strategic insights and advice you need to help your business navigate disruption and outperform your goals. Our research is designed to help you understand the technologies that will impact your business–using a number of trusted research methodologies that have been proven to help organizations like yours get to business outcomes faster.

On Wednesday, January 17th, 2024, join Aragon Research CEO and Lead Analyst, Jim Lundy for a complimentary webinar as they walk you through Aragon’s Q1 2024 research agenda.

This webinar will cover:

Register Here


 

Blog 1: Introducing the Digital Operations Blog Series

Blog 2: Digital Operations: Keeping Your Infrastructure Secure

Blog 3: Digital Operations: Cloud Computing

Blog 4: Cybersecurity Attacks Have Been Silently Escalating

Blog 5: Automation—The Key to Success in Today’s Digital World

Blog 6: Infrastructure—Making the Right Choices in a Digital World

Blog 7: Open-Source Software—Is Your Supply Chain at Risk?

Blog 8: IBM AIU—A System on a Chip Designed For AI

Blog 9: IBM Quantum: The Osprey Is Here

Blog 10: The Persistence of Log4j

Blog 11: AWS re:Invent 2022—Focus on Zero-ETL for AWS

Blog 12: AWS re:Invent 2022—The Customer Is Always Right

Blog 13: How Good is the New ChatGPT?

Blog 14: The U.S. Department of Defense Embraces Multi-Cloud

Blog 15: 2022 Digital Operations—The Year in Review

Blog 16: Lucky Number 13 for Intel—Intel Is Back on Top

Blog 17: Quantum Decryption—The Holy Grail for Cybercriminals

Blog 18: Microsoft and OpenAI—Intelligent Partnership

Blog 19: ChatGPT—The First One Is Free

Blog 20: Bing and ChatGPT—Your Co-Pilot When Searching the Web

Blog 21: ESXiArgs—Ransomware Attack on VMware

Blog 22: The Cost of Supply Chain Security—$250M in Sales

Blog 23: OpenAI Delivers on APIs—Accelerating the Adoption of ChatGPT

Blog 24: OpenAI Delivers on Plugins—Is ChatGPT The New Generative Content Platform?

Blog 25: Microsoft Security Copilot—Defending the Enterprise at the Speed of AI

Blog 26: Operation Cookie Monster Takes a Huge Bite Out of The Dark Web

Blog 27: AWS Bedrock—Amazon’s Generative AI Launch

Blog 28: Google Cloud Security AI Workbench – Conversational Security

Blog 29: World Password Day – Is This the Last Anniversary

Blog 30: Intel Partners to Enter the Generative AI Race—Aurora genAI

Blog 31: Charlotte AI – CrowdStrike Enters the Generative AI Cybersecurity Race

Blog 32: NICE Catches the Generative AI Wave

Blog 33: AMD Instinct MI300X—A New Challenger to Nvidia

Blog 34: Storm-0558—Chinese Cyber Attack on US Government Organizations

Blog 35: Network Resilience Coalition—Making the Network Safer

Blog 36: Frontier Model Forum—Power Players Unite to Make AI Safer

Blog 37: Intel Is Back in the Foundry Business—Entering the Angstrom Era

Blog 38: Check Point Acquires Perimeter 81—Securing Remote Access

Blog 39: PSA—Akira Ransomware Targeting Cisco VPNs

Blog 40: Volt Typhoon: A Chinese Cyberwar Threat to Critical US Infrastructure

Exit mobile version