Microsoft reveals custom chips to power AI workloads on Azure

2023-11-20
关注

  •  

Microsoft has unveiled its first custom silicon, launching two chips, the Maia 100 (M100) AI accelerator and the Cobolt 100 CPU, designed to handle artificial intelligence and general purpose workloads on its Azure cloud platform.

Microsoft has unveiled two custom chips for its Azure Cloud platform. (Photo courtesy of Microsoft)

The two chips represent a first foray into semiconductors for Microsoft and see the company follow in the footsteps of its public cloud rivals, Amazon’s AWS and Google Cloud, which run their own chips in their data centres alongside those provided by vendors such as Nvidia.

Maia 100 AI accelerator and Cobolt 100 CPU unveiled

Both the new chips will be available early next year, and are designed on Arm architecture, which is increasingly being deployed in cloud datacentres as an alternative to semiconductors built using Intel’s x86 process, the long-time market leader. Microsoft already offers some Arm-based CPUs on Azure, having struck a partnership with Ampere Computing last year, but claims Cobolt will deliver a 40% performance increase compared to Ampere’s chips.

The Maia 100 will apparently “power some of the largest internal AI workloads running on Microsoft Azure”, such as the Microsoft Copilot AI assistant and the Azure OpenAI Service, which allows MSFT’s cloud customers to access services from AI lab OpenAI, the creator of ChatGPT.

Microsoft is “building the infrastructure to support AI innovation, and we are reimagining every aspect of our data centres to meet the needs of our customers”, said Scott Guthrie, executive vice-president of the company’s cloud and AI group. “At the scale we operate, it’s important for us to optimise and integrate every layer of the infrastructure stack to maximize performance, diversify our supply chain and give customers infrastructure choice.”

Customers will be able to choose from a wider range of chips from other vendors, too, with Microsoft introducing virtual machines featuring Nvidia’s H100 Tensor Core GPUs, the most powerful AI chip currently on the market. It also plans to add the vendor’s H200 Tensor Core GPU, launched this week, to its fleet next year to “support larger model inferencing with no reduction in latency”.

It is also adding accelerated virtual machines using AMD’s top-of-the-range MI300X design to Azure.

Microsoft was an early adopter of AI tools through its partnership with OpenAI, in which it invested billions of dollars earlier this year. OpenAI CEO Sam Altman is enthusiastic about the M100’s potential, and said: “We were excited when Microsoft first shared their designs for the Maia chip, and we’ve worked together to refine and test it with our models.”

Content from our partners

Collaboration along the entire F&B supply chain can optimise and enhance business

Collaboration along the entire F&B supply chain can optimise and enhance business

Inside ransomware's hidden costs

Inside ransomware’s hidden costs

The future of data-driven enterprises is transforming data centre requirements

The future of data-driven enterprises is transforming data centre requirements

Altman added that he believes Azure’s AI architecture “paves the way for training more capable models and making those models cheaper for our customers.”

View all newsletters Sign up to our newsletters Data, insights and analysis delivered to you By The Tech Monitor team

Will Microsoft’s new chips give it an AI edge?

Microsoft is the last of the public cloud market’s big three to launch its own processors. Amazon offers its own range of Arm-based Graviton processors as an option to AWS customers, while Google uses in-house tensor processing units, or TPUs, for its AI systems.

James Sanders, principal analyst for cloud and infrastructure at CCS Insight, said: “Microsoft notes that Cobalt delivers up to 40% performance improvement over current generations of Azure Arm chips. Rather than depend on external vendors to deliver the part Microsoft needs, building this in-house and manufacturing it at a partner fab provides Microsoft greater flexibility to gain the compute power it needs.”

Sanders argues that the benefits of developing the Maia 100 are clear. He said: “With Microsoft’s investment in OpenAI, and the burgeoning popularity of OpenAI products such as ChatGPT as well as Microsoft’s Copilot functionality in Office, GitHub, Bing, and others, the creation of a custom accelerator was inevitable. At the scale which Microsoft is operating, bringing this computing capacity online while delivering the unit economics to make this direction viably profitable, requires a custom accelerator.”

Read more: Google Cloud launches Vertex AI data residency regions

  •  

  • en
您觉得本篇内容如何
评分

相关产品

EN 650 & EN 650.3 观察窗

EN 650.3 version is for use with fluids containing alcohol.

Acromag 966EN 温度信号调节器

这些模块为多达6个输入通道提供了一个独立的以太网接口。多量程输入接收来自各种传感器和设备的信号。高分辨率,低噪音,A/D转换器提供高精度和可靠性。三路隔离进一步提高了系统性能。,两种以太网协议可用。选择Ethernet Modbus TCP\/IP或Ethernet\/IP。,i2o功能仅在6通道以太网Modbus TCP\/IP模块上可用。,功能

雷克兰 EN15F 其他

品牌;雷克兰 型号; EN15F 功能;防化学 名称;防化手套

Honeywell USA CSLA2EN 电流传感器

CSLA系列感应模拟电流传感器集成了SS490系列线性霍尔效应传感器集成电路。该传感元件组装在印刷电路板安装外壳中。这种住房有四种配置。正常安装是用0.375英寸4-40螺钉和方螺母(没有提供)插入外壳或6-20自攻螺钉。所述传感器、磁通收集器和壳体的组合包括所述支架组件。这些传感器是比例测量的。

TMP Pro Distribution C012EN RF 音频麦克风

C012E射频从上到下由实心黄铜制成,非常适合于要求音质的极端环境,具有非常坚固的外壳。内置的幻像电源模块具有完全的射频保护,以防止在800 Mhz-1.2 Ghz频段工作的GSM设备的干扰。极性模式:心形频率响应:50赫兹-18千赫灵敏度:-47dB+\/-3dB@1千赫

ValueTronics DLRO200-EN 毫欧表

"The DLRO200-EN ducter ohmmeter is a dlro from Megger."

评论

您需要登录才可以回复|注册

提交评论

提取码
复制提取码
点击跳转至百度网盘