Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Reform UK proposes visa freezes tied to deportation returns

    February 24, 2026

    ILO and IMO mark 20 years of maritime labour rights

    February 24, 2026

    Navneet AI and IIT Gandhinagar Announce Landmark MoU to Strengthen AI Education in India

    February 24, 2026
    Trending
    • Reform UK proposes visa freezes tied to deportation returns
    • ILO and IMO mark 20 years of maritime labour rights
    • Navneet AI and IIT Gandhinagar Announce Landmark MoU to Strengthen AI Education in India
    • TELUS Digital showcases AI transformation in telecom: Unlocking value with innovative use cases at Mobile World Congress 2026
    • Star Health Launches 32 Arogya Seva Kendras; to Provide Primary Healthcare Access to over a Million Lives
    • Tuinmaximaal Opens Several New Showrooms in Germany
    • Frontier Biotech and GSK Enter Global Exclusive Licensing Agreement for Small Interfering RNA Therapeutics
    • Mars Celebrates Latest European Decarbonization Milestone with New Renewable Electricity Contract in Sweden
    • Home
    • Contact Us
    Evening Post LondonEvening Post London
    Tuesday, February 24
    • Automotive
    • Business
    • Entertainment
    • Health
    • Lifestyle
    • Luxury
    • News
    • Sports
    • Technology
    • Travel
    Evening Post LondonEvening Post London
    Home » Maia 200 boosts Microsoft Azure with new AI inference silicon
    Technology

    Maia 200 boosts Microsoft Azure with new AI inference silicon

    January 28, 2026
    Facebook WhatsApp Twitter Pinterest LinkedIn Telegram Tumblr Email Reddit VKontakte

    EuroWire, SAN FRANCISCO: Microsoft on Jan. 26 introduced Maia 200, the second generation of its in-house artificial intelligence accelerator, built to run AI models in production across Azure data centres. The company said Maia 200 is designed for inference, the stage where trained models generate responses to live requests, and will be used to support a range of Microsoft AI services.

    Maia 200 boosts Microsoft Azure with new AI inference silicon
    Maia 200 brings high-bandwidth HBM3e memory to Microsoft’s Azure AI inference platform. (AI-generated image)

    Maia 200 is manufactured on TSMC’s 3-nanometer process and includes more than 140 billion transistors, Microsoft said. The chip pairs compute with a new memory system that includes 216 gigabytes of HBM3e high-bandwidth memory and about 272 megabytes of on-chip SRAM, aimed at sustaining large-scale token generation and other inference-heavy workloads.

    Microsoft said Maia 200 delivers more than 10 petaflops of performance at 4-bit precision and about 5 petaflops at 8-bit precision, formats commonly used to run modern generative AI efficiently. The company also said the system is designed around a 750-watt power envelope and is built with scalable networking so chips can be linked for larger deployments.

    The company said the new hardware has begun coming online in an Azure U.S. Central data centre in Iowa, with an additional location planned in Arizona. Microsoft described Maia 200 as its most efficient inference system deployed to date, reporting a 30% improvement in performance per dollar compared with its existing inference systems.

    AI inference focus and Azure deployment

    Microsoft said Maia 200 is intended to support AI products and services that rely on high-volume, low-latency model execution, including workloads running in Azure and Microsoft’s own applications. The company said it has designed the chip and the surrounding system as part of an end-to-end infrastructure approach that includes silicon, servers, networking and software for deploying AI models at scale.

    Alongside the chip, Microsoft announced early access to a Maia software development kit for developers and researchers working on model optimization. The company said the tooling is aimed at helping teams compile and tune models for Maia-based systems, and is structured to fit into common AI development workflows used for deploying inference in the cloud.

    Performance claims and model support

    Microsoft said Maia 200 is built to run large language models and advanced reasoning systems, and that it will be used for internal and hosted model deployments in Azure. The company has positioned the chip as a production inference accelerator, distinguishing it from training-focused systems that are typically used to build models before deployment.

    Microsoft has accelerated custom silicon work as demand has grown for compute to serve generative AI applications, where costs and availability of accelerators can affect how quickly services scale. Maia 200 follows Maia 100, which Microsoft introduced in 2023, and represents the company’s latest iteration of its dedicated AI accelerator line for datacenter inference.

    Related Posts

    Reform UK proposes visa freezes tied to deportation returns

    February 24, 2026

    ILO and IMO mark 20 years of maritime labour rights

    February 24, 2026

    EU commits 225 million euros for next-gen flu vaccines

    February 24, 2026

    EU seeks US clarity after Supreme Court tariff ruling

    February 23, 2026

    Frontex reports 60% drop in EU border crossings in January

    February 22, 2026

    France questions EU Commission role at Board of Peace

    February 21, 2026
    Latest News

    Reform UK proposes visa freezes tied to deportation returns

    February 24, 2026

    ILO and IMO mark 20 years of maritime labour rights

    February 24, 2026

    EU commits 225 million euros for next-gen flu vaccines

    February 24, 2026

    EU seeks US clarity after Supreme Court tariff ruling

    February 23, 2026

    Frontex reports 60% drop in EU border crossings in January

    February 22, 2026

    France questions EU Commission role at Board of Peace

    February 21, 2026

    US warns EU on curbs for American arms makers

    February 21, 2026

    WMO issues guidance to map national renewable resources

    February 21, 2026
    © 2024 Evening Post London | All Rights Reserved
    • Home
    • Contact Us

    Type above and press Enter to search. Press Esc to cancel.