Home Google launches advanced Arm-based ‘Axion’ chip

Google launches advanced Arm-based ‘Axion’ chip


  • Google introduces a new Arm-based central processor, Axion, for AI operations.
  • Axion CPU offers 30% improved performance over general-purpose Arm chips and outperforms Intel processors by 50%.
  • Google also enhances its TPU AI chips with TPU v5p, designed for training large generative AI models.

Google has unveiled the details of a new version of its data center artificial intelligence chips and announced an Arm-based central processor.

An Arm processor is a type of CPU that uses the RISC architecture, simplifying instructions the computer has to process. Google’s Tensor Processing Units (TPUs) are one of the only alternatives to Nvidia’s sophisticated AI chips. However, developers can only use them via Google’s Cloud Platform and cannot purchase them directly.

However, Google’s new Axion CPU will first support the company’s AI operations before becoming available to Google Cloud’s business clients later this year. The company stated that its performance is better than that of x86 chips and general-purpose Arm chips in the cloud. The Axion chips will be used to run YouTube ads, power the Google Earth Engine, and support various other Google services.

The Axion Arm-based CPU will deliver 30 per cent improved performance over “general-purpose Arm chips” and outperform Intel’s current processors by 50 per cent.

“We’re making it easy for customers to bring their existing workloads to Arm,” Google Cloud’s vice president and general manager of compute and machine learning infrastructure, Mark Lohmeyer, told Reuters. “Axion is built on open foundations but customers using Arm anywhere can easily adopt Axion without re-architecting or re-writing their apps.”

Lohmeyer also stated on a blog that the tech giant is improving its TPU AI chips: “TPU v5p is a next-generation accelerator that is purpose-built to train some of the largest and most demanding generative AI models.” The Alphabet subsidiary announced that the new TPU v5p chip is designed to operate in pods containing 8,960 chips, delivering twice the raw performance of the previous TPU generation. The TPU v5p is already available via Google’s cloud.

Google’s new cloud AI Hypercomputer architecture features

Google stated it has made significant improvements to its hypercomputer architecture, focusing on performance-optimized hardware enhancements. This includes the general availability of Cloud TPU v5p and A3 Mega VMs, which are powered by NVIDIA H100 Tensor Core GPUs. These updates are said to offer higher performance for large-scale training and come with enhanced networking capabilities.

It has optimized its storage portfolio for AI workloads with the introduction of Hyperdisk ML, a new block storage service designed for AI inference/serving workloads. New caching capabilities in Cloud Storage FUSE and Parallelstore have also been introduced, improving training and inferencing throughput and latency.

On the software front, Google has introduced several open source developments. This includes JetStream, a throughput- and memory-optimized inference engine for large language models (LLMs), which provides higher performance per dollar on open models like Gemma 7B.

Google is also introducing new flexible consumption options to better accommodate varying workload needs. This includes the Dynamic Workload Scheduler, which features a calendar mode for start time assurance and a flex start mode designed for optimized economics, further enhancing the efficiency and flexibility of Google’s cloud computing offerings.

Featured image: Canva

About ReadWrite’s Editorial Process

The ReadWrite Editorial policy involves closely monitoring the tech industry for major developments, new product launches, AI breakthroughs, video game releases and other newsworthy events. Editors assign relevant stories to staff writers or freelance contributors with expertise in each particular topic area. Before publication, articles go through a rigorous round of editing for accuracy, clarity, and to ensure adherence to ReadWrite's style guidelines.

Suswati Basu
Tech journalist

Suswati Basu is a multilingual, award-winning editor and the founder of the intersectional literature channel, How To Be Books. She was shortlisted for the Guardian Mary Stott Prize and longlisted for the Guardian International Development Journalism Award. With 18 years of experience in the media industry, Suswati has held significant roles such as head of audience and deputy editor for NationalWorld news, digital editor for Channel 4 News and ITV News. She has also contributed to the Guardian and received training at the BBC As an audience, trends, and SEO specialist, she has participated in panel events alongside Google. Her…

Get the biggest tech headlines of the day delivered to your inbox

    By signing up, you agree to our Terms and Privacy Policy. Unsubscribe anytime.

    Tech News

    Explore the latest in tech with our Tech News. We cut through the noise for concise, relevant updates, keeping you informed about the rapidly evolving tech landscape with curated content that separates signal from noise.

    In-Depth Tech Stories

    Explore tech impact in In-Depth Stories. Narrative data journalism offers comprehensive analyses, revealing stories behind data. Understand industry trends for a deeper perspective on tech's intricate relationships with society.

    Expert Reviews

    Empower decisions with Expert Reviews, merging industry expertise and insightful analysis. Delve into tech intricacies, get the best deals, and stay ahead with our trustworthy guide to navigating the ever-changing tech market.