Authors
- Kaoutar El Maghraouir*
- Kim Tran*
- Kurtis Ruby*
- Borja Godoy*
- Jordan Murray*
- Manuel Le Gallo-Bourdeau*
- Todd Deshane*
- Pablo Gonzalez*
- Diego Moreda*
- Hadjer Benmeziane*
- Corey Liam Lammie*
- Julian Büchel*
- Malte J. Rasch
- Abu Sebastian*
- Vijay Narayanan*
* External authors
Venue
- SSE 2024
Date
- 2024
Analog AI as a Service: A Cloud Platform for In-Memory Computing
Kaoutar El Maghraouir*
Kim Tran*
Kurtis Ruby*
Borja Godoy*
Jordan Murray*
Manuel Le Gallo-Bourdeau*
Todd Deshane*
Pablo Gonzalez*
Diego Moreda*
Hadjer Benmeziane*
Corey Liam Lammie*
Julian Büchel*
Malte J. Rasch
Abu Sebastian*
Vijay Narayanan*
* External authors
SSE 2024
2024
Abstract
This paper introduces the Analog AI Cloud Composer platform, a service that allows users to access Analog In-Memory Computing (AIMC) simulation and computing resources over the cloud. We introduce the concept of an Analog AI as a Service (AAaaS). AIMC offers a novel approach for decreasing both the latency and energy usage associated with Deep Neural Network (DNN) inference and training. This platform democratizes access to AIMC computing, making it available to a broader audience, including researchers, developers, and businesses. Emphasizing a user-friendly, no-code approach, AAaaS integrates the Analog Hardware Acceleration Kit (AIHWKit) simulation platform within a fully managed cloud environment. We discuss the architecture of the Analog AI Cloud Composer (AAICC), focusing on its key services such as inference, training, and AIMC hardware access. The platform's design, grounded in cloud services and guidelines, ensures a secure, data-centric user experience with robust control and validation mechanisms.
Related Publications
Analog in-memory accelerators present a promising solution for energy-efficient training and inference of large vision or language models. While the inference on analog accelerators has been studied recently, the analog training perspective is under-explored. Recent studies …
Analog in-memory computing is a promising future technology for efficiently accelerating deep learning networks. While using in-memory computing to accelerate the inference phase has been studied extensively, accelerating the training phase has received less attention, despi…
We developed a phase-change memory (PCM), with SiSbTe material, that showed state-independent resistance drift (v~0.04) at 65°C over the entire analog conductance range. We evaluated this PCM for In Memory Compute (IMC) applications simulating the performance of BERT model w…
JOIN US
Shape the Future of AI with Sony AI
We want to hear from those of you who have a strong desire
to shape the future of AI.