Ecosystem
Backend.AI

Backend.AI

-

AI Infra OS

  • OpenSource
    26.1
  • Enterprise
    R2

Transforming GPU complexity into simplicity

Backend.AI is an open-source AI Infrastructure OS that frees developers and data scientists from hardware complexity. Backend.AI supports 11+ heterogeneous accelerators from GPUs to NPUs, and ASICs like TPUs, plus seamless high-performance storage integration. Unlock your AI and HPC potential from training to inference with Backend.AI.

Turning infrastructure from a limitation into a competitive edge

Illustration of on-premise server, hybrid, and cloud setup

Backend.AI ensures reliable AI operations across any environment, whether on‑premises, in the cloud, or in hybrid settings. Even in isolated offline networks, it delivers the same stability and usability. Whatever your infrastructure, Backend.AI is ready to power your AI business with confidence.

Illustration of on-premise server, hybrid, and cloud setup

A vendor‑neutral platform supporting 11+ AI accelerators

Backend.AI supports the latest NVIDIA GPUs including Blackwell, as well as Intel Gaudi, AMD Instinct, Rebellions ATOM+, Furiosa RNGD, and more. Choose the right accelerator for your workload and budget, and even run heterogeneous hardware side by side. Adding new accelerator types is effortless. Backend.AI manages them consistently through a unified abstraction layer and flexible plugin architecture.

Logo of NVIDIA, Intel, Graphcore, Rebellions, Furiosa, Tenstorrent, AMD, HyperAccel, Groq, Sambanova, and more.

NVIDIA DGX™-Ready Software, designed for DGX systems

Image of Cluster Management software level diagram. GPU, Network, Storage at bottom and Backend.AI managing all of them.

Backend.AI is validated as NVIDIA DGX™‑Ready Software, designed to unlock the full performance of DGX systems. Built on NVIDIA’s container runtime, it enhances enterprise‑grade AI operations. With advanced GPU sharing, fine‑grained allocation, and topology‑aware scheduling, teams can efficiently utilize resources and ensure stable, consistent performance across DGX clusters.

Image of Cluster Management software level diagram. GPU, Network, Storage at bottom and Backend.AI managing all of them.

Proprietary CUDA virtualization layer: Container-level GPU virtualization

Backend.AI's in-house CUDA virtualization layer lets you split and merge GPUs at the container level with total flexibility. Share a single GPU across multiple users or aggregate several into one massive workload. This works perfectly for light tasks like education and inference as well as heavy model training. Lablup's internationally patented container-level GPU virtualization ensures the smartest, waste-free utilization of GPUs in AI and HPC environments.

Illustration explaining GPU Virtualization

Perfect integration with High-performance storage

Backend.AI integrates seamlessly with high-performance storage systems for AI and HPC workloads. Vendor-optimized configurations ensure stable, efficient performance and unlock full storage potential. Use parallel file systems for throughput, distributed storage for scalability, and low-latency solutions for fast data loading, checkpointing, and result storage.

Easy infrastructure management via User-friendly WebUI

Backend.AI’s intuitive WebUI makes cluster monitoring and control simple and visual. View GPU, CPU, and memory usage with active sessions in real time. Manage sessions, adjust quotas, and set permissions per user, project, or tenant to handle complex policies easily. Developers and researchers can skip setup and dive straight into experiments and services.

Backend.AI WebUI Select Environment screenshot

We're here for you!

Complete the form and we'll be in touch soon

Contact Us

Headquarter & HPC Lab

KR Office: 8F, 577, Seolleung-ro, Gangnam-gu, Seoul, Republic of Korea US Office: 3003 N First st, Suite 221, San Jose, CA 95134

© Lablup Inc. All rights reserved.

We value your privacy

We use cookies to enhance your browsing experience, analyze site traffic, and understand where our visitors are coming from. By clicking "Accept All", you consent to our use of cookies. Learn more