ResourcesWhitepaper

Backend.AI + Intel Gaudi 3: Benchmarks and Deep Integration

Comprehensive whitepaper on Backend.AI and Intel Gaudi 3 AI Accelerators performance

A comprehensive whitepaper with detailed benchmarks showing how Backend.AI and Intel Gaudi 3 deliver remarkable performance for both small and large language model inference workloads.

Remarkable performance benchmarks of Backend.AI with Intel Gaudi 3 AI Accelerators.


This whitepaper presents detailed benchmarks of Backend.AI with Intel Gaudi 3 AI Accelerators, covering small model (Llama-3.1-8B-Instruct) and large model (Llama-3.1-70B-Instruct) performance. It includes deployment recommendations, optimization strategies, and demonstrates how the Sokovan orchestrator powers AI workloads on Intel Gaudi 3.

Related Services

Backend.AI

Backend.AI is a vendor-agnostic accelerated workload hosting platform based on our own home-grown orchestration and job scheduler, running on top of either cloud or on-premises (air-gapped) clusters.

Explore service
Intel

Bringing choice to gen AI with performance, scalability and efficiency. Meet your new high-performance option built to handle your AI workloads, your way.

Learn more

We're here for you!

Complete the form and we'll be in touch soon

Contact Us

Headquarter & HPC Lab

KR Office: 8F, 577, Seolleung-ro, Gangnam-gu, Seoul, Republic of Korea US Office: 3003 N First st, Suite 221, San Jose, CA 95134

© Lablup Inc. All rights reserved.

We value your privacy

We use cookies to enhance your browsing experience, analyze site traffic, and understand where our visitors are coming from. By clicking "Accept All", you consent to our use of cookies. Learn more