What is GPUStack?

A GPU cluster manager

for running Large Language Models.

GPUStack allows you to create a unified cluster from any brand of GPUs in Apple MacBooks, Windows PCs, and Linux servers. Administrators can deploy LLMs from popular repositories such as Hugging Face. Developers can then access LLMs just as easily as accessing public LLM services from vendors like OpenAI or Microsoft Azure.

What can GPUStack
help you?

For AI Developers

For Admins

Integration & Serving

GPUStack aggregates all GPU resources within a cluster. It is designed to support all GPU vendors, including Nvidia, Apple, AMD, Intel, Qualcomm, and others. GPUStack is compatible with a laptops, desktops, workstations, and servers running MacOS, Windows, and Linux.

GPUStack can host any LLMs (on the cloud or on premises) and to make their functions available so that applications can incorporate AI into their systems.

Huggingface
ModelScope
Ollama
vLLM
LLaMA.cpp
Nvidia
Apple
AMD
Intel
Qualcomm
Huawei Ascend
OpenAI
Meta
Mistral
Qwen
BaiChuan
Yi
DeepSeek
OpenBMB

Latest Seal News

Read the latest news from our blog.