Skip to content

Commit b36dea7

Browse files
committed
Document Sync by Tina
1 parent 51bb18e commit b36dea7

File tree

2 files changed

+14
-7
lines changed

2 files changed

+14
-7
lines changed

docs/stable/intro.md

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,10 @@ sidebar_position: 1
99

1010
ServerlessLLM is a **fast** and **easy-to-use** serving system designed for **affordable** multi-LLM serving, also known as LLM-as-a-Service. ServerlessLLM is ideal for environments with multiple LLMs that need to be served on limited GPU resources, as it enables efficient dynamic loading of LLMs onto GPUs. By elastically scaling model instances and multiplexing GPUs, ServerlessLLM can significantly reduce costs compared to traditional GPU-dedicated serving systems while still providing low-latency (Time-to-First-Token, TTFT) LLM completions.
1111

12+
ServerlessLLM now supports NVIDIA and AMD GPUs, including following hardware:
13+
* NVIDIA GPUs: Compute Capability 7.0+ (e.g, V100, A100, RTX A6000, GeForce RTX 3060)
14+
* AMD GPUs: ROCm 6.2.0+ (tested on MI100s and MI200s)
15+
1216
## Documentation
1317

1418
### Getting Started
@@ -25,6 +29,7 @@ ServerlessLLM is a **fast** and **easy-to-use** serving system designed for **af
2529
### ServerlessLLM Store
2630

2731
- [Quickstart](./store/quickstart.md)
32+
- [ROCm Installation(Experimental)](./store/installation_with_rocm.md)
2833

2934
### ServerlessLLM CLI
3035

docs/stable/store/installation_with_rocm.md

Lines changed: 9 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,15 @@ sidebar_position: 1
44

55
# Installation with ROCm (Experimental)
66

7+
## Latest Tested Version
8+
+ v0.5.1
9+
10+
## Tested Hardware
11+
+ OS: Ubuntu 22.04
12+
+ ROCm: 6.2
13+
+ PyTorch: 2.3.0
14+
+ GPU: MI100s (gfx908), MI200s (gfx90a)
15+
716
## Build the wheel from source and install
817
ServerlessLLM Store (`sllm-store`) currently provides experimental support for ROCm platform. Due to an internal bug in ROCm, serverless-llm-store may face a GPU memory leak in ROCm before version 6.2.0, as noted in [issue](https://github.com/ROCm/HIP/issues/3580).
918

@@ -162,13 +171,6 @@ cd ServerlessLLM/sllm_store/build
162171
ctest --output-on-failure
163172
```
164173
165-
## Tested Hardware
166-
+ OS: Ubuntu 22.04
167-
+ ROCm: 6.2
168-
+ PyTorch: 2.3.0
169-
+ GPU: MI100s (gfx908), MI200s (gfx90a)
170-
171-
172174
## Known issues
173175
174176
1. GPU memory leak in ROCm before version 6.2.0.

0 commit comments

Comments
 (0)