Skip to content

Commit

Permalink
Update README.md (#322)
Browse files Browse the repository at this point in the history
Updated to move to ROCm 6.3 and post the issue with saving Tunable Ops due to PyTorch bug.
  • Loading branch information
t-parry authored Dec 12, 2024
1 parent a1aaa74 commit 7efa6e0
Showing 1 changed file with 5 additions and 6 deletions.
11 changes: 5 additions & 6 deletions docs/dev-docker/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,11 +10,11 @@ This documentation shows some reference performance numbers and the steps to rep

It includes:

- ROCm™ 6.2.2
- ROCm™ 6.3

- vLLM 0.6.3

- PyTorch 2.5dev (nightly)
- PyTorch 2.6dev (nightly)

## System configuration

Expand All @@ -23,7 +23,7 @@ The performance data below was measured on a server with MI300X accelerators wit
| System | MI300X with 8 GPUs |
|---|---|
| BKC | 24.13 |
| ROCm | version ROCm 6.2.2 |
| ROCm | version ROCm 6.3 |
| amdgpu | build 2009461 |
| OS | Ubuntu 22.04 |
| Linux Kernel | 5.15.0-117-generic |
Expand All @@ -45,9 +45,8 @@ You can pull the image with `docker pull rocm/vllm-dev:main`

### What is New

- MoE optimizations for Mixtral 8x22B, FP16
- Llama 3.2 stability improvements
- Llama 3.3 support
- ROCm 6.3 support
- Potential bug with Tunable Ops not saving due to a PyTorch issue


Gemms are tuned using PyTorch's Tunable Ops feature (https://github.com/pytorch/pytorch/blob/main/aten/src/ATen/cuda/tunable/README.md)
Expand Down

0 comments on commit 7efa6e0

Please sign in to comment.