Skip to content

Commit cb0392e

Browse files
committed
feat: add requirement for superoffload
1 parent f196311 commit cb0392e

2 files changed

Lines changed: 18 additions & 3 deletions

File tree

training/DeepSpeed-SuperOffload/README.md

Lines changed: 9 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -11,11 +11,17 @@ With common sequence length and batch size, SuperOffload can deliver up to ~500
1111

1212
## Quick Start
1313

14-
### 1. No custom model code required
14+
### 1. Install dependencies
15+
16+
```bash
17+
pip install -r requirements.txt
18+
```
19+
20+
### 2. No custom model code required
1521

1622
All examples use Hugging Face Transformers and DeepSpeed ZeRO Stage 3, no custom modeling code required.
1723

18-
### 2. Enable SuperOffload (one line)
24+
### 3. Enable SuperOffload (one line)
1925

2026
Add the `super_offload` flag to the `offload_optimizer` block in the ZeRO Stage 3 DeepSpeed config:
2127

@@ -34,7 +40,7 @@ Add the `super_offload` flag to the `offload_optimizer` block in the ZeRO Stage
3440

3541
To fall back to ZeRO-Offload, remove `"super_offload": true` (and optionally `cpuadam_cores_perc`).
3642

37-
### 3. Run a fine-tuning script
43+
### 4. Run a fine-tuning script
3844

3945
Fine-tune GPT-OSS-20B (1× GH200):
4046

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,9 @@
1+
torch>=2.5.1
2+
deepspeed>=0.17.0
3+
datasets>=4.0.0
4+
transformers>=4.56.1
5+
numpy>=1.21.0
6+
flash-attn>=2.0.0
7+
wandb
8+
packaging
9+
psutil

0 commit comments

Comments
 (0)