Documentation Index
Fetch the complete documentation index at: https://www.radixark.com/llms.txt
Use this file to discover all available pages before exploring further.
1. Model Introduction
GLM-Z1-9B-0414 is a dense reasoning-tuned model from Zhipu AI’s GLM-4 series, sized for single-node experimentation. Key highlights:- Dense 9 B architecture: fits comfortably on a single 8-GPU node.
- Reasoning-tuned: post-trained for step-by-step reasoning under the GLM-Z1 line.
- Compatible RL recipe: GRPO with DAPO-style rollout, drop-in replacement for other dense Qwen / LLaMA-class workloads.
2. Supported Variants
| Model | HF ID |
|---|---|
| GLM-Z1-9B-0414 | zai-org/GLM-Z1-9B-0414 |
3. Environment Setup
3.1 Download model + datasets
3.2 HF → Megatron torch_dist conversion
4. Launch
4.1 Quick start
5. Recipe Configuration
5.1 Parallelism
| Script | TP | PP | CP | EP | max_tokens_per_gpu | actor / rollout GPUs | GPUs |
|---|---|---|---|---|---|---|---|
run-glm4-9B.sh | 2 | 1 | 2 | 1 | 4608 | 4 / 4 (non-colocate) | 8 (1 × 8) |
run-glm4-9B-4xgpu-radixtree.sh | 2 | 1 | 1 | 1 | 2304 | 4 / 2 | 4 (1 × 4) |
5.2 Algorithm
GRPO across both scripts:5.3 Rollout & SGLang
5.4 Optimizer
CPU Adam is not enabled in either launcher.5.5 Notable quirks
run-glm4-9B.shruns actor and rollout on disjoint GPUs (non-colocate).

