Training examples ordered from simplest to most advanced.
| Example | GPUs | Backend | Model | Nodes | Difficulty |
|---|---|---|---|---|---|
| hf-quickstart | 3 | HuggingFace | Qwen3-8B | 1 | Easiest |
| qwen3-8b-single-node | 4+ | SGLang | Qwen3-8B | 1 | Easy |
| kimi-k25-2node-h200 | 16 | SGLang | Kimi-K2.5 | 2 | Advanced |
| kimi-k25-3node-h100 | 24 | SGLang | Kimi-K2.5 | 3 | Advanced |
If you just want to try TorchSpec locally, start with hf-quickstart (3 GPUs, no SGLang dependency):
./examples/hf-quickstart/run.shFor production workloads with async inference, use qwen3-8b-single-node:
./examples/qwen3-8b-single-node/run.shExamples use SGLang by default. To use vLLM instead:
# Use vLLM backend with qwen3-8b-single-node example
./examples/qwen3-8b-single-node/run.sh \
--config configs/vllm_qwen3_8b.yaml \Sample training data is in data/sample_conversations.jsonl. All examples that use local data point to this file by default.