Skip to content

Latest commit

 

History

History
38 lines (26 loc) · 1.2 KB

File metadata and controls

38 lines (26 loc) · 1.2 KB

Examples

Training examples ordered from simplest to most advanced.

Example GPUs Backend Model Nodes Difficulty
hf-quickstart 3 HuggingFace Qwen3-8B 1 Easiest
qwen3-8b-single-node 4+ SGLang Qwen3-8B 1 Easy
kimi-k25-2node-h200 16 SGLang Kimi-K2.5 2 Advanced
kimi-k25-3node-h100 24 SGLang Kimi-K2.5 3 Advanced

Quick start

If you just want to try TorchSpec locally, start with hf-quickstart (3 GPUs, no SGLang dependency):

./examples/hf-quickstart/run.sh

For production workloads with async inference, use qwen3-8b-single-node:

./examples/qwen3-8b-single-node/run.sh

Switching inference backends

Examples use SGLang by default. To use vLLM instead:

# Use vLLM backend with qwen3-8b-single-node example
./examples/qwen3-8b-single-node/run.sh \
    --config configs/vllm_qwen3_8b.yaml \

Data

Sample training data is in data/sample_conversations.jsonl. All examples that use local data point to this file by default.