The Dataset Viewer has been disabled on this dataset.
Transformers Continuous Batching Scripts
GPU inference scripts using transformers' native continuous batching (CB). No vLLM dependency required.
Why transformers CB?
- Instant new model support - works with any model supported by transformers, including newly released architectures. No waiting for vLLM to add support.
- No dependency headaches - no vLLM, flashinfer, or custom wheel indexes. Just
transformers+accelerate. - Simple HF Jobs setup - no Docker image needed. Just
hf jobs uv run. - ~95% of vLLM throughput - uses PagedAttention and continuous scheduling for near-vLLM performance.
Available Scripts
generate-responses.py
Generate responses for prompts in a dataset. Supports chat messages and plain text prompts.
Quick Start
# Local (requires GPU)
uv run generate-responses.py \
username/input-dataset \
username/output-dataset \
--prompt-column question
# HF Jobs (single GPU)
hf jobs uv run --flavor l4x1 -s HF_TOKEN \
https://huggingface.co/datasets/uv-scripts/transformers-inference/raw/main/generate-responses.py \
username/input-dataset \
username/output-dataset \
--prompt-column question \
--max-tokens 1024
# HF Jobs (multi-GPU for larger models)
hf jobs uv run --flavor l4x4 -s HF_TOKEN \
https://huggingface.co/datasets/uv-scripts/transformers-inference/raw/main/generate-responses.py \
username/input-dataset \
username/output-dataset \
--model-id Qwen/Qwen3-30B-A3B-Instruct-2507 \
--messages-column messages \
--max-batch-tokens 2048 \
--max-tokens 4096
Example with SmolTalk2
# Generate responses for SmolTalk2 chat data
hf jobs uv run --flavor l4x1 -s HF_TOKEN \
https://huggingface.co/datasets/uv-scripts/transformers-inference/raw/main/generate-responses.py \
HuggingFaceTB/smoltalk2 username/smoltalk2-responses \
--subset SFT \
--split OpenHermes_2.5_no_think \
--messages-column messages \
--max-tokens 256
Parameters
| Parameter | Default | Description |
|---|---|---|
--model-id |
Qwen/Qwen3-4B-Instruct-2507 |
Any HF causal LM model |
--messages-column |
messages |
Column with chat messages |
--prompt-column |
- | Column with plain text prompts (alternative to messages) |
--output-column |
response |
Name for the generated response column |
--temperature |
0.7 |
Sampling temperature |
--top-p |
0.8 |
Top-p (nucleus) sampling |
--top-k |
20 |
Top-k sampling |
--max-tokens |
4096 |
Maximum tokens to generate per response |
--repetition-penalty |
1.0 |
Repetition penalty |
--max-batch-tokens |
512 |
Token budget per scheduling step (see below) |
--dtype |
bfloat16 |
Model precision (bfloat16, float16, float32) |
--attn-implementation |
`paged | sdpa` |
--max-samples |
all | Limit to N samples (useful for testing) |
--hf-token |
- | HF token (or use HF_TOKEN env var) |
--skip-long-prompts |
True |
Skip prompts exceeding context length |
Tuning --max-batch-tokens
This is the key performance parameter. It controls how many tokens the continuous batching scheduler processes per step:
- Too low (e.g., 128): GPU underutilized, slow throughput
- Too high (e.g., 8192): May cause out-of-memory errors
- Default 512: Conservative, works on most GPUs
- Recommended for A100/H100: 2048-4096
- Recommended for L4: 512-1024
If you hit OOM errors, reduce this value or switch to --dtype float16.
Current Limitations
- Single GPU only -
device_map="auto"(pipeline parallelism) doesn't work with CB's PagedAttention cache. Transformers does have tensor parallelism (tp_plan="auto") for supported models, but it requirestorchrunand is undocumented with CB. For now, use a model that fits on one GPU (e.g., 8B in bf16 on A10G/L4 with 24GB). - Text-only - no vision-language model support yet.
When to use this vs vLLM
| Transformers CB | vLLM | |
|---|---|---|
| Best for | New/niche models, simple setup, avoiding dependency issues | Maximum throughput, production serving |
| Model support | Any transformers model, immediately | Popular models, may lag on new architectures |
| Dependencies | transformers + accelerate |
vllm + flashinfer + custom indexes |
| Docker image | Not needed | vllm/vllm-openai recommended |
| Multi-GPU | Single GPU only (for now) | Tensor parallelism |
| Performance | ~95% of vLLM for text generation | Fastest for supported models |
| VLM support | Not yet | Yes |
Rule of thumb: Use transformers CB when you want simplicity and broad model support. Use vLLM when you need maximum throughput with well-supported models.
- Downloads last month
- -