VLLM example inference server usage:
uv pip install vllm>=0.17.0 --torch-backend=auto
uv pip install huggingface-hub>=1.6.0 transformers>=5.3.0
vllm serve hhzm/qwen-3.5-9b-meow --reasoning-parser qwen3 --enable-auto-tool-choice --tool-call-parser qwen3_coder --enable-prefix-caching
- Downloads last month
- 52
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for hhzm/qwen-3.5-9b-meow
Base model
Qwen/Qwen3.5-9B-Base Finetuned
trohrbaugh/Qwen3.5-9B-heretic-v2