VLLM example inference server usage:

uv pip install vllm>=0.17.0 --torch-backend=auto
uv pip install huggingface-hub>=1.6.0 transformers>=5.3.0

vllm serve hhzm/qwen-3.5-9b-meow --reasoning-parser qwen3 --enable-auto-tool-choice --tool-call-parser qwen3_coder --enable-prefix-caching
Downloads last month
52
Safetensors
Model size
9B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for hhzm/qwen-3.5-9b-meow

Finetuned
(18)
this model

Dataset used to train hhzm/qwen-3.5-9b-meow