gpt-oss-20b-semanticwiki GGUF

GGUF conversion of GhostScientist/gpt-oss-20b-semanticwiki, fine-tuned on SemanticWiki data.

Available Quantizations

File Quant Description
gpt-oss-20b-semanticwiki-f16.gguf F16 Full precision
gpt-oss-20b-semanticwiki-q8_0.gguf Q8_0 8-bit (recommended for 32GB+ RAM)
gpt-oss-20b-semanticwiki-q5_k_m.gguf Q5_K_M 5-bit medium
gpt-oss-20b-semanticwiki-q4_k_m.gguf Q4_K_M 4-bit medium (smallest)

Usage

With Ollama

huggingface-cli download GhostScientist/gpt-oss-20b-semanticwiki-gguf gpt-oss-20b-semanticwiki-q8_0.gguf
echo "FROM ./gpt-oss-20b-semanticwiki-q8_0.gguf" > Modelfile
ollama create gpt-oss-semanticwiki -f Modelfile
ollama run gpt-oss-semanticwiki

With llama.cpp

./llama-cli -m gpt-oss-20b-semanticwiki-q8_0.gguf -p "Your prompt"

Model Details

  • Base Model: openai/gpt-oss-20b (22B params, 3.6B active - MoE)
  • Fine-tuned Model: GhostScientist/gpt-oss-20b-semanticwiki
  • Dataset: GhostScientist/semanticwiki-data
  • Training: SFT with LoRA using TRL
Downloads last month
104
GGUF
Model size
21B params
Architecture
gpt-oss
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for GhostScientist/gpt-oss-20b-semanticwiki-gguf

Quantized
(192)
this model

Dataset used to train GhostScientist/gpt-oss-20b-semanticwiki-gguf