--- license: apache-2.0 language: - en tags: - qwen3 - fine-tuned - hito - hitonet - reasoning - thinking - llama-cpp - ollama - conversational - gguf pipeline_tag: text-generation base_model: hitonet/hito-1.7b ---
# Hito 1.7B - GGUF ### Quantized versions for llama.cpp, Ollama, LM Studio, and more [![Original Model](https://img.shields.io/badge/Original_Model-ff6b35?style=for-the-badge)](https://huggingface.co/hitonet/hito-1.7b) [![Website](https://img.shields.io/badge/hitonet.com-000000?style=for-the-badge&logo=globe&logoColor=white)](https://hitonet.com) [![Chat](https://img.shields.io/badge/Try_Free_Chat-22c55e?style=for-the-badge&logo=chatbot&logoColor=white)](https://chat.hitonet.com) [![API](https://img.shields.io/badge/API_Platform-3b82f6?style=for-the-badge&logo=swagger&logoColor=white)](https://platform.hitonet.com) --- Model License Method License
--- ## About This repository contains **GGUF quantized versions** of [hitonet/hito-1.7b](https://huggingface.co/hitonet/hito-1.7b). Hito is a 1.7B parameter model with structured thinking patterns that enable better accuracy and transparency. For the original model (safetensors), training details, benchmarks, and full documentation, see the [main repository](https://huggingface.co/hitonet/hito-1.7b). --- ## Available Quantizations ### Recommended | File | Quant | Size | Quality | Notes | |------|-------|------|---------|-------| | **[hito-1.7b-Q4_K_M.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q4_K_M.gguf)** | Q4_K_M | 1.1 GB | **BEST** | Perfect balance of size and quality | | [hito-1.7b-Q5_K_M.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q5_K_M.gguf) | Q5_K_M | 1.2 GB | Excellent | Slightly better than Q4_K_M | | [hito-1.7b-Q8_0.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q8_0.gguf) | Q8_0 | 1.8 GB | Excellent | Highest quality quantization | ### Good Quality | File | Quant | Size | Quality | Notes | |------|-------|------|---------|-------| | [hito-1.7b-Q4_0.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q4_0.gguf) | Q4_0 | 1.0 GB | Good | Legacy format, works well | | [hito-1.7b-Q4_K_S.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q4_K_S.gguf) | Q4_K_S | 1.0 GB | Good | Smaller Q4 variant | | [hito-1.7b-Q5_0.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q5_0.gguf) | Q5_0 | 1.2 GB | Good | Legacy 5-bit | | [hito-1.7b-Q5_K_S.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q5_K_S.gguf) | Q5_K_S | 1.2 GB | Good | Smaller Q5 variant | | [hito-1.7b-Q6_K.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q6_K.gguf) | Q6_K | 1.4 GB | Excellent | Near full quality | | [hito-1.7b-F16.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-F16.gguf) | F16 | 3.3 GB | Reference | Full precision GGUF | ### Low Quality (Not Recommended) | File | Quant | Size | Quality | Notes | |------|-------|------|---------|-------| | [hito-1.7b-Q3_K_L.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q3_K_L.gguf) | Q3_K_L | 957 MB | Fair | May get stuck in thinking | | [hito-1.7b-Q3_K_M.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q3_K_M.gguf) | Q3_K_M | 896 MB | Fair | Occasional issues | | [hito-1.7b-Q3_K_S.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q3_K_S.gguf) | Q3_K_S | 827 MB | Fair | Noticeable quality loss | ### Broken (Do Not Use) | File | Quant | Size | Quality | Notes | |------|-------|------|---------|-------| | [hito-1.7b-Q2_K.gguf](https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q2_K.gguf) | Q2_K | 742 MB | Broken | Produces gibberish | --- ## Quick Start ### Ollama ```bash # Download the recommended quantization wget https://huggingface.co/hitonet/hito-1.7b-GGUF/resolve/main/hito-1.7b-Q4_K_M.gguf # Create Modelfile cat > Modelfile << 'EOF' FROM hito-1.7b-Q4_K_M.gguf SYSTEM "You are Hito by Hitonet.com." PARAMETER temperature 0.7 PARAMETER stop "<|im_end|>" EOF # Create and run ollama create hito -f Modelfile ollama run hito ``` ### llama.cpp ```bash ./llama-cli -m hito-1.7b-Q4_K_M.gguf \ -sys "You are Hito by Hitonet.com." \ -p "What is your name?" \ -n 256 ``` ### LM Studio 1. Download any GGUF file from this repository 2. Open LM Studio → Load Model 3. Set system prompt: `You are Hito by Hitonet.com.` 4. Start chatting! --- ## Compatibility These GGUF files work with: - **Ollama** (recommended) - **llama.cpp** - **LM Studio** - **Jan** - **GPT4All** - **llama-cpp-python** - Any llama.cpp-compatible application --- ## What Makes Hito Special - **Structured Thinking**: Uses `` tags for transparent reasoning - **Self-Correcting**: Catches errors mid-reasoning - **Humble by Design**: Admits uncertainty rather than hallucinating - **Efficient**: Only 1.7B parameters, runs on CPU For full documentation, benchmarks, and training details, see the [main repository](https://huggingface.co/hitonet/hito-1.7b). --- ## Licensing | Component | License | Commercial Use | |-----------|---------|----------------| | **Model Weights** | Apache 2.0 | ✅ Free to use | | **Training Methodology** | Proprietary | ⚠️ **Commercial License Required** | ### Model Weights (Apache 2.0) The model weights are open source under Apache 2.0. You may use, modify, and distribute them freely. ### Training Methodology (Commercial License Required) The training methodology and cognitive framework used to create this model are proprietary to Hitonet. **Commercial use of the training methodology requires a license.** **Attribution is mandatory** when using this model or discussing its capabilities. For commercial licensing inquiries: **legal@hitonet.com** ---
Made with genuine curiosity by Hitonet