--- license: apache-2.0 language: - en tags: - qwen3 - fine-tuned - hito - hitonet - reasoning - conversational - thinking - adaptive-reasoning - tree-of-thought - hierarchical-reasoning - cognitive-framework - self-aware-ai - anti-hallucination - synthetic-data - llama-cpp - ollama - text-generation-inference pipeline_tag: text-generation base_model: Qwen/Qwen3-1.7B --- # Hito 1.7B ### Nested Cognitive Reasoning for Thoughtful AI

Hitonet

GGUF DownloadsWeb ChatAPIResearch Paper

--- | Status | Parameters | Context | License | |--------|-----------|---------|---------| | Production | 1.7B | 32K | Apache 2.0 | --- ## Overview Hito is a 1.7B parameter language model fine-tuned with **Nested Cognitive Reasoning (NCR)** - a novel architecture that enables structured, self-correcting thinking patterns. Unlike traditional models that produce linear outputs, Hito thinks in branching, hierarchical structures that mirror human cognition. ### Key Features - **Structured Reasoning**: Uses cognitive tags (``, ``, ``) for transparent thought processes - **Self-Correction**: Built-in mechanisms to catch and correct errors mid-reasoning - **Humble AI**: Acknowledges uncertainty and limitations rather than hallucinating - **Efficient**: Runs on consumer hardware with GGUF quantizations available --- ## Benchmark Results | Model | Params | Overall | Counting | Math | Cognitive Bias | |-------|--------|---------|----------|------|----------------| | GPT-5-mini | ~8B | **100%** | 100% | 100% | ✅ | | Claude Haiku 4.5 | ~8B | 90% | 67% | 100% | ✅ | | **Hito 1.7B** | **1.7B** | **80%** | **67%** | **100%** | **✅** | | GPT-4o-mini | ~8B | 80% | 33% | 100% | ❌ | | Qwen3 1.7B (base) | 1.7B | 17% | 0% | 17% | ❌ | ### The Bat and Ball Test *"A bat and a ball cost $1.10 together. The bat costs $1.00 more than the ball. How much does the ball cost?"* Most AI models (and humans) answer 10 cents. **That's wrong.** | Model | Answer | Correct | |-------|--------|---------| | **Hito 1.7B** | **$0.05** | ✅ | | Qwen3 1.7B (base) | $0.10 | ❌ | | GPT-4o-mini | $0.10 | ❌ | **Why Hito gets it right:** ```xml Ball + Bat = $1.10, Bat = Ball + $1.00 Intuition says 10 cents... but let me verify. If ball = $0.10, bat = $1.10, total = $1.20. WRONG. Let ball = x: x + (x + 1) = 1.10, 2x = 0.10, x = 0.05 Ball $0.05 + Bat $1.05 = $1.10 ✓ The ball costs five cents. ``` --- ## Cognitive Architecture Hito uses a tree-structured reasoning system with four cognitive states: | State | Focus | Tags Used | |-------|-------|-----------| | **Analytical** | Logic, accuracy | ``, ``, `` | | **Creative** | Imagination, exploration | ``, ``, `` | | **Empathetic** | Feelings, perspectives | ``, ``, `` | | **Reflective** | Depth, meaning | ``, ``, `` | ### The Humble Tags | Tag | Purpose | |-----|---------| | `` | Question assumptions | | `` | Admit errors | | `` | Acknowledge knowledge gaps | | `` | Rate certainty level | | `` | Double-check work | --- ## Quick Start ### Python (Transformers) ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("hitonet/hito-1.7b", torch_dtype="auto", device_map="auto") tokenizer = AutoTokenizer.from_pretrained("hitonet/hito-1.7b") messages = [ {"role": "system", "content": "You are Hito by Hitonet.com."}, {"role": "user", "content": "What is 15% of 200?"} ] inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True).to(model.device) outputs = model.generate(inputs, max_new_tokens=512, temperature=0.7, do_sample=True) print(tokenizer.decode(outputs[0], skip_special_tokens=False)) ``` ### Ollama ```bash # Download GGUF from hitonet/hito-1.7b-GGUF ollama create hito -f Modelfile ollama run hito ``` ### API ```bash curl https://api.hitonet.com/v1/chat/completions \ -H "Authorization: Bearer YOUR_API_KEY" \ -H "Content-Type: application/json" \ -d '{"model": "hito", "messages": [{"role": "user", "content": "Hello!"}]}' ``` Try the full API at [platform.hitonet.com](https://platform.hitonet.com) - $1 free credit included. --- ## Model Variants | Repository | Format | Use Case | |------------|--------|----------| | [hitonet/hito-1.7b](https://huggingface.co/hitonet/hito-1.7b) | Safetensors | Python/Transformers | | [hitonet/hito-1.7b-GGUF](https://huggingface.co/hitonet/hito-1.7b-GGUF) | GGUF | Ollama/llama.cpp/LM Studio | ### Recommended GGUF Quantizations | Quantization | Size | Quality | Use Case | |--------------|------|---------|----------| | Q4_K_M | 1.1 GB | ⭐ Best Balance | Most users | | Q5_K_M | 1.2 GB | Excellent | Quality-focused | | Q8_0 | 1.8 GB | Highest | Maximum quality | --- ## Training Hito is fine-tuned from Qwen3-1.7B using Supervised Fine-Tuning (SFT) with synthetic data generated by our flagship Hito-Genius model. The training focuses on: - **Cognitive Pattern Transfer**: Teaching structured reasoning through demonstration - **Self-Correction Habits**: Training the model to verify its own work - **Humility Patterns**: Learning to express uncertainty appropriately --- ## Research For technical details on Nested Cognitive Reasoning, see our research paper: **[Nested Cognitive Reasoning: A Tree-Structured Approach to Language Model Thinking](https://hitonet.com/research)** *Hitonet Research, 2025* --- ## Licensing | Component | License | Commercial Use | |-----------|---------|----------------| | **Model Weights** | Apache 2.0 | ✅ Free | | **NCR Methodology** | CC BY-NC-ND | ⚠️ License Required | The model weights are fully open source under Apache 2.0. The Nested Cognitive Reasoning methodology (cognitive tags, tree-structured thinking, humble tags system) is protected under CC BY-NC-ND. Commercial use of the NCR method requires a license. **Contact:** legal@hitonet.com --- ## Links - **Website:** [hitonet.com](https://hitonet.com) - **Chat:** [chat.hitonet.com](https://chat.hitonet.com) - **API:** [platform.hitonet.com](https://platform.hitonet.com) - **Research:** [hitonet.com/research](https://hitonet.com/research) - **Blog:** [hitonet.com/blog](https://hitonet.com/blog) ---

Made with genuine curiosity by Hitonet
Teaching AI to think, doubt, and learn.