HyzeMini (GGUF)
Lightweight GGUF builds of HyzeMini for fast local inference
π hyzebot.vercel.app β’ π hyzedocs.vercel.app β’ π§ hyzecode.vercel.app
π Overview
HyzeMini (GGUF) provides quantized GGUF versions of the HyzeMini model, optimized for local execution using tools like llama.cpp, LM Studio, Ollama, and other GGUF-compatible runtimes.
This version keeps the same Space + General Chat focus, while enabling:
- β‘ Faster inference
- π§ Lower memory usage
- π» CPU-friendly execution
π§ Model Details
- Base model: HyzeAI / HyzeMini
- Parameters: ~0.1B
- Architecture: Transformer (LLaMA-style)
- Format: GGUF
- Language: English
- License: Apache-2.0
π§ͺ Available Quantizations
(Exact files may vary depending on upload)
Common GGUF variants include:
Q2_Kβ Ultra-low memory, fastestQ4_K_Mβ Balanced quality & speed (recommended)Q5_K_Mβ Higher quality, slightly slowerQ8_0β Best quality, highest memory usage
π‘ If youβre unsure, start with Q4_K_M.
βοΈ Usage
llama.cpp
./main -m HyzeMini-Q4_K_M.gguf -p "Tell me a cool space fact:"
---
- Downloads last month
- 70
Hardware compatibility
Log In
to add your hardware
We're not able to determine the quantization variants.
Model tree for HyzeAI/HyzeMiniGGUF
Base model
HyzeAI/HyzeMini