| --- |
| license: mit |
| library_name: transformers |
| tags: |
| - mlx |
| - open4bits |
| base_model: deepseek-ai/DeepSeek-R1 |
| pipeline_tag: text-generation |
| --- |
| |
| # Open4bits / DeepSeek-R1-MLX-2Bit |
|
|
| This repository provides the **DeepSeek-R1 model quantized to 2-bit in MLX format**, published by Open4bits to enable highly efficient local inference with minimal memory usage and broad hardware compatibility. |
|
|
| The underlying DeepSeek-R1 model and architecture are **developed and owned by DeepSeek AI**. This repository contains only a 2-bit quantized MLX conversion of the original model weights. |
|
|
| The model is designed for lightweight, high-performance text generation and instruction-following tasks, making it well suited for resource-constrained and local deployments. |
|
|
| --- |
|
|
| ## Model Overview |
|
|
| DeepSeek-R1 is a transformer-based large language model developed for strong general language understanding and generation. |
| This release provides a **2-bit quantized checkpoint in MLX format**, enabling efficient inference on CPUs and supported accelerators with reduced memory footprint. |
|
|
| Open4bits has started supporting **MLX models** to broaden compatibility with emerging quantization formats and efficient runtimes. |
|
|
| --- |
|
|
| ## Model Details |
|
|
| * **Base Model:** DeepSeek-R1 |
| * **Quantization:** 2-bit |
| * **Format:** MLX |
| * **Task:** Text generation, instruction following |
| * **Weight tying:** Preserved |
| * **Compatibility:** MLX-enabled inference engines and efficient runtimes |
|
|
| This quantized release is designed to balance strong generation performance with low resource requirements. |
|
|
| --- |
|
|
| ## Intended Use |
|
|
| This model is intended for: |
|
|
| * Local text generation and conversational applications |
| * CPU-based or low-resource deployments |
| * Research, prototyping, and experimentation |
| * Self-hosted or offline AI systems |
|
|
| --- |
|
|
| ## Limitations |
|
|
| * Reduced performance compared to full-precision variants |
| * Output quality depends on prompt design and inference settings |
| * Not specifically tuned for highly specialized or domain-specific tasks |
|
|
| --- |
|
|
| ## License |
|
|
| This model follows the **MIT** as defined by the base model creators. |
| Users must comply with the licensing conditions of the base DeepSeek-R1 model. |
|
|
| --- |
|
|
| ## Support |
|
|
| If you find this model useful, please consider supporting the project. |
| Your support helps Open4bits continue releasing and maintaining high-quality, efficient open models for the community. |
|
|
|
|