Octen-Embedding-8B-mlx

Pre-converted MLX weights for Octen-Embedding-8B, ready to run on Apple Silicon.

Why this exists

The original model requires a ~30 minute conversion step and ~32GB temporary disk space. This repo provides the already-converted MLX weights so you can start embedding immediately.

Usage

With octen-embeddings-server:

# Clone the server
git clone https://github.com/c-h-/octen-embeddings-server.git
cd octen-embeddings-server
pip install -r requirements.txt

# Download pre-converted weights (instead of running convert_model.py)
huggingface-cli download chulcher/Octen-Embedding-8B-mlx --local-dir models/Octen-Embedding-8B-mlx

# Start the server
python3 server.py

The server exposes an OpenAI-compatible /v1/embeddings endpoint at http://localhost:8100.

Hardware Requirements

Component Requirement
CPU Apple Silicon (M1/M2/M3/M4)
RAM 20 GB+
Disk ~16 GB for weights
OS macOS 13+

Performance

Octen-Embedding-8B ranks #1 on MTEB/RTEB with a score of 0.8045, outperforming commercial embedding APIs.

Typical latency on Apple Silicon: ~50-200ms per text depending on length.

License

Apache 2.0 (same as base model)

Downloads last month
205
Safetensors
Model size
8B params
Tensor type
BF16
·
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for chulcher/Octen-Embedding-8B-mlx

Base model

Qwen/Qwen3-8B-Base
Finetuned
(1)
this model