Moonshine Base -- GGUF

GGUF conversions and quantisations of UsefulSensors/moonshine-base for use with CrispStrobe/CrispASR.

Available variants

File Quant Size Notes
moonshine-base.gguf F32 235 MB Full precision
moonshine-base-q4_k.gguf Q4_K 40 MB Best size/quality tradeoff

Model details

  • Architecture: Conv1d stem + 8L transformer encoder + 8L transformer decoder (416d, 8 heads, partial RoPE, SiLU/GELU)
  • Parameters: 61M
  • Languages: English only
  • License: MIT
  • Source: UsefulSensors/moonshine-base

Usage with CrispASR

# Auto-download (English tiny only)
./build/bin/crispasr --backend moonshine -m auto -f audio.wav

# Explicit model path
./build/bin/crispasr --backend moonshine -m moonshine-base-q4_k.gguf -f audio.wav

Notes

  • Moonshine models run on CPU only (GPU not needed for these small models)
  • Tokenizer (tokenizer.bin) must be in the same directory as the model file
  • Base models use head_dim=52 which works on CPU flash_attn
Downloads last month
207
GGUF
Model size
61.5M params
Architecture
moonshine
Hardware compatibility
Log In to add your hardware

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for cstr/moonshine-base-GGUF

Quantized
(2)
this model