Model Gallery

31 models from 1 repositories

Filter by type:

Filter by tags:

omnilingual-0.3b-ctc-q8-sherpa
Omnilingual ASR CTC 300M (int8) is a multilingual automatic speech recognition model supporting 1,600+ languages. Based on Meta's omniASR_CTC_300M architecture (Wav2Vec2 with CTC head), quantized to int8 for efficient inference. Uses the sherpa-onnx backend with ONNX Runtime.

Repository: localaiLicense: apache-2.0

streaming-zipformer-en-sherpa
Streaming English ASR: sherpa-onnx zipformer transducer (int8, chunk-16 left-128). Low-latency real-time transcription with endpoint detection via sherpa-onnx's online recognizer. English-only; for multilingual offline ASR see omnilingual-0.3b-ctc-q8-sherpa.

Repository: localaiLicense: apache-2.0

vibevoice-cpp
VibeVoice Realtime 0.5B (C++ / GGML, Q8_0) - native C++ port of Microsoft VibeVoice via the vibevoice-cpp backend. 24kHz mono TTS with voice cloning from a single reference voice prompt. Default voice prompt: en-Carter_man.

Repository: localaiLicense: mit

vibevoice-cpp-asr
VibeVoice ASR 7B (C++ / GGML, Q4_K) - long-form speech-to-text with speaker diarization. Returns per-speaker JSON segments with start/end timestamps. English-only. ~10 GB download.

Repository: localaiLicense: mit

qwen3-asr-1.7b
Qwen3-ASR is an automatic speech recognition model supporting multiple languages and batch inference.

Repository: localaiLicense: apache-2.0

qwen3-asr-0.6b
Qwen3-ASR is an automatic speech recognition model supporting multiple languages and batch inference.

Repository: localaiLicense: apache-2.0

qwen3-asr-0.6b
Qwen3-ASR is an automatic speech recognition model supporting multiple languages and batch inference.

Repository: localaiLicense: apache-2.0

qwen3-asr-1.7b
Qwen3-ASR is an automatic speech recognition model supporting multiple languages and batch inference.

Repository: localaiLicense: apache-2.0

whisper-1
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-base-q5_1
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-base
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-base-en-q5_1
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-base-en
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-large-q5_0
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-medium
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-medium-q5_0
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-small-q5_1
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-small
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-small-en-q5_1
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-small-en
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

whisper-small-q5_1
Port of OpenAI's Whisper model in C/C++

Repository: localaiLicense: mit

Page 1