STT/ASR - GGUF
Collection
Multilingual quantized models in GGUF format • 8 items • Updated
This model is a quantized gguf-format export of Qwen/Qwen3-ForcedAligner-0.6B for ease of use in edge devices and CPU-based inference environments. The original model is transformed into gguf with F16 tensors by the script convert_hf_to_gguf.py and then further quantized, if needed, using the tool quantize from the same repo.
The requirements can be installed as
$ pip install git+https://github.com/femelo/py-qwen3-asr-cpp
from py_qwen3_asr_cpp.model import Qwen3ASRModel
# Initialize the model (it handles downloading from this repo)
model = Qwen3ASRModel(
asr_model="qwen3-asr-0.6b-f16",
align_model="qwen3-forced-aligner-0.6b-f16",
n_threads=4
)
# Transcribe from file
result, alignment = model.transcribe_and_align("audio.mp3")
print(f"Detected Language: {result.language}")
print(f"Transcription: {result.text}")
Refer to https://github.com/femelo/py-qwen3-asr-cpp for more details.
The license is derived from the original model: Apache 2.0. For more details, please refer to Qwen/Qwen3-ForcedAligner-0.6B.
16-bit