license: apache-2.0

OVOS - Qwen3 Forced Aligner 0.6B F16 (GGUF)

This model is a quantized gguf-format export of Qwen/Qwen3-ForcedAligner-0.6B for ease of use in edge devices and CPU-based inference environments. The original model is transformed into gguf with F16 tensors by the script convert_hf_to_gguf.py and then further quantized, if needed, using the tool quantize from the same repo.

Requirements

The requirements can be installed as

$ pip install git+https://github.com/femelo/py-qwen3-asr-cpp

Usage

from py_qwen3_asr_cpp.model import Qwen3ASRModel

# Initialize the model (it handles downloading from this repo)
model = Qwen3ASRModel(
    asr_model="qwen3-asr-0.6b-f16",
    align_model="qwen3-forced-aligner-0.6b-f16",
    n_threads=4
)

# Transcribe from file
result, alignment = model.transcribe_and_align("audio.mp3")
print(f"Detected Language: {result.language}")
print(f"Transcription: {result.text}")

Refer to https://github.com/femelo/py-qwen3-asr-cpp for more details.

Licensing

The license is derived from the original model: Apache 2.0. For more details, please refer to Qwen/Qwen3-ForcedAligner-0.6B.

Downloads last month
42
GGUF
Model size
0.9B params
Architecture
qwen3-asr
Hardware compatibility
Log In to add your hardware

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including OpenVoiceOS/qwen3-forced-aligner-0.6b-f16