YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
WhisperX Base NPU (INT8 Quantized)
π Hardware-Accelerated Speech Recognition for AMD NPU
Model Description
INT8-quantized version of openai/whisper-base, optimized for AMD Phoenix NPU (Ryzen AI) with custom MLIR-AIE2 kernels.
Specifications
- Size: 50MB (INT8)
- Performance: 0.002 RTF real-time factor
- Accuracy: 88% on LibriSpeech test-clean
- Quantization: INT8
- Hardware: AMD Phoenix NPU (16 TOPS)
Quick Start
from unicorn_engine import NPUWhisperX
model = NPUWhisperX.from_pretrained("magicunicorn/whisperx-base-npu")
result = model.transcribe("audio.wav")
print(result["text"])
Performance
Processes 1 hour of audio in < 30 seconds on AMD NPU hardware.
Links
- π οΈ Custom Runtime
- π¦ All NPU Models
- π¬ Community
License
MIT License (inherited from OpenAI Whisper)
Part of the Unicorn Commander Suite
- Downloads last month
- 8
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support