YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

WhisperX Base NPU (INT8 Quantized)

πŸš€ Hardware-Accelerated Speech Recognition for AMD NPU

Model Description

INT8-quantized version of openai/whisper-base, optimized for AMD Phoenix NPU (Ryzen AI) with custom MLIR-AIE2 kernels.

Specifications

  • Size: 50MB (INT8)
  • Performance: 0.002 RTF real-time factor
  • Accuracy: 88% on LibriSpeech test-clean
  • Quantization: INT8
  • Hardware: AMD Phoenix NPU (16 TOPS)

Quick Start

from unicorn_engine import NPUWhisperX

model = NPUWhisperX.from_pretrained("magicunicorn/whisperx-base-npu")
result = model.transcribe("audio.wav")
print(result["text"])

Performance

Processes 1 hour of audio in < 30 seconds on AMD NPU hardware.

Links

License

MIT License (inherited from OpenAI Whisper)


Part of the Unicorn Commander Suite

Downloads last month
8
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support