Spaces:
Running
Running
metadata
title: clip-tagger
emoji: π΅
colorFrom: blue
colorTo: purple
sdk: static
pinned: false
license: mit
π΅ clip-tagger
π Live Demo β’ π¦ Model Repository
Custom audio tagging in the browser using CLAP (Contrastive Language-Audio Pre-training)
Instantly tag any audio with AI that learns from your corrections. Upload files or record directly in your browser - everything runs locally, no servers needed.
β¨ Features
- π€ Audio Input: Upload files or record directly from your microphone
- π§ Smart Tagging: CLAP model identifies speech, music, ambient sounds, and more
- π Personalized Learning: Correct tags and add custom ones - the model adapts to your domain
- πΎ Persistent Memory: Your corrections are saved and improve future predictions
- π Export Ready: Export tagged data and trained models for sharing
- π Privacy First: Everything runs in your browser - no data leaves your device
π How It Works
- Drop an audio file or click record
- Review AI-generated tags with confidence scores
- Correct tags with β/β buttons or add custom tags
- Watch the model learn from your feedback in real-time
- Export results or share your trained model
π§ Technical Details
- Model: Xenova/clap-htsat-unfused (~45MB)
- Framework: Transformers.js + React
- Storage: IndexedDB for user feedback and model weights
- Deployment: Ready for Hugging Face Spaces
π― Use Cases
- Voice memo organization
- Music library tagging
- Audio content moderation
- Podcast categorization
- Sound effect libraries
- Research datasets
Powered by Transformers.js β’ Runs entirely in your browser