tasty-musicgen-small

License: CC BY 4.0 arXiv

tasty-musicgen-small is a musicgen-small fine-tuned on a patched version of the Taste & Affect Music Database. It generates music that's supposed to induce gustatory synesthesia perceptions based on multimodal research. It generates mono audio in 32khz.

Code and Dataset

Code and the dataset used to train this model are available at: https://osf.io/xs5jy/.

How to use

Here is a showcase on how to use the model with the transformer library, it is also possible to make inference with the audiocraft library, for a detailed explanation we suggest to read the official MusicGEN guide by Hugging Face

from transformers import pipeline
import scipy

synthesiser = pipeline("text-to-audio", "csc-unipd/tasty-musicgen-small")

music = synthesiser("sweet music for fine restaurents", forward_params={"do_sample": True})

scipy.io.wavfile.write("musicgen_out.wav", rate=music["sampling_rate"], data=music["audio"])

Citation

If you use this model, code or the data in your research, please cite the following article:

@article{10.3389/fcomp.2025.1575741,
  author={Spanio, Matteo  and Zampini, Massimiliano  and Rodà, Antonio  and Pierucci, Franco }, 
  title={A multimodal symphony: integrating taste and sound through generative AI},    
  journal={Frontiers in Computer Science},   
  volume={Volume 7 - 2025},
  year={2025},
  url={https://www.frontiersin.org/journals/computer-science/articles/10.3389/fcomp.2025.1575741},
  doi={10.3389/fcomp.2025.1575741},
  issn={2624-9898},
  abstract={In recent decades, neuroscientific and psychological research has identified direct relationships between taste and auditory perception. This article explores multimodal generative models capable of converting taste information into music, building on this foundational research. We provide a brief review of the state of the art in this field, highlighting key findings and methodologies. We present an experiment in which a fine-tuned version of a generative music model (MusicGEN) is used to generate music based on detailed taste descriptions provided for each musical piece. The results are promising: according to the participants' evaluations (n = 111), the fine-tuned model produces music that more coherently reflects the input taste descriptions compared to the non-fine-tuned model. This study represents a significant step toward understanding and developing embodied interactions between AI, sound, and taste, opening new possibilities in the field of generative AI.}
}
Downloads last month
7
Safetensors
Model size
591M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for csc-unipd/tasty-musicgen-small

Finetuned
(14)
this model