MusiLingo is a novel system for music caption generation and music-related query responses, bridging the gap between music audio and textual contexts and creating the MusicInstruct dataset from captions in the MusicCaps datasets, tailored for open-ended music inquiries.
Large Language Models (LLMs) have shown immense potential in multimodal applications, yet the convergence of textual and musical domains remains not well-explored. To address this gap, we present MusiLingo, a novel system for music caption generation and music-related query responses. MusiLingo employs a single projection layer to align music representations from the pre-trained frozen music audio model MERT with a frozen LLM, bridging the gap between music audio and textual contexts. We train it on an extensive music caption dataset and fine-tune it with instructional data. Due to the scarcity of high-quality music Q&A datasets, we created the MusicInstruct (MI) dataset from captions in the MusicCaps datasets, tailored for open-ended music inquiries. Empirical evaluations demonstrate its competitive performance in generating music captions and composing music-related Q&A pairs. Our introduced dataset enables notable advancements beyond previous ones.
Ge Zhang
6 papers
Zihao Deng
1 papers
Yi Ma
2 papers
Yudong Liu
1 papers
Rongchen Guo
1 papers