Skip to content

Commit 90bf0ec

Browse files
author
蒄骰
committed
Merge branch 'main' of github.com:ddlBoJack/SLAM-LLM into ygr_pr2
2 parents 819ce1f + f32b8a2 commit 90bf0ec

File tree

6 files changed

+2068
-13
lines changed

6 files changed

+2068
-13
lines changed

README.md

Lines changed: 45 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,8 @@ developers to train custom multimodal large language model (MLLM), focusing on <
2828
6. [Citation](#citation)
2929

3030
# News
31-
- [Update Oct. 12, 2024] Recipes for [SLAM-AAC](examples/slam_aac/README.md) have been supported.
31+
- [Update Nov. 5, 2024] Recipes for [speech emotion captioning (SEC)](examples/sec_emotioncaps/README.md) with [emotion2vec](https://github.com/ddlBoJack/emotion2vec) as the encoder has been supported.
32+
- [Update Oct. 12, 2024] Recipes for [SLAM-AAC](examples/slam_aac/README.md) with [EAT](https://github.com/cwx-worst-one/EAT) as the encoder have been supported.
3233
- [Update Sep. 28, 2024] Recipes for [CoT-ST](examples/st_covost2/README.md) have been supported.
3334
- [Update Sep. 25, 2024] Recipes for [DRCap](examples/drcap_zeroshot_aac/README.md) have been supported.
3435
- [Update Jun. 12, 2024] Recipes for [MaLa-ASR](examples/mala_asr_slidespeech/README.md) have been supported.
@@ -90,6 +91,7 @@ We provide reference implementations of various LLM-based speech, audio, and mus
9091

9192
- Text-to-Speech (TTS)
9293
- [VALL-E-X](examples/vallex/README.md)
94+
- [Speech Emotion Captioning (SEC)](examples/sec_emotioncaps/README.md)
9395

9496
- **Audio Task**
9597
- [Automated Audio Captioning (AAC)](examples/aac_audiocaps/README.md)
@@ -118,7 +120,10 @@ command-line (shell file) > Hydra configuration (yaml file) > dataclass configur
118120
- We borrow code from [Fairseq](https://github.com/facebookresearch/fairseq) for deepspeed configuration.
119121
- We thank the contributors for providing diverse recipes.
120122

121-
## Citation
123+
# Citation
124+
125+
## Speech Task
126+
122127
SLAM-ASR:
123128
```
124129
@article{ma2024embarrassingly,
@@ -128,7 +133,27 @@ SLAM-ASR:
128133
year={2024}
129134
}
130135
```
136+
Mala-ASR:
137+
```
138+
@article{yang2024mala,
139+
title={MaLa-ASR: Multimedia-Assisted LLM-Based ASR},
140+
author={Yang, Guanrou and Ma, Ziyang and Yu, Fan and Gao, Zhifu and Zhang, Shiliang and Chen, Xie},
141+
journal={Proc. INTERSPEECH},
142+
year={2024}
143+
}
144+
```
145+
CoT-ST:
146+
```
147+
@article{du2024cot,
148+
title={CoT-ST: Enhancing LLM-based Speech Translation with Multimodal Chain-of-Thought},
149+
author={Du, Yexing and Ma, Ziyang and Yang, Yifan and Deng, Keqi and Chen, Xie and Yang, Bo and Xiang, Yang and Liu, Ming and Qin, Bing},
150+
journal={arXiv preprint arXiv:2409.19510},
151+
year={2024}
152+
}
153+
```
131154

155+
156+
## Audio Task
132157
SLAM-AAC:
133158
```
134159
@article{chen2024slam,
@@ -138,5 +163,21 @@ SLAM-AAC:
138163
year={2024}
139164
}
140165
```
141-
142-
166+
DRCap:
167+
```
168+
@article{li2024drcap,
169+
title={DRCap: Decoding CLAP Latents with Retrieval-augmented Generation for Zero-shot Audio Captioning},
170+
author={Li, Xiquan and Chen, Wenxi and Ma, Ziyang and Xu, Xuenan and Liang, Yuzhe and Zheng, Zhisheng and Kong, Qiuqiang and Chen, Xie},
171+
journal={arXiv preprint arXiv:2410.09472},
172+
year={2024}
173+
}
174+
```
175+
BAT:
176+
```
177+
@article{zheng2024bat,
178+
title={BAT: Learning to Reason about Spatial Sounds with Large Language Models},
179+
author={Zheng, Zhisheng and Peng, Puyuan and Ma, Ziyang and Chen, Xie and Choi, Eunsol and Harwath, David},
180+
journal={Proc. ICML},
181+
year={2024}
182+
}
183+
```

examples/drcap_zeroshot_aac/README.md

Lines changed: 18 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
# DRCap_Zeroshot_Audio_Captioning
22

33
## Introduction
4-
DRCap is a data-efficient and flexible audio captioning system requiring text-only data for training and can quickly adapt to new domains without additional fine-tuning.
4+
[DRCap](https://www.arxiv.org/abs/2410.09472) is a data-efficient and flexible audio captioning system requiring text-only data for training and can quickly adapt to new domains without additional fine-tuning. It uses projection decoding and retrieval-augmented generation to perform zero-shot audio captioning.
55

66
![](assets/model.png)
77

@@ -14,7 +14,7 @@ You could download our pretrained CLAP model and linear mapping network through
1414
* LLM [vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-7b-v1.5)
1515

1616
## Inference
17-
You could modify the variables `run_dir`, `audio_encoder_dir`, `output_dir`, `llm_path` in `scripts/inference_drcap.sh` to match the paths where the downloaded checkpoints are located. Additionally, update the `source` in `data/audiocaps_test.jsonl` to ensure the audio paths point to your audio files, and then run:
17+
You could modify the variables `run_dir`, `audio_encoder_dir`, `output_dir`, `llm_path` in `scripts/inference_drcap.sh` to match the paths where the downloaded checkpoints are located. Additionally, update the `source` in `data_examples/audiocaps_test.jsonl` to ensure the audio paths point to your audio files, and then run:
1818

1919
```shell
2020
bash scripts/inference_drcap.sh
@@ -24,10 +24,10 @@ bash scripts/inference_drcap.sh
2424
## Data preparation
2525
Prepare your `jsonl` data file in the following format:
2626
```json
27-
{"key": "Y7fmOlUlwoNg_1", "target": "Constant rattling noise and sharp vibrations", "text": "Constant rattling noise and sharp vibrations"}
28-
{"key": "Y6BJ455B1aAs_1", "target": "A rocket flies by followed by a loud explosion and fire crackling as a truck engine runs idle", "text": "A rocket flies by followed by a loud explosion and fire crackling as a truck engine runs idle"}
27+
{"key": "Y7fmOlUlwoNg_1", "target": "Constant rattling noise and sharp vibrations", "text": "Constant rattling noise and sharp vibrations", "similar_captions": ["The engine of a small machine pulling chains", "A market vendor is producing a rhythmic sound with metal forceps.", "A masonry machine is in operation at a fair."]}
28+
{"key": "Y6BJ455B1aAs_1", "target": "A rocket flies by followed by a loud explosion and fire crackling as a truck engine runs idle", "text": "A rocket flies by followed by a loud explosion and fire crackling as a truck engine runs idle", "similar_captions": ["An engine is revving, with fire and an explosion.", "An explosion is heard after an engine cuts out.", "A car speeding past with a large boom"]}
2929
```
30-
Please note that only textual data is required for training. However, for zero-shot inference, audio files are also necessary. You could find an example of the jsonl file in `data/audiocaps_test.jsonl`
30+
Please note that only textual data is required for training. However, for zero-shot inference, audio files are also necessary. You could find an example of the jsonl file in `data_examples/audiocaps_test.jsonl`
3131

3232
Run the following command to do the retrieval-augmentation and create the text embedding support for evaluation:
3333
```shell
@@ -42,4 +42,16 @@ bash scripts/finetune_drcap.sh
4242
For training only the linear layer (without using LoRA or other PEFT methods), you can set the following parameters: `use_peft=false` and `freeze_llm=true`. To turn off the RAG, you could set `use_arg=false` and `rag_first=false`
4343

4444
## Acknowledgement
45-
The code of training the CLAP model is based on the [WavCaps](https://github.com/XinhaoMei/WavCaps) repo, we thank the contributors for open-sourcing their work.
45+
The code of training the CLAP model is based on the [WavCaps](https://github.com/XinhaoMei/WavCaps) repo, we thank the contributors for open-sourcing their work.
46+
47+
48+
## Citation
49+
You can refer to our paper for more results
50+
```
51+
@article{li2024drcap,
52+
title={DRCap: Decoding CLAP Latents with Retrieval-augmented Generation for Zero-shot Audio Captioning},
53+
author={Li, Xiquan and Chen, Wenxi and Ma, Ziyang and Xu, Xuenan and Liang, Yuzhe and Zheng, Zhisheng and Kong, Qiuqiang and Chen, Xie},
54+
journal={arXiv preprint arXiv:2410.09472},
55+
year={2024}
56+
}
57+
```

0 commit comments

Comments
 (0)