File tree Expand file tree Collapse file tree 7 files changed +21
-0
lines changed
examples/slam_aac/scripts Expand file tree Collapse file tree 7 files changed +21
-0
lines changed Original file line number Diff line number Diff line change @@ -9,6 +9,8 @@ run_dir=/data/wenxi.chen/SLAM-LLM
99cd $run_dir
1010code_dir=examples/slam_aac
1111
12+ encoder_fairseq_dir=/fairseq/EAT # path to the fairseq directory of the encoder model
13+
1214audio_encoder_path=/data/xiquan.li/models/EAT-base_epoch30_ft.pt
1315llm_path=/data/xiquan.li/models/vicuna-7b-v1.5
1416
@@ -38,6 +40,7 @@ hydra.run.dir=$output_dir \
3840++model_config.encoder_path=$audio_encoder_path \
3941++model_config.encoder_dim=768 \
4042++model_config.encoder_projector=linear \
43+ ++model_config.encoder_fairseq_dir=$encoder_fairseq_dir \
4144++dataset_config.encoder_projector_ds_rate=${encoder_projector_ds_rate} \
4245++dataset_config.dataset=audio_dataset \
4346++dataset_config.train_data_path=$train_jsonl_path \
Original file line number Diff line number Diff line change @@ -9,6 +9,8 @@ run_dir=/data/wenxi.chen/SLAM-LLM
99cd $run_dir
1010code_dir=examples/slam_aac
1111
12+ encoder_fairseq_dir=/fairseq/EAT # path to the fairseq directory of the encoder model
13+
1214audio_encoder_path=/data/xiquan.li/models/EAT-base_epoch30_ft.pt
1315llm_path=/data/xiquan.li/models/vicuna-7b-v1.5
1416
@@ -38,6 +40,7 @@ hydra.run.dir=$output_dir \
3840++model_config.encoder_path=$audio_encoder_path \
3941++model_config.encoder_dim=768 \
4042++model_config.encoder_projector=linear \
43+ ++model_config.encoder_fairseq_dir=$encoder_fairseq_dir \
4144++dataset_config.encoder_projector_ds_rate=${encoder_projector_ds_rate} \
4245++dataset_config.dataset=audio_dataset \
4346++dataset_config.train_data_path=$train_jsonl_path \
Original file line number Diff line number Diff line change @@ -10,6 +10,8 @@ audio_encoder_path=/data/xiquan.li/models/EAT-base_epoch30_ft.pt
1010llm_path=/data/xiquan.li/models/vicuna-7b-v1.5
1111clap_dir=/data/xiquan.li/models/clap
1212
13+ encoder_fairseq_dir=/fairseq/EAT # path to the fairseq directory of the encoder model
14+
1315encoder_projector_ds_rate=5
1416
1517inference_data_path=/data/wenxi.chen/data/audiocaps/new_test.jsonl
@@ -41,6 +43,7 @@ for num_beams in "${beam_range[@]}"; do
4143 ++model_config.encoder_projector=linear \
4244 ++model_config.encoder_projector_ds_rate=$encoder_projector_ds_rate \
4345 ++model_config.normalize=true \
46+ ++model_config.encoder_fairseq_dir=$encoder_fairseq_dir \
4447 ++dataset_config.encoder_projector_ds_rate=$encoder_projector_ds_rate \
4548 ++dataset_config.dataset=audio_dataset \
4649 ++dataset_config.val_data_path=$inference_data_path \
Original file line number Diff line number Diff line change @@ -6,6 +6,8 @@ run_dir=/data/wenxi.chen/SLAM-LLM
66cd $run_dir
77code_dir=examples/slam_aac
88
9+ encoder_fairseq_dir=/fairseq/EAT # path to the fairseq directory of the encoder model
10+
911audio_encoder_path=/data/xiquan.li/models/EAT-base_epoch30_ft.pt
1012llm_path=/data/xiquan.li/models/vicuna-7b-v1.5
1113
@@ -31,6 +33,7 @@ python $code_dir/inference_aac_batch.py \
3133 ++model_config.encoder_projector=linear \
3234 ++model_config.encoder_projector_ds_rate=$encoder_projector_ds_rate \
3335 ++model_config.normalize=true \
36+ ++model_config.encoder_fairseq_dir=$encoder_fairseq_dir \
3437 ++dataset_config.encoder_projector_ds_rate=$encoder_projector_ds_rate \
3538 ++dataset_config.dataset=audio_dataset \
3639 ++dataset_config.val_data_path=$inference_data_path \
Original file line number Diff line number Diff line change @@ -6,6 +6,8 @@ run_dir=/data/wenxi.chen/SLAM-LLM
66cd $run_dir
77code_dir=examples/slam_aac
88
9+ encoder_fairseq_dir=/fairseq/EAT # path to the fairseq directory of the encoder model
10+
911audio_encoder_path=/data/xiquan.li/models/EAT-base_epoch30_ft.pt
1012llm_path=/data/xiquan.li/models/vicuna-7b-v1.5
1113clap_dir=/data/xiquan.li/models/clap
@@ -41,6 +43,7 @@ for num_beams in "${beam_range[@]}"; do
4143 ++model_config.encoder_projector=linear \
4244 ++model_config.encoder_projector_ds_rate=$encoder_projector_ds_rate \
4345 ++model_config.normalize=true \
46+ ++model_config.encoder_fairseq_dir=$encoder_fairseq_dir \
4447 ++dataset_config.encoder_projector_ds_rate=$encoder_projector_ds_rate \
4548 ++dataset_config.dataset=audio_dataset \
4649 ++dataset_config.val_data_path=$inference_data_path \
Original file line number Diff line number Diff line change @@ -6,6 +6,8 @@ run_dir=/data/wenxi.chen/SLAM-LLM
66cd $run_dir
77code_dir=examples/slam_aac
88
9+ encoder_fairseq_dir=/fairseq/EAT # path to the fairseq directory of the encoder model
10+
911audio_encoder_path=/data/xiquan.li/models/EAT-base_epoch30_ft.pt
1012llm_path=/data/xiquan.li/models/vicuna-7b-v1.5
1113
@@ -31,6 +33,7 @@ python $code_dir/inference_aac_batch.py \
3133 ++model_config.encoder_projector=linear \
3234 ++model_config.encoder_projector_ds_rate=$encoder_projector_ds_rate \
3335 ++model_config.normalize=true \
36+ ++model_config.encoder_fairseq_dir=$encoder_fairseq_dir \
3437 ++dataset_config.encoder_projector_ds_rate=$encoder_projector_ds_rate \
3538 ++dataset_config.dataset=audio_dataset \
3639 ++dataset_config.val_data_path=$inference_data_path \
Original file line number Diff line number Diff line change @@ -9,6 +9,8 @@ run_dir=/data/wenxi.chen/SLAM-LLM
99cd $run_dir
1010code_dir=examples/slam_aac
1111
12+ encoder_fairseq_dir=/fairseq/EAT # path to the fairseq directory of the encoder model
13+
1214audio_encoder_path=/data/xiquan.li/models/EAT-base_epoch30_ft.pt
1315llm_path=/data/xiquan.li/models/vicuna-7b-v1.5
1416
@@ -34,6 +36,7 @@ hydra.run.dir=$output_dir \
3436++model_config.encoder_path=$audio_encoder_path \
3537++model_config.encoder_dim=768 \
3638++model_config.encoder_projector=linear \
39+ ++model_config.encoder_fairseq_dir=$encoder_fairseq_dir \
3740++dataset_config.encoder_projector_ds_rate=${encoder_projector_ds_rate} \
3841++dataset_config.dataset=audio_dataset \
3942++dataset_config.train_data_path=$train_jsonl_path \
You can’t perform that action at this time.
0 commit comments