Skip to content

Commit c16e3a9

Browse files
committed
fix model name, README and fsdp config for examples/sec_emotioncaps
1 parent f37fe73 commit c16e3a9

File tree

3 files changed

+7
-17
lines changed

3 files changed

+7
-17
lines changed

examples/sec_emotioncaps/README.md

Lines changed: 0 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -40,16 +40,3 @@ If you do have sufficient relevant data, you can train the model yourself.
4040
```
4141
bash finetune_emotion2vec_qformer_vicuna_7b.sh
4242
```
43-
44-
## Citation
45-
46-
You can refer to the paper for more results.
47-
48-
```
49-
@article{ma2024embarrassingly,
50-
title={An Embarrassingly Simple Approach for LLM with Strong ASR Capacity},
51-
author={Ma, Ziyang and Yang, Guanrou and Yang, Yifan and Gao, Zhifu and Wang, Jiaming and Du, Zhihao and Yu, Fan and Chen, Qian and Zheng, Siqi and Zhang, Shiliang and others},
52-
journal={arXiv preprint arXiv:2402.08846},
53-
year={2024}
54-
}
55-
```

examples/sec_emotioncaps/model/slam_model_sec.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ def model_factory(train_config, model_config, **kwargs):
2525
encoder_projector = setup_encoder_projector(
2626
train_config, model_config, **kwargs
2727
)
28-
model = slam_model_asr(
28+
model = slam_model_sec(
2929
encoder,
3030
llm,
3131
encoder_projector,
@@ -55,7 +55,7 @@ def model_factory(train_config, model_config, **kwargs):
5555
return model, tokenizer
5656

5757

58-
class slam_model_asr(slam_model):
58+
class slam_model_sec(slam_model):
5959
def __init__(
6060
self,
6161
encoder,

examples/sec_emotioncaps/sec_config.py

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,8 @@
11
from dataclasses import dataclass, field
22
from typing import Optional, List
3+
4+
from torch.distributed.fsdp import ShardingStrategy
5+
36
@dataclass
47
class ModelConfig:
58
file: str = "examples/sec_emotioncaps/model/slam_model_sec.py:model_factory"
@@ -108,8 +111,8 @@ class DataConfig:
108111
class FSDPConfig:
109112
mixed_precision: bool = True
110113
use_fp16: bool = False
111-
# sharding_strategy: str = "FULL_SHARD" #ShardingStrategy = ShardingStrategy.FULL_SHARD
112-
sharding_strategy: str = "NO_SHARD" #MZY: set NO_SHARD when use DDP
114+
# sharding_strategy = "FULL_SHARD" #ShardingStrategy = ShardingStrategy.FULL_SHARD
115+
sharding_strategy: ShardingStrategy = "NO_SHARD" #ShardingStrategy.NO_SHARD #MZY: set NO_SHARD when use DDP
113116
checkpoint_type: str = "SHARDED_STATE_DICT" # alternatively can use SHARDED_STATE_DICT save one file per rank, and can resize the world-size.
114117
fsdp_activation_checkpointing: bool = True
115118
fsdp_cpu_offload: bool = False

0 commit comments

Comments
 (0)