-
|
Hi, I’ve been working with the F5-TTS model, and I’m interested in integrating a Byte Pair Encoding (BPE) tokenizer instead of the character-level tokenizer currently used. I’m familiar with how BPE tokenization works and the general steps for implementing it, but I’m unsure of the specific files and methods in the F5-TTS codebase that need to be modified to support this change. Could you kindly guide me on:
Thanks in advance! |
Beta Was this translation helpful? Give feedback.
Replies: 4 comments 13 replies
-
|
core codes with model here https://github.com/SWivid/F5-TTS/tree/main/src/f5_tts/model basically change the text format is fine F5-TTS/src/f5_tts/model/cfm.py Line 214 in c2cf31e F5-TTS/src/f5_tts/model/cfm.py Lines 227 to 233 in c2cf31e
inference process is the same, you could make a func e.g. char_to_bpe(), and pass the result in as |
Beta Was this translation helpful? Give feedback.
-
|
What was the reason you shifted to bpe tokenizer? just curios and it might help me because i am also facing issues in vocab learning . |
Beta Was this translation helpful? Give feedback.
-
|
Ohh gotcha |
Beta Was this translation helpful? Give feedback.
-
|
@Alykasym can you share your settings.json file?? i alos tried to train f5-tts with open source datya espcially for hindi but my results are not that great as its should be espcially the pronunciation, whats your data size and how many epochs you trained!! i was training with 4090 i tried to train it for a week still the results were not that upto mark !! |
Beta Was this translation helpful? Give feedback.
core codes with model here https://github.com/SWivid/F5-TTS/tree/main/src/f5_tts/model
basically change the text format is fine
F5-TTS/src/f5_tts/model/cfm.py
Line 214 in c2cf31e
F5-TTS/src/f5_tts/model/cfm.py
Lines 227 to 233 in c2cf31e
textis currently input like: [['h', 'o', 'w', ' ', 'a', 'r', 'e', ' ', 'y', 'o', 'u', '?',], ['i', ''', 'm', ' ', 'f', 'i', '…