Skip to content

Commit 3613abb

Browse files
committed
update readme and requirements bump
1 parent 8c980aa commit 3613abb

File tree

3 files changed

+173
-153
lines changed

3 files changed

+173
-153
lines changed

README.md

Lines changed: 71 additions & 42 deletions
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,15 @@
55
- Outputs into a clean .cha file by utterance.
66
- As always, TRUST NOTHING GENERATED BY AI, and always verify
77

8+
## REQUIREMENTS:
9+
*See the [installation](#installation) section for detailed step by step instructions.*
10+
1. Internet (for install and download of AI models to be run locally)
11+
1. FFMPEG (should be available via environment PATH)
12+
1. GIT (should be available via environment PATH)
13+
1. PYTHON 3.11+ (should be available via environment PATH)
14+
1. Huggingface key/account (for download of AI models to be run locally)
15+
1. _[Optional]_ CUDA (requires NVIDIA GPUs, for faster runtime)
16+
817
## Installation:
918
1. **Internet**
1019
- For installation of the transcriber
@@ -46,7 +55,8 @@
4655
`git --version`
4756
If you don’t have it installed already, it should prompt you to install it. If it does not, see the detailed guide on installing git above.
4857
- Windows:
49-
- Install git for windows from the offical git website: [https://git-scm.com/downloads/win](https://git-scm.com/downloads/win)
58+
- Automatic install with command line: `winget install -e --id Git.Git`
59+
- Or manual install git for windows from the offical git website: [https://git-scm.com/downloads/win](https://git-scm.com/downloads/win)
5060
- Click on the first install link. It should look like the following
5161
> Click here to download the latest (##.##.##) x64 version of Git for Windows.
5262
- When installing, use the default configuration for all steps except:
@@ -67,21 +77,22 @@
6777
- MacOS: install ffmpeg to be made available by the command line
6878
- `brew install ffmpeg`
6979
- Windows:
70-
- Install here [instant download link to zip file](https://github.com/BtbN/FFmpeg-Builds/releases/download/latest/ffmpeg-master-latest-win64-gpl-shared.zip) or follow the instructions on the offical ffmpeg webiste here [https://ffmpeg.org/download.html](https://ffmpeg.org/download.html).
71-
- If you downloaded a zip file follow these steps:
72-
- Unzip (or move the unzipped files) to `C:/Program Files/ffmpeg`
73-
The results should look like the following:
74-
![example](docs/ffmpeg_install_windows_manual.png)
75-
- Add the `C:/Program Files/ffmpeg/bin` (or wherever you downloaded it to + `/bin`) to your system environment PATH variable
76-
- in windows bar search, type "`env`" and follow the arrows in the image below
77-
- when you click "new" enter `C:\Program Files\ffmpeg\bin`. DO NOT EDIT ANY OTHER LINE THAT IS ALREADY THERE!
78-
![example](docs/update_system_env_path.png)
80+
- Automatic install with command line: `winget install ffmpeg`
81+
- OR manual install here [instant download link to zip file](https://github.com/BtbN/FFmpeg-Builds/releases/download/latest/ffmpeg-master-latest-win64-gpl-shared.zip) or follow the instructions on the offical ffmpeg webiste here [https://ffmpeg.org/download.html](https://ffmpeg.org/download.html).
82+
- If you installed manually, follow these steps:
83+
- Unzip (or move the unzipped files) to `C:/Program Files/ffmpeg`
84+
The results should look like the following:
85+
![example](docs/ffmpeg_install_windows_manual.png)
86+
- Add the `C:/Program Files/ffmpeg/bin` (or wherever you downloaded it to + `/bin`) to your system environment PATH variable
87+
- in windows bar search, type "`env`" and follow the arrows in the image below
88+
- when you click "new" enter `C:\Program Files\ffmpeg\bin`. DO NOT EDIT ANY OTHER LINE THAT IS ALREADY THERE!
89+
![example](docs/update_system_env_path.png)
7990

8091
1. **CUDA**
8192
- If you can, use CUDA to increase performance when running the AI.
82-
- __NOTE: You will probably need a NVIDIA GPU for this.__
93+
- __NOTE: You will need a NVIDIA GPU for this.__
8394
- [Install CUDA](https://docs.nvidia.com/cuda/cuda-quick-start-guide/)
84-
- If this sounds like gibberish to you, it is safe to skip this step. Your AI models will run off of just your CPU.
95+
- If this sounds like gibberish to you, it is safe to skip this step. Your AI models will run off of your CPU.
8596

8697
1. **Python's tkinter** needs to be installed
8798
- If you installed python with the installer and selected the tk/tcl and IDLE options, you should have this step completed already.
@@ -118,25 +129,25 @@ Now that you have all of the requirements installed we will install the Transcri
118129
- MacOS/Linux
119130
- `source ./.venv/bin/activate`
120131

121-
1. <a id="pip install"></a>`pip install -r requirements.txt`
122-
- if the pip install fails try installing each package one at a time, by line as it is written in the requirements.txt file.
123-
- if that fails, [report the issue to the maintainer or your point of contact](https://github.com/Noah-Jaffe/Transcribble/issues)
124-
- The only requirements not in the txt files are `torch` `torchaudio` and `torchvision`
125-
1. Install requirements not in requirements.txt
132+
1. Install python packages with the command: <a id="pip install"></a>`pip install -r requirements.txt`
133+
- If the pip install fails try installing each package one at a time, by line as it is written in the [requirements.txt](requirements.txt) file.
134+
- If that fails, [report the issue to the maintainer or your point of contact](https://github.com/Noah-Jaffe/Transcribble/issues)
135+
- The only requirements not in the txt files are `torch` `torchaudio` and `torchvision` (continue reading...)
136+
1. Install requirements not in the requirements.txt
126137

127-
- If you are using CUDA:
138+
1. **If you are using CUDA**:
128139
- You may need to `pip uninstall torch torchaudio torchvision` before doing this next step.
129140
- See here to reinstall `torch` `torchaudio` and `torchvision` with your appropriate build: [https://pytorch.org/get-started/locally/](https://pytorch.org/get-started/locally/)
130141
- Example for if you have a CUDA compatible device for 12.6 you would run: `pip install torch torchvision torchaudio --no-cache -U --index-url https://download.pytorch.org/whl/cu126`
131-
- Otherwise if you dont have a CUDA compatible GPU or dont know what that means then run the next line:
142+
1. **Otherwise if you dont have a CUDA compatible GPU or dont know what that means then run the next line**:
132143
- `pip install torch torchaudio torchvision`
133144
1. Put your [Huggingface token](https://huggingface.co/docs/hub/en/security-tokens) in a file named `.hftoken`.
134145
- `.hftoken` only, `.hftoken.txt` will not work
135146
![Your directory should look something like this](docs/hftoken.png)
136147
- Generating the token:
137-
- You may need to create an account, you can skip all the optional steps except for verifying your email.
138-
- You can name the token anything, just save the token to a file called `.hftoken` to your machine.
139-
- The token does not need any special permissions, you can deselect all of the options.
148+
- You may need to [create a huggingface account](https://huggingface.co/join), you can skip all the optional steps except for verifying your email.
149+
- You can name the token anything on the setup page, just save the resulting token to a file called `.hftoken` to your machine.
150+
- When asked to setup the token permissions, please note that the token does not need any special permissions and *you can deselect **all*** of the options.
140151

141152
1. If you want to make a one click startup script, you could do so now.
142153
- Example for windows:
@@ -158,11 +169,11 @@ Now that you have all of the requirements installed we will install the Transcri
158169
```
159170
_You may need to run `chmod +x Transcriber.sh` on the new .sh file to give it permissions to execute_
160171
- _NOTE: You might need to do some additional research for how to do this properly for your machine._
161-
1. Continue to the [using the transcriber](#using-the-transcriber) step.
172+
1. Continue to the [using Transcribble](#using-transcribble) step.
162173

163174
---
164175

165-
## Using the transcriber
176+
## Using Transcribble
166177
1. Start the application
167178
- If using a virtual environment start/activate that now. [See here](#start-activate-venv)
168179
- Run the program with
@@ -185,25 +196,10 @@ Please see the [CITATION](CITATION.cff) file for citing this work.
185196

186197
> Jaffe, N., & Lurie, S. (2025). *Jaffe-Lurie Transcribble* [Computer software]. GitHub. https://github.com/Noah-Jaffe/Transcribble
187198
199+
---
188200

189201
# Frequently asked questions:
190202

191-
### Help, I got a `TypeError: ... not supported between instances of 'NoneType' and 'float'`!
192-
193-
If you run into the following errors specifically in the Whisper (Step 1), know that this is an error of the AI and may not entirely be in our control.
194-
> `TypeError: '<=' not supported between instances of 'NoneType' and 'float'`
195-
196-
> `TypeError: '>' not supported between instances of 'NoneType' and 'float'`
197-
198-
One patch for this is to use python package `transformers==4.38.2`.
199-
200-
To attempt patch #1:
201-
- If using a virtual environment start/activate that now. [See here](#start-activate-venv)
202-
- Run `pip install transformers==4.38.2`
203-
204-
If you are already using this version of transformers, I'm sorry but the only known workaround is to split up the original file into smaller segments, and then run the smaller segments through the transcriber again. Eventually you may hit a small section of the original audio file that crashes constantly, for that you will have to transcribe by hand.
205-
206-
207203
### How to update my Transcribble app?
208204

209205
If you followed the instructions on this readme file, or installed this repo with git:
@@ -243,6 +239,38 @@ so that it looks like
243239
morphosyntax,
244240
```
245241

242+
### How do I use this in more fine detail?
243+
244+
See the `__doc__` in the [transcribe_proc.py](transcribe_proc.py) file.
245+
246+
### I want to split a file into files of smaller (but relatively equal length) chunks!
247+
248+
Run `python splitfile.py`
249+
1. Select the files to be split
250+
2. Input the number of seperations and the seconds of overlap between each split file
251+
3. Click split file
252+
4. Find the output files next to the input files as `<file_name>.#.<ext>`
253+
- e.g. Input file `myfile.mp3` split into 2 files will result in files: `myfile.1.mp3` `myfile.2.mp3` (the original file will not be changed or moved)
254+
255+
256+
### Help, I got a `TypeError: ... not supported between instances of 'NoneType' and 'float'`!
257+
258+
If you run into the following errors specifically in the Whisper (Step 1), know that this is an error of the AI and may not entirely be in our control.
259+
> `TypeError: '<=' not supported between instances of 'NoneType' and 'float'`
260+
261+
> `TypeError: '>' not supported between instances of 'NoneType' and 'float'`
262+
263+
One patch for this is to use python package `transformers==4.38.2`.
264+
265+
To attempt patch #1:
266+
- If using a virtual environment start/activate that now. [See here](#start-activate-venv)
267+
- Run `pip install transformers==4.38.2`
268+
269+
If you are already using this version of transformers, I'm sorry but the only known workaround is to split up the original file into smaller segments, and then run the smaller segments through the transcriber again. Eventually you may hit a small section of the original audio file that crashes constantly, for that you will have to transcribe by hand.
270+
271+
Note that this issue should be fixed as of Transcribble release 1.1.1!
272+
273+
246274
---
247275

248276
### Backlog ideas:
@@ -253,5 +281,6 @@ so that it looks like
253281
- [ ] Select subframe of time to transcribe from?
254282
- [ ] Better error handling
255283
- [ ] Checkboxes for pipeline steps?
256-
257-
284+
- [ ] Research Inter-sentential code-switching
285+
- [ ] Custom setup for auto run script on output .cha file?
286+
- [ ] Remove reliance on batchalign?

0 commit comments

Comments
 (0)