|
| 1 | +# Transcriber |
| 2 | +- Uses [`openai-whisper`](https://github.com/openai/whisper) to transcribe audio or video files. |
| 3 | +- Runs locally. |
| 4 | +- Outputs into a clean workbook by trascribed section and word. |
| 5 | +- As always, TRUST NOTHING GENERATED BY AI, and always verify |
| 6 | + |
| 7 | +## Requirements: |
| 8 | +- Python3+ |
| 9 | +- ffmpeg (included in repo) |
| 10 | +- beefy pc(?) |
| 11 | +- internet (for model download, on the first time you use it) |
| 12 | +--- |
| 13 | +# How to use: |
| 14 | +0. install requirements |
| 15 | + - `pip install -r requirements.txt` |
| 16 | +1. Start application. |
| 17 | + - `python ./transcribe.py` |
| 18 | +2. Select files (button) to be transcribed. |
| 19 | +3. Select AI model to use from the drop down. |
| 20 | + - Hover over the drop down to see some selection guidance. Choose one that your device can handle. |
| 21 | + - [Read more about the models here.](https://github.com/openai/whisper?tab=readme-ov-file#available-models-and-languages) |
| 22 | +4. Start transcript (button). |
| 23 | +5. Wait for final results popup to appear. |
| 24 | +6. Review resulting transcripts. |
| 25 | +--- |
| 26 | +#### What the running program looks like. |
| 27 | + |
| 28 | +#### Completion notification. |
| 29 | + |
| 30 | +#### Output files can be found next to the input files. |
| 31 | + |
| 32 | +#### Output contents is 2 sheets, one by segment and one by word. |
| 33 | + |
| 34 | +##### Don't blame me if the AI is bad. (Same input as previous pic, but got different output.) |
| 35 | + |
| 36 | + |
| 37 | +--- |
| 38 | +### Backlog ideas: |
| 39 | +- [ ] Additional AI sources? |
| 40 | +- [ ] Advanced/Runtime configuration of AI parameters? |
| 41 | +- [ ] Bundle into single executable to be more user friendly? |
| 42 | +- [ ] Select subframe of time to transcribe from? |
0 commit comments