Skip to content

Commit 8f437e7

Browse files
committed
0.0.1
0 parents  commit 8f437e7

10 files changed

+398
-0
lines changed

.gitignore

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
.venv
2+
.vscode
3+
models
4+
*.mp4
5+
*.mkv
6+
*.mp3
7+
*.xlsx

README.md

Lines changed: 42 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,42 @@
1+
# Transcriber
2+
- Uses [`openai-whisper`](https://github.com/openai/whisper) to transcribe audio or video files.
3+
- Runs locally.
4+
- Outputs into a clean workbook by trascribed section and word.
5+
- As always, TRUST NOTHING GENERATED BY AI, and always verify
6+
7+
## Requirements:
8+
- Python3+
9+
- ffmpeg (included in repo)
10+
- beefy pc(?)
11+
- internet (for model download, on the first time you use it)
12+
---
13+
# How to use:
14+
0. install requirements
15+
- `pip install -r requirements.txt`
16+
1. Start application.
17+
- `python ./transcribe.py`
18+
2. Select files (button) to be transcribed.
19+
3. Select AI model to use from the drop down.
20+
- Hover over the drop down to see some selection guidance. Choose one that your device can handle.
21+
- [Read more about the models here.](https://github.com/openai/whisper?tab=readme-ov-file#available-models-and-languages)
22+
4. Start transcript (button).
23+
5. Wait for final results popup to appear.
24+
6. Review resulting transcripts.
25+
---
26+
#### What the running program looks like.
27+
![What the running program looks like.](docs\readme_demo.png)
28+
#### Completion notification.
29+
![Completion notification.](docs\readme_transcripts_complete.png)
30+
#### Output files can be found next to the input files.
31+
![Output files can be found next to the input files.](docs\readme_output_location.png)
32+
#### Output contents is 2 sheets, one by segment and one by word.
33+
![Output contents is 2 sheets, one by segment and one by word.](docs\readme_demo_outputs.png)
34+
##### Don't blame me if the AI is bad. (Same input as previous pic, but got different output.)
35+
![Same input but different output](docs\readme_same_input_different_output.png)
36+
37+
---
38+
### Backlog ideas:
39+
- [ ] Additional AI sources?
40+
- [ ] Advanced/Runtime configuration of AI parameters?
41+
- [ ] Bundle into single executable to be more user friendly?
42+
- [ ] Select subframe of time to transcribe from?

docs/readme_demo.png

13 KB
Loading

docs/readme_demo_outputs.png

64.7 KB
Loading

docs/readme_output_location.png

8.08 KB
Loading
25.6 KB
Loading
8.08 KB
Loading

ffmpeg.exe

72.2 MB
Binary file not shown.

requirements.txt

Lines changed: 28 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,28 @@
1+
# used:
2+
ffmpeg-python==0.2.0
3+
openai-whisper==20240930
4+
XlsxWriter==3.2.2
5+
# deps:
6+
certifi==2025.1.31
7+
charset-normalizer==3.4.1
8+
colorama==0.4.6
9+
filelock==3.17.0
10+
fsspec==2025.2.0
11+
future==1.0.0
12+
idna==3.10
13+
jinja2==3.1.5
14+
llvmlite==0.43.0
15+
MarkupSafe==3.0.2
16+
more-itertools==10.6.0
17+
mpmath==1.3.0
18+
networkx==3.2.1
19+
numba==0.60.0
20+
numpy==2.0.2
21+
regex==2024.11.6
22+
requests==2.32.3
23+
sympy==1.13.1
24+
tiktoken==0.9.0
25+
torch==2.6.0
26+
tqdm==4.67.1
27+
typing-extensions==4.12.2
28+
urllib3==2.3.0

0 commit comments

Comments
 (0)