This project is working with Python 3.8.8+, you should have it installed on your machine.
pip install -r requirements.txt
The frontend is inside the frontend/AudioTranscribeAI
folder.
The application is built with node.js and vue.js.
So to start the server, you need to install the node.js and yarn package manager.
npm install --global yarn
# if you are in the root directory
yarn --cwd frontend/AudioTranscribeAI/
# or
cd frontend/AudioTranscribeAI
# if you are in the frontend/AudioTranscribeAI directory
yarn
Because we are using separate backend and frontend architecture, so we need to run both of them.
python app.py
# if you are in the root directory
yarn --cwd frontend/AudioTranscribeAI/ dev
# or
cd frontend/AudioTranscribeAI
# if you are in the frontend/AudioTranscribeAI directory
yarn dev
yarn --cwd frontend/AudioTranscribeAI/ build # if you are in the root directory
# or
cd frontend/AudioTranscribeAI
yarn build # if you are in the frontend/AudioTranscribeAI directory
✨Model: openai/whisper-small
✨Model:
- Default: TinyLlama/TinyLlama-1.1B-Chat-v1.0
- Alternative: meta-llama/Llama-2-7b-chat-hf
✨Model spacy: en_core_web_sm
- PyDictionary package
- pywikibot package
- Text processing with nltk package
For aduio model evaluation, we use the LibriSpeech dataset.
cd asr && python eval_asr.py
For the LLM evaluation on summarization, direct use the llm.py
:
python llm/llm.py