Skip to content

Mini Project on Trying to Finetune whisper-tiny model using Deutsch dataset

Notifications You must be signed in to change notification settings

HanCreation/Whisper-Tiny-German

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 

Repository files navigation

Whisper-Tiny-German Fine-Tuning Mini Project

Overview

This mini project focuses on fine-tuning the Whisper-Tiny model by OpenAI using the German dataset from Common Voice 11. Before finetuning the model achieved 43.488% in the test dataset. Due to computational limitations, I experimented with two variants of the model:

  1. Model V1 (100k Train Data):

    • Trained for 5 hours with 4000 Steps
    • Achieved a Word Error Rate (WER) of 31% on the test dataset
  2. Model V2 (200k Train Data):

    • Trained for 10 hours with 8000 Steps
    • Achieved a WER of 32% on the test dataset

Computer Spec = AMD Ryzen 9 5900HS, 8GB of RAM available, RTX 3060 Laptop GPU 6GB of VRAM

Try Gradio Demo for the ASR and Translation (With gtranslate translator)

You can try the model in this google colab: Demo

Model Availability

You can find the pre-trained models on Hugging Face:

*This project is for school project

Created and Trained by Han 2024

Helped by friends in my project group

About

Mini Project on Trying to Finetune whisper-tiny model using Deutsch dataset

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published