zibib's picture
Update README.md
06dc58c verified
---
library_name: transformers
base_model: ivrit-ai/whisper-large-v3-turbo
tags:
- generated_from_trainer
model-index:
- name: whisper-large-v3-turbo-ivrit-ai-coursera-fine-tuned
results: []
datasets:
- imvladikon/hebrew_speech_coursera
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# whisper-large-v3-turbo-ivrit-ai-coursera-fine-tuned
This model is a fine-tuned version of [ivrit-ai/whisper-large-v3-turbo](https://huggingface.co/ivrit-ai/whisper-large-v3-turbo) on the dataset imvladikon/hebrew_speech_coursera.
It achieves the following results on the evaluation set:
- Loss: 0.2829
## Model description
This model created for my work for the Open University Of Israel.
[Here](https://colab.research.google.com/gist/zibib3/373bbc36c305899e29c1a91b9a834c97/.ipynb) you can see the notebook that used to
create this model, and [here](https://www.youtube.com/live/rEoG9vF0GAo) you can find me displaying the notebook.
I think that this model is useless becaus it has lower performance from its base model.
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 5
- eval_batch_size: 5
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: constant_with_warmup
- lr_scheduler_warmup_ratio: 0.1
- lr_scheduler_warmup_steps: 500
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:------:|:-----:|:---------------:|
| 0.1907 | 0.1641 | 500 | 0.2266 |
| 0.2283 | 0.3283 | 1000 | 0.2217 |
| 0.2253 | 0.4924 | 1500 | 0.2154 |
| 0.2257 | 0.6566 | 2000 | 0.2080 |
| 0.2138 | 0.8207 | 2500 | 0.2102 |
| 0.2153 | 0.9849 | 3000 | 0.2056 |
| 0.1615 | 1.1490 | 3500 | 0.2128 |
| 0.1588 | 1.3132 | 4000 | 0.1677 |
| 0.1628 | 1.4773 | 4500 | 0.1656 |
| 0.168 | 1.6415 | 5000 | 0.1798 |
| 0.167 | 1.8056 | 5500 | 0.1710 |
| 0.1663 | 1.9698 | 6000 | 0.1828 |
| 0.1297 | 2.1339 | 6500 | 0.1722 |
| 0.1196 | 2.2981 | 7000 | 0.1762 |
| 0.1336 | 2.4622 | 7500 | 0.1779 |
| 0.1258 | 2.6264 | 8000 | 0.1821 |
| 0.1275 | 2.7905 | 8500 | 0.1796 |
| 0.1331 | 2.9547 | 9000 | 0.1786 |
| 0.0988 | 3.1188 | 9500 | 0.1982 |
| 0.0933 | 3.2830 | 10000 | 0.1888 |
| 0.0963 | 3.4471 | 10500 | 0.1927 |
| 0.0946 | 3.6113 | 11000 | 0.1979 |
| 0.1018 | 3.7754 | 11500 | 0.2031 |
| 0.1027 | 3.9396 | 12000 | 0.1971 |
| 0.0795 | 4.1037 | 12500 | 0.2016 |
| 0.0698 | 4.2679 | 13000 | 0.2017 |
| 0.0736 | 4.4320 | 13500 | 0.2058 |
| 0.0747 | 4.5962 | 14000 | 0.2033 |
| 0.0768 | 4.7603 | 14500 | 0.2057 |
| 0.0801 | 4.9245 | 15000 | 0.2076 |
| 0.067 | 5.0886 | 15500 | 0.2196 |
| 0.0539 | 5.2528 | 16000 | 0.2185 |
| 0.0563 | 5.4169 | 16500 | 0.2220 |
| 0.0594 | 5.5811 | 17000 | 0.2265 |
| 0.0651 | 5.7452 | 17500 | 0.2176 |
| 0.0655 | 5.9094 | 18000 | 0.2227 |
| 0.0533 | 6.0735 | 18500 | 0.2387 |
| 0.0441 | 6.2377 | 19000 | 0.2334 |
| 0.0474 | 6.4018 | 19500 | 0.2343 |
| 0.0506 | 6.5660 | 20000 | 0.2387 |
| 0.0504 | 6.7301 | 20500 | 0.2373 |
| 0.0502 | 6.8943 | 21000 | 0.2318 |
| 0.0441 | 7.0584 | 21500 | 0.2524 |
| 0.0375 | 7.2226 | 22000 | 0.2533 |
| 0.0379 | 7.3867 | 22500 | 0.2491 |
| 0.0382 | 7.5509 | 23000 | 0.2635 |
| 0.0427 | 7.7150 | 23500 | 0.2506 |
| 0.0439 | 7.8792 | 24000 | 0.2430 |
| 0.043 | 8.0433 | 24500 | 0.2575 |
| 0.0296 | 8.2075 | 25000 | 0.2617 |
| 0.0309 | 8.3716 | 25500 | 0.2797 |
| 0.0366 | 8.5358 | 26000 | 0.2689 |
| 0.0351 | 8.6999 | 26500 | 0.2687 |
| 0.0384 | 8.8641 | 27000 | 0.2643 |
| 0.0365 | 9.0282 | 27500 | 0.2688 |
| 0.0265 | 9.1924 | 28000 | 0.2903 |
| 0.0299 | 9.3565 | 28500 | 0.2742 |
| 0.0347 | 9.5207 | 29000 | 0.2754 |
| 0.0311 | 9.6848 | 29500 | 0.2744 |
| 0.0345 | 9.8490 | 30000 | 0.2829 |
### Framework versions
- Transformers 4.48.1
- Pytorch 2.6.0+cu124
- Datasets 3.4.1
- Tokenizers 0.21.1