w2v-bert-2.0-bemgen-female-model
This model is a fine-tuned version of facebook/w2v-bert-2.0 on the BEMGEN - BEM dataset. It achieves the following results on the evaluation set:
- Loss: 0.1984
- Wer: 0.3901
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0003
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 3000
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Wer |
---|---|---|---|---|
No log | 0.1028 | 100 | 0.7203 | 0.9424 |
No log | 0.2055 | 200 | 0.7191 | 1.2778 |
No log | 0.3083 | 300 | 0.8854 | 0.9047 |
No log | 0.4111 | 400 | 0.7127 | 0.9246 |
0.8843 | 0.5139 | 500 | 0.7602 | 0.8745 |
0.8843 | 0.6166 | 600 | 0.5975 | 0.8196 |
0.8843 | 0.7194 | 700 | 0.5665 | 0.8119 |
0.8843 | 0.8222 | 800 | 0.5523 | 0.7963 |
0.8843 | 0.9250 | 900 | 0.5271 | 0.7702 |
0.5422 | 1.0277 | 1000 | 0.4329 | 0.6983 |
0.5422 | 1.1305 | 1100 | 0.4232 | 0.6759 |
0.5422 | 1.2333 | 1200 | 0.3969 | 0.6648 |
0.5422 | 1.3361 | 1300 | 0.3574 | 0.6024 |
0.5422 | 1.4388 | 1400 | 0.3778 | 0.6090 |
0.3586 | 1.5416 | 1500 | 0.3328 | 0.5881 |
0.3586 | 1.6444 | 1600 | 0.2965 | 0.5361 |
0.3586 | 1.7472 | 1700 | 0.2929 | 0.5120 |
0.3586 | 1.8499 | 1800 | 0.2829 | 0.5372 |
0.3586 | 1.9527 | 1900 | 0.2802 | 0.5226 |
0.2874 | 2.0555 | 2000 | 0.2988 | 0.5407 |
0.2874 | 2.1583 | 2100 | 0.2835 | 0.5164 |
0.2874 | 2.2610 | 2200 | 0.2721 | 0.5103 |
0.2874 | 2.3638 | 2300 | 0.2618 | 0.4979 |
0.2874 | 2.4666 | 2400 | 0.2547 | 0.4701 |
0.1943 | 2.5694 | 2500 | 0.2240 | 0.4379 |
0.1943 | 2.6721 | 2600 | 0.2204 | 0.4419 |
0.1943 | 2.7749 | 2700 | 0.2127 | 0.4221 |
0.1943 | 2.8777 | 2800 | 0.2062 | 0.4104 |
0.1943 | 2.9805 | 2900 | 0.1988 | 0.3947 |
0.154 | 3.0832 | 3000 | 0.1984 | 0.3905 |
Framework versions
- Transformers 4.47.1
- Pytorch 2.5.1+cu124
- Datasets 3.2.0
- Tokenizers 0.21.0
- Downloads last month
- 3
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
Model tree for csikasote/w2v-bert-2.0-bemgen-female-model
Base model
facebook/w2v-bert-2.0