Initial training on seed20; DeBERTa-v3 toxicity classifier.
Browse files- README.md +16 -16
- model.safetensors +1 -1
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -21,12 +21,12 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 21 |
|
| 22 |
This model is a fine-tuned version of [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) on an unknown dataset.
|
| 23 |
It achieves the following results on the evaluation set:
|
| 24 |
-
- Loss: 0.
|
| 25 |
-
- Accuracy: 0.
|
| 26 |
-
- Precision: 0.
|
| 27 |
-
- Recall: 0.
|
| 28 |
-
- F1: 0.
|
| 29 |
-
- Auc: 0.
|
| 30 |
|
| 31 |
## Model description
|
| 32 |
|
|
@@ -47,7 +47,7 @@ More information needed
|
|
| 47 |
The following hyperparameters were used during training:
|
| 48 |
- learning_rate: 5e-05
|
| 49 |
- train_batch_size: 32
|
| 50 |
-
- eval_batch_size:
|
| 51 |
- seed: 13
|
| 52 |
- gradient_accumulation_steps: 8
|
| 53 |
- total_train_batch_size: 256
|
|
@@ -59,19 +59,19 @@ The following hyperparameters were used during training:
|
|
| 59 |
|
| 60 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Auc |
|
| 61 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:------:|
|
| 62 |
-
| No log | 1.0 | 141 | 0.
|
| 63 |
-
| No log | 2.0 | 282 | 0.
|
| 64 |
-
| No log | 3.0 | 423 | 0.
|
| 65 |
-
| 0.
|
| 66 |
-
| 0.
|
| 67 |
-
| 0.
|
| 68 |
-
| 0.
|
| 69 |
-
| 0.
|
| 70 |
|
| 71 |
|
| 72 |
### Framework versions
|
| 73 |
|
| 74 |
- Transformers 4.57.1
|
| 75 |
-
- Pytorch 2.
|
| 76 |
- Datasets 4.4.1
|
| 77 |
- Tokenizers 0.22.1
|
|
|
|
| 21 |
|
| 22 |
This model is a fine-tuned version of [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) on an unknown dataset.
|
| 23 |
It achieves the following results on the evaluation set:
|
| 24 |
+
- Loss: 0.3694
|
| 25 |
+
- Accuracy: 0.8054
|
| 26 |
+
- Precision: 0.7440
|
| 27 |
+
- Recall: 0.9942
|
| 28 |
+
- F1: 0.8511
|
| 29 |
+
- Auc: 0.8908
|
| 30 |
|
| 31 |
## Model description
|
| 32 |
|
|
|
|
| 47 |
The following hyperparameters were used during training:
|
| 48 |
- learning_rate: 5e-05
|
| 49 |
- train_batch_size: 32
|
| 50 |
+
- eval_batch_size: 8
|
| 51 |
- seed: 13
|
| 52 |
- gradient_accumulation_steps: 8
|
| 53 |
- total_train_batch_size: 256
|
|
|
|
| 59 |
|
| 60 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Auc |
|
| 61 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:------:|
|
| 62 |
+
| No log | 1.0 | 141 | 0.4441 | 0.8012 | 0.7428 | 0.9861 | 0.8473 | 0.8880 |
|
| 63 |
+
| No log | 2.0 | 282 | 0.3568 | 0.8042 | 0.7453 | 0.9875 | 0.8495 | 0.8905 |
|
| 64 |
+
| No log | 3.0 | 423 | 0.3691 | 0.8052 | 0.7444 | 0.9926 | 0.8508 | 0.8922 |
|
| 65 |
+
| 0.4062 | 4.0 | 564 | 0.3701 | 0.8054 | 0.7440 | 0.9942 | 0.8511 | 0.8908 |
|
| 66 |
+
| 0.4062 | 5.0 | 705 | 0.3925 | 0.8051 | 0.7436 | 0.9944 | 0.8509 | 0.8915 |
|
| 67 |
+
| 0.4062 | 6.0 | 846 | 0.3891 | 0.8056 | 0.7498 | 0.9793 | 0.8493 | 0.8921 |
|
| 68 |
+
| 0.4062 | 7.0 | 987 | 0.3860 | 0.8070 | 0.7573 | 0.9638 | 0.8482 | 0.8943 |
|
| 69 |
+
| 0.3208 | 8.0 | 1128 | 0.3909 | 0.8073 | 0.7603 | 0.9575 | 0.8475 | 0.8939 |
|
| 70 |
|
| 71 |
|
| 72 |
### Framework versions
|
| 73 |
|
| 74 |
- Transformers 4.57.1
|
| 75 |
+
- Pytorch 2.8.0+cu129
|
| 76 |
- Datasets 4.4.1
|
| 77 |
- Tokenizers 0.22.1
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 737719272
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:49df3870ca7de2415f1c78353313e1c28f3ada9129d2e1fb58df386ad4cc8556
|
| 3 |
size 737719272
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5841
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:288663d4dae564292a49ddcd99d203282db760c60bcffd017f24a988a8d2a398
|
| 3 |
size 5841
|