Hanish2007 commited on
Commit
a67bafb
·
verified ·
1 Parent(s): bd9e268

End of training

Browse files
Files changed (5) hide show
  1. README.md +26 -15
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. tokenizer_config.json +1 -1
  5. training_args.bin +2 -2
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: distilbert/distilbert-base-uncased
4
  tags:
@@ -17,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.3497
21
  - Accuracy: 0.6355
22
 
23
  ## Model description
@@ -41,29 +42,39 @@ The following hyperparameters were used during training:
41
  - train_batch_size: 32
42
  - eval_batch_size: 32
43
  - seed: 42
44
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - num_epochs: 10
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
- | 3.319 | 1.0 | 65 | 2.7462 | 0.4230 |
53
- | 2.6272 | 2.0 | 130 | 2.0795 | 0.5010 |
54
- | 2.137 | 3.0 | 195 | 1.6683 | 0.5770 |
55
- | 1.469 | 4.0 | 260 | 1.4721 | 0.6101 |
56
- | 1.2405 | 5.0 | 325 | 1.3497 | 0.6355 |
57
- | 1.1023 | 6.0 | 390 | 1.2936 | 0.6335 |
58
- | 0.9206 | 7.0 | 455 | 1.2855 | 0.6316 |
59
- | 0.8374 | 8.0 | 520 | 1.2579 | 0.6355 |
60
- | 0.794 | 9.0 | 585 | 1.2525 | 0.6335 |
61
- | 0.7388 | 10.0 | 650 | 1.2478 | 0.6316 |
 
 
 
 
 
 
 
 
 
 
62
 
63
 
64
  ### Framework versions
65
 
66
- - Transformers 4.44.0
67
  - Pytorch 2.4.0
68
  - Datasets 2.20.0
69
- - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: distilbert/distilbert-base-uncased
5
  tags:
 
18
 
19
  This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.2600
22
  - Accuracy: 0.6355
23
 
24
  ## Model description
 
42
  - train_batch_size: 32
43
  - eval_batch_size: 32
44
  - seed: 42
45
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
+ - num_epochs: 20
48
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
+ | 3.3759 | 1.0 | 65 | 2.8220 | 0.4152 |
54
+ | 2.7146 | 2.0 | 130 | 2.0778 | 0.5127 |
55
+ | 2.1537 | 3.0 | 195 | 1.6326 | 0.6121 |
56
+ | 1.4176 | 4.0 | 260 | 1.4415 | 0.6062 |
57
+ | 1.1537 | 5.0 | 325 | 1.2944 | 0.6316 |
58
+ | 1.0093 | 6.0 | 390 | 1.2600 | 0.6355 |
59
+ | 0.7806 | 7.0 | 455 | 1.2770 | 0.6199 |
60
+ | 0.6639 | 8.0 | 520 | 1.2654 | 0.6296 |
61
+ | 0.5922 | 9.0 | 585 | 1.2733 | 0.6296 |
62
+ | 0.4659 | 10.0 | 650 | 1.3403 | 0.6179 |
63
+ | 0.3928 | 11.0 | 715 | 1.3584 | 0.6179 |
64
+ | 0.3347 | 12.0 | 780 | 1.3825 | 0.6179 |
65
+ | 0.3175 | 13.0 | 845 | 1.4199 | 0.6101 |
66
+ | 0.2582 | 14.0 | 910 | 1.4277 | 0.6179 |
67
+ | 0.2097 | 15.0 | 975 | 1.4421 | 0.6179 |
68
+ | 0.2308 | 16.0 | 1040 | 1.4636 | 0.6101 |
69
+ | 0.1753 | 17.0 | 1105 | 1.4857 | 0.6199 |
70
+ | 0.1632 | 18.0 | 1170 | 1.4894 | 0.6277 |
71
+ | 0.1564 | 19.0 | 1235 | 1.5043 | 0.6160 |
72
+ | 0.1494 | 20.0 | 1300 | 1.5040 | 0.6179 |
73
 
74
 
75
  ### Framework versions
76
 
77
+ - Transformers 4.46.3
78
  - Pytorch 2.4.0
79
  - Datasets 2.20.0
80
+ - Tokenizers 0.20.3
config.json CHANGED
@@ -96,6 +96,6 @@
96
  "sinusoidal_pos_embds": false,
97
  "tie_weights_": true,
98
  "torch_dtype": "float32",
99
- "transformers_version": "4.44.0",
100
  "vocab_size": 30522
101
  }
 
96
  "sinusoidal_pos_embds": false,
97
  "tie_weights_": true,
98
  "torch_dtype": "float32",
99
+ "transformers_version": "4.46.3",
100
  "vocab_size": 30522
101
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72c8c1cf03c6864b2956e5235f34b02bcd3ea51f808ba94f5e30e78c84d468bb
3
  size 267937160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f077b2f4176aaa251c8e69019c07d05e68074219b085310a8bda208411beb20c
3
  size 267937160
tokenizer_config.json CHANGED
@@ -41,7 +41,7 @@
41
  "special": true
42
  }
43
  },
44
- "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
 
41
  "special": true
42
  }
43
  },
44
+ "clean_up_tokenization_spaces": false,
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e9a68dcb35447df75f0612c9054154ae4966d21ea11f108d3320c9e29bf8278
3
- size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c138e5f4cf55dcf5dfd35b0ebfd4dca743f0d212a315a64a5b90924b8f7f72d
3
+ size 5240