j05hr3d commited on
Commit
9663a8e
·
verified ·
1 Parent(s): b6f4e2b

Model save

Browse files
Files changed (3) hide show
  1. README.md +19 -1
  2. adapter_model.safetensors +1 -1
  3. trainer_state.json +170 -10
README.md CHANGED
@@ -18,6 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
  # SFT-Qwen2.5-Coder-3B_long_v1
19
 
20
  This model is a fine-tuned version of [Qwen/Qwen2.5-Coder-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-3B-Instruct) on the None dataset.
 
 
21
 
22
  ## Model description
23
 
@@ -38,7 +40,7 @@ More information needed
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0001
40
  - train_batch_size: 1
41
- - eval_batch_size: 8
42
  - seed: 42
43
  - gradient_accumulation_steps: 8
44
  - total_train_batch_size: 8
@@ -47,6 +49,22 @@ The following hyperparameters were used during training:
47
  - lr_scheduler_warmup_ratio: 0.03
48
  - num_epochs: 3
49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  ### Framework versions
51
 
52
  - PEFT 0.18.0
 
18
  # SFT-Qwen2.5-Coder-3B_long_v1
19
 
20
  This model is a fine-tuned version of [Qwen/Qwen2.5-Coder-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-3B-Instruct) on the None dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: 0.7569
23
 
24
  ## Model description
25
 
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 0.0001
42
  - train_batch_size: 1
43
+ - eval_batch_size: 1
44
  - seed: 42
45
  - gradient_accumulation_steps: 8
46
  - total_train_batch_size: 8
 
49
  - lr_scheduler_warmup_ratio: 0.03
50
  - num_epochs: 3
51
 
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss |
55
+ |:-------------:|:------:|:----:|:---------------:|
56
+ | 0.9906 | 0.2807 | 20 | 0.9487 |
57
+ | 0.8528 | 0.5614 | 40 | 0.8620 |
58
+ | 0.8721 | 0.8421 | 60 | 0.8238 |
59
+ | 0.8059 | 1.1123 | 80 | 0.8018 |
60
+ | 0.8141 | 1.3930 | 100 | 0.7868 |
61
+ | 0.7353 | 1.6737 | 120 | 0.7767 |
62
+ | 0.6779 | 1.9544 | 140 | 0.7647 |
63
+ | 0.6273 | 2.2246 | 160 | 0.7629 |
64
+ | 0.6983 | 2.5053 | 180 | 0.7597 |
65
+ | 0.6958 | 2.7860 | 200 | 0.7569 |
66
+
67
+
68
  ### Framework versions
69
 
70
  - PEFT 0.18.0
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64150ddf1b93e1201ed89104b4610933cb5157950a32dbc4064c780f61df1545
3
  size 239536272
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b1bb12b446b82d4cdb5000895076f8c66e6544e2d36eaf932edcacf8b081d64
3
  size 239536272
trainer_state.json CHANGED
@@ -1,20 +1,180 @@
1
  {
2
- "best_global_step": null,
3
- "best_metric": null,
4
- "best_model_checkpoint": null,
5
- "epoch": 0.2807017543859649,
6
  "eval_steps": 20,
7
- "global_step": 20,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 0.2807017543859649,
14
- "grad_norm": 0.47223395109176636,
15
  "learning_rate": 9.425837320574164e-05,
16
- "loss": 0.9903,
17
  "step": 20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  }
19
  ],
20
  "logging_steps": 20,
@@ -37,13 +197,13 @@
37
  "should_epoch_stop": false,
38
  "should_evaluate": false,
39
  "should_log": false,
40
- "should_save": false,
41
- "should_training_stop": false
42
  },
43
  "attributes": {}
44
  }
45
  },
46
- "total_flos": 2320839556472832.0,
47
  "train_batch_size": 1,
48
  "trial_name": null,
49
  "trial_params": null
 
1
  {
2
+ "best_global_step": 200,
3
+ "best_metric": 0.7568597793579102,
4
+ "best_model_checkpoint": "j05hr3d/SFT-Qwen2.5-Coder-3B_long_v1/checkpoint-200",
5
+ "epoch": 3.0,
6
  "eval_steps": 20,
7
+ "global_step": 216,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 0.2807017543859649,
14
+ "grad_norm": 0.47658494114875793,
15
  "learning_rate": 9.425837320574164e-05,
16
+ "loss": 0.9906,
17
  "step": 20
18
+ },
19
+ {
20
+ "epoch": 0.2807017543859649,
21
+ "eval_loss": 0.9487136006355286,
22
+ "eval_runtime": 14.7059,
23
+ "eval_samples_per_second": 4.352,
24
+ "eval_steps_per_second": 4.352,
25
+ "step": 20
26
+ },
27
+ {
28
+ "epoch": 0.5614035087719298,
29
+ "grad_norm": 0.4057779908180237,
30
+ "learning_rate": 8.468899521531101e-05,
31
+ "loss": 0.8528,
32
+ "step": 40
33
+ },
34
+ {
35
+ "epoch": 0.5614035087719298,
36
+ "eval_loss": 0.8620332479476929,
37
+ "eval_runtime": 14.5037,
38
+ "eval_samples_per_second": 4.413,
39
+ "eval_steps_per_second": 4.413,
40
+ "step": 40
41
+ },
42
+ {
43
+ "epoch": 0.8421052631578947,
44
+ "grad_norm": 0.5583078861236572,
45
+ "learning_rate": 7.511961722488039e-05,
46
+ "loss": 0.8721,
47
+ "step": 60
48
+ },
49
+ {
50
+ "epoch": 0.8421052631578947,
51
+ "eval_loss": 0.8238226175308228,
52
+ "eval_runtime": 14.7369,
53
+ "eval_samples_per_second": 4.343,
54
+ "eval_steps_per_second": 4.343,
55
+ "step": 60
56
+ },
57
+ {
58
+ "epoch": 1.1122807017543859,
59
+ "grad_norm": 0.4012566804885864,
60
+ "learning_rate": 6.555023923444976e-05,
61
+ "loss": 0.8059,
62
+ "step": 80
63
+ },
64
+ {
65
+ "epoch": 1.1122807017543859,
66
+ "eval_loss": 0.8017938137054443,
67
+ "eval_runtime": 14.6894,
68
+ "eval_samples_per_second": 4.357,
69
+ "eval_steps_per_second": 4.357,
70
+ "step": 80
71
+ },
72
+ {
73
+ "epoch": 1.3929824561403508,
74
+ "grad_norm": 0.7860776782035828,
75
+ "learning_rate": 5.5980861244019145e-05,
76
+ "loss": 0.8141,
77
+ "step": 100
78
+ },
79
+ {
80
+ "epoch": 1.3929824561403508,
81
+ "eval_loss": 0.7868363857269287,
82
+ "eval_runtime": 14.7,
83
+ "eval_samples_per_second": 4.354,
84
+ "eval_steps_per_second": 4.354,
85
+ "step": 100
86
+ },
87
+ {
88
+ "epoch": 1.6736842105263157,
89
+ "grad_norm": 0.48762819170951843,
90
+ "learning_rate": 4.641148325358852e-05,
91
+ "loss": 0.7353,
92
+ "step": 120
93
+ },
94
+ {
95
+ "epoch": 1.6736842105263157,
96
+ "eval_loss": 0.7766555547714233,
97
+ "eval_runtime": 14.7914,
98
+ "eval_samples_per_second": 4.327,
99
+ "eval_steps_per_second": 4.327,
100
+ "step": 120
101
+ },
102
+ {
103
+ "epoch": 1.9543859649122806,
104
+ "grad_norm": 0.5413577556610107,
105
+ "learning_rate": 3.6842105263157895e-05,
106
+ "loss": 0.6779,
107
+ "step": 140
108
+ },
109
+ {
110
+ "epoch": 1.9543859649122806,
111
+ "eval_loss": 0.7647180557250977,
112
+ "eval_runtime": 14.7626,
113
+ "eval_samples_per_second": 4.335,
114
+ "eval_steps_per_second": 4.335,
115
+ "step": 140
116
+ },
117
+ {
118
+ "epoch": 2.2245614035087717,
119
+ "grad_norm": 0.8345220685005188,
120
+ "learning_rate": 2.7272727272727273e-05,
121
+ "loss": 0.6273,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 2.2245614035087717,
126
+ "eval_loss": 0.7629338502883911,
127
+ "eval_runtime": 14.7511,
128
+ "eval_samples_per_second": 4.339,
129
+ "eval_steps_per_second": 4.339,
130
+ "step": 160
131
+ },
132
+ {
133
+ "epoch": 2.5052631578947366,
134
+ "grad_norm": 0.5914349555969238,
135
+ "learning_rate": 1.770334928229665e-05,
136
+ "loss": 0.6983,
137
+ "step": 180
138
+ },
139
+ {
140
+ "epoch": 2.5052631578947366,
141
+ "eval_loss": 0.7597119808197021,
142
+ "eval_runtime": 14.7516,
143
+ "eval_samples_per_second": 4.339,
144
+ "eval_steps_per_second": 4.339,
145
+ "step": 180
146
+ },
147
+ {
148
+ "epoch": 2.7859649122807015,
149
+ "grad_norm": 0.5911722183227539,
150
+ "learning_rate": 8.133971291866028e-06,
151
+ "loss": 0.6958,
152
+ "step": 200
153
+ },
154
+ {
155
+ "epoch": 2.7859649122807015,
156
+ "eval_loss": 0.7568597793579102,
157
+ "eval_runtime": 14.6958,
158
+ "eval_samples_per_second": 4.355,
159
+ "eval_steps_per_second": 4.355,
160
+ "step": 200
161
+ },
162
+ {
163
+ "epoch": 3.0,
164
+ "step": 216,
165
+ "total_flos": 2.684496203948851e+16,
166
+ "train_loss": 0.7727184339805886,
167
+ "train_runtime": 1493.2791,
168
+ "train_samples_per_second": 1.145,
169
+ "train_steps_per_second": 0.145
170
+ },
171
+ {
172
+ "epoch": 3.0,
173
+ "eval_loss": 0.7568597793579102,
174
+ "eval_runtime": 14.794,
175
+ "eval_samples_per_second": 4.326,
176
+ "eval_steps_per_second": 4.326,
177
+ "step": 216
178
  }
179
  ],
180
  "logging_steps": 20,
 
197
  "should_epoch_stop": false,
198
  "should_evaluate": false,
199
  "should_log": false,
200
+ "should_save": true,
201
+ "should_training_stop": true
202
  },
203
  "attributes": {}
204
  }
205
  },
206
+ "total_flos": 2.684496203948851e+16,
207
  "train_batch_size": 1,
208
  "trial_name": null,
209
  "trial_params": null