Training in progress, step 400, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 411094576
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e653775cc249175678648a653044416209f6475bf29e43394e2ab2b03341ab5
|
3 |
size 411094576
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 209193780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2fe895de504a208490a1cf2ef58a932627f92684431ae1bdba71258bdcbb6bf
|
3 |
size 209193780
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22087d1b9076daa7f9d9ca391795de63314bd5b40201ba03c337e4e05bebf4a3
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:505f9225762b105f8ca5168f44d99b2f8467174f4ade85f1cc95f684fbd828e0
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"best_metric": 0.
|
3 |
-
"best_model_checkpoint": "miner_id_24/checkpoint-
|
4 |
-
"epoch": 0.
|
5 |
"eval_steps": 100,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -2139,6 +2139,714 @@
|
|
2139 |
"eval_samples_per_second": 19.65,
|
2140 |
"eval_steps_per_second": 4.912,
|
2141 |
"step": 300
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2142 |
}
|
2143 |
],
|
2144 |
"logging_steps": 1,
|
@@ -2167,7 +2875,7 @@
|
|
2167 |
"attributes": {}
|
2168 |
}
|
2169 |
},
|
2170 |
-
"total_flos":
|
2171 |
"train_batch_size": 8,
|
2172 |
"trial_name": null,
|
2173 |
"trial_params": null
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.8248791694641113,
|
3 |
+
"best_model_checkpoint": "miner_id_24/checkpoint-400",
|
4 |
+
"epoch": 0.005317363518233572,
|
5 |
"eval_steps": 100,
|
6 |
+
"global_step": 400,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
2139 |
"eval_samples_per_second": 19.65,
|
2140 |
"eval_steps_per_second": 4.912,
|
2141 |
"step": 300
|
2142 |
+
},
|
2143 |
+
{
|
2144 |
+
"epoch": 0.004001316047470763,
|
2145 |
+
"grad_norm": 0.9838849902153015,
|
2146 |
+
"learning_rate": 3.6743228687829595e-05,
|
2147 |
+
"loss": 0.9883,
|
2148 |
+
"step": 301
|
2149 |
+
},
|
2150 |
+
{
|
2151 |
+
"epoch": 0.0040146094562663465,
|
2152 |
+
"grad_norm": 1.8244836330413818,
|
2153 |
+
"learning_rate": 3.642797750674629e-05,
|
2154 |
+
"loss": 1.0104,
|
2155 |
+
"step": 302
|
2156 |
+
},
|
2157 |
+
{
|
2158 |
+
"epoch": 0.004027902865061931,
|
2159 |
+
"grad_norm": 1.7432421445846558,
|
2160 |
+
"learning_rate": 3.6113307705935396e-05,
|
2161 |
+
"loss": 1.0969,
|
2162 |
+
"step": 303
|
2163 |
+
},
|
2164 |
+
{
|
2165 |
+
"epoch": 0.004041196273857514,
|
2166 |
+
"grad_norm": 1.5802562236785889,
|
2167 |
+
"learning_rate": 3.579923276480387e-05,
|
2168 |
+
"loss": 0.9227,
|
2169 |
+
"step": 304
|
2170 |
+
},
|
2171 |
+
{
|
2172 |
+
"epoch": 0.004054489682653099,
|
2173 |
+
"grad_norm": 1.689244270324707,
|
2174 |
+
"learning_rate": 3.5485766137276894e-05,
|
2175 |
+
"loss": 0.8085,
|
2176 |
+
"step": 305
|
2177 |
+
},
|
2178 |
+
{
|
2179 |
+
"epoch": 0.004067783091448682,
|
2180 |
+
"grad_norm": 1.4541411399841309,
|
2181 |
+
"learning_rate": 3.5172921251221455e-05,
|
2182 |
+
"loss": 1.0265,
|
2183 |
+
"step": 306
|
2184 |
+
},
|
2185 |
+
{
|
2186 |
+
"epoch": 0.0040810765002442665,
|
2187 |
+
"grad_norm": 1.967889666557312,
|
2188 |
+
"learning_rate": 3.486071150787128e-05,
|
2189 |
+
"loss": 0.9648,
|
2190 |
+
"step": 307
|
2191 |
+
},
|
2192 |
+
{
|
2193 |
+
"epoch": 0.00409436990903985,
|
2194 |
+
"grad_norm": 1.8028714656829834,
|
2195 |
+
"learning_rate": 3.4549150281252636e-05,
|
2196 |
+
"loss": 1.1209,
|
2197 |
+
"step": 308
|
2198 |
+
},
|
2199 |
+
{
|
2200 |
+
"epoch": 0.004107663317835434,
|
2201 |
+
"grad_norm": 2.3967385292053223,
|
2202 |
+
"learning_rate": 3.423825091761153e-05,
|
2203 |
+
"loss": 0.9468,
|
2204 |
+
"step": 309
|
2205 |
+
},
|
2206 |
+
{
|
2207 |
+
"epoch": 0.004120956726631018,
|
2208 |
+
"grad_norm": 1.7051564455032349,
|
2209 |
+
"learning_rate": 3.392802673484193e-05,
|
2210 |
+
"loss": 0.8184,
|
2211 |
+
"step": 310
|
2212 |
+
},
|
2213 |
+
{
|
2214 |
+
"epoch": 0.004134250135426602,
|
2215 |
+
"grad_norm": 1.4805026054382324,
|
2216 |
+
"learning_rate": 3.361849102191533e-05,
|
2217 |
+
"loss": 1.0213,
|
2218 |
+
"step": 311
|
2219 |
+
},
|
2220 |
+
{
|
2221 |
+
"epoch": 0.004147543544222186,
|
2222 |
+
"grad_norm": 1.5888340473175049,
|
2223 |
+
"learning_rate": 3.330965703831146e-05,
|
2224 |
+
"loss": 0.9103,
|
2225 |
+
"step": 312
|
2226 |
+
},
|
2227 |
+
{
|
2228 |
+
"epoch": 0.00416083695301777,
|
2229 |
+
"grad_norm": 1.7905009984970093,
|
2230 |
+
"learning_rate": 3.300153801345028e-05,
|
2231 |
+
"loss": 0.9134,
|
2232 |
+
"step": 313
|
2233 |
+
},
|
2234 |
+
{
|
2235 |
+
"epoch": 0.004174130361813354,
|
2236 |
+
"grad_norm": 1.981239914894104,
|
2237 |
+
"learning_rate": 3.2694147146125345e-05,
|
2238 |
+
"loss": 0.9723,
|
2239 |
+
"step": 314
|
2240 |
+
},
|
2241 |
+
{
|
2242 |
+
"epoch": 0.004187423770608938,
|
2243 |
+
"grad_norm": 1.7049620151519775,
|
2244 |
+
"learning_rate": 3.2387497603938326e-05,
|
2245 |
+
"loss": 1.0175,
|
2246 |
+
"step": 315
|
2247 |
+
},
|
2248 |
+
{
|
2249 |
+
"epoch": 0.004200717179404522,
|
2250 |
+
"grad_norm": 1.575150966644287,
|
2251 |
+
"learning_rate": 3.2081602522734986e-05,
|
2252 |
+
"loss": 0.8881,
|
2253 |
+
"step": 316
|
2254 |
+
},
|
2255 |
+
{
|
2256 |
+
"epoch": 0.004214010588200106,
|
2257 |
+
"grad_norm": 1.5370079278945923,
|
2258 |
+
"learning_rate": 3.177647500604252e-05,
|
2259 |
+
"loss": 0.9056,
|
2260 |
+
"step": 317
|
2261 |
+
},
|
2262 |
+
{
|
2263 |
+
"epoch": 0.00422730399699569,
|
2264 |
+
"grad_norm": 1.9242894649505615,
|
2265 |
+
"learning_rate": 3.147212812450819e-05,
|
2266 |
+
"loss": 0.8755,
|
2267 |
+
"step": 318
|
2268 |
+
},
|
2269 |
+
{
|
2270 |
+
"epoch": 0.0042405974057912734,
|
2271 |
+
"grad_norm": 2.297606945037842,
|
2272 |
+
"learning_rate": 3.116857491533947e-05,
|
2273 |
+
"loss": 0.9929,
|
2274 |
+
"step": 319
|
2275 |
+
},
|
2276 |
+
{
|
2277 |
+
"epoch": 0.004253890814586858,
|
2278 |
+
"grad_norm": 1.8832775354385376,
|
2279 |
+
"learning_rate": 3.086582838174551e-05,
|
2280 |
+
"loss": 0.852,
|
2281 |
+
"step": 320
|
2282 |
+
},
|
2283 |
+
{
|
2284 |
+
"epoch": 0.004267184223382441,
|
2285 |
+
"grad_norm": 1.8884977102279663,
|
2286 |
+
"learning_rate": 3.056390149238022e-05,
|
2287 |
+
"loss": 0.9385,
|
2288 |
+
"step": 321
|
2289 |
+
},
|
2290 |
+
{
|
2291 |
+
"epoch": 0.004280477632178026,
|
2292 |
+
"grad_norm": 2.0682528018951416,
|
2293 |
+
"learning_rate": 3.0262807180786647e-05,
|
2294 |
+
"loss": 0.8438,
|
2295 |
+
"step": 322
|
2296 |
+
},
|
2297 |
+
{
|
2298 |
+
"epoch": 0.004293771040973609,
|
2299 |
+
"grad_norm": 2.1671676635742188,
|
2300 |
+
"learning_rate": 2.996255834484296e-05,
|
2301 |
+
"loss": 0.7709,
|
2302 |
+
"step": 323
|
2303 |
+
},
|
2304 |
+
{
|
2305 |
+
"epoch": 0.0043070644497691934,
|
2306 |
+
"grad_norm": 2.2052109241485596,
|
2307 |
+
"learning_rate": 2.9663167846209998e-05,
|
2308 |
+
"loss": 0.8592,
|
2309 |
+
"step": 324
|
2310 |
+
},
|
2311 |
+
{
|
2312 |
+
"epoch": 0.004320357858564777,
|
2313 |
+
"grad_norm": 1.8438620567321777,
|
2314 |
+
"learning_rate": 2.936464850978027e-05,
|
2315 |
+
"loss": 0.752,
|
2316 |
+
"step": 325
|
2317 |
+
},
|
2318 |
+
{
|
2319 |
+
"epoch": 0.004333651267360361,
|
2320 |
+
"grad_norm": 1.9005708694458008,
|
2321 |
+
"learning_rate": 2.9067013123128613e-05,
|
2322 |
+
"loss": 0.7647,
|
2323 |
+
"step": 326
|
2324 |
+
},
|
2325 |
+
{
|
2326 |
+
"epoch": 0.004346944676155945,
|
2327 |
+
"grad_norm": 2.098482847213745,
|
2328 |
+
"learning_rate": 2.8770274435964355e-05,
|
2329 |
+
"loss": 0.8867,
|
2330 |
+
"step": 327
|
2331 |
+
},
|
2332 |
+
{
|
2333 |
+
"epoch": 0.004360238084951529,
|
2334 |
+
"grad_norm": 2.2485973834991455,
|
2335 |
+
"learning_rate": 2.8474445159585235e-05,
|
2336 |
+
"loss": 0.701,
|
2337 |
+
"step": 328
|
2338 |
+
},
|
2339 |
+
{
|
2340 |
+
"epoch": 0.004373531493747113,
|
2341 |
+
"grad_norm": 2.077705144882202,
|
2342 |
+
"learning_rate": 2.8179537966332887e-05,
|
2343 |
+
"loss": 0.896,
|
2344 |
+
"step": 329
|
2345 |
+
},
|
2346 |
+
{
|
2347 |
+
"epoch": 0.004386824902542697,
|
2348 |
+
"grad_norm": 2.2783312797546387,
|
2349 |
+
"learning_rate": 2.7885565489049946e-05,
|
2350 |
+
"loss": 0.7303,
|
2351 |
+
"step": 330
|
2352 |
+
},
|
2353 |
+
{
|
2354 |
+
"epoch": 0.00440011831133828,
|
2355 |
+
"grad_norm": 2.306527853012085,
|
2356 |
+
"learning_rate": 2.759254032053888e-05,
|
2357 |
+
"loss": 0.817,
|
2358 |
+
"step": 331
|
2359 |
+
},
|
2360 |
+
{
|
2361 |
+
"epoch": 0.004413411720133865,
|
2362 |
+
"grad_norm": 2.3645269870758057,
|
2363 |
+
"learning_rate": 2.7300475013022663e-05,
|
2364 |
+
"loss": 0.8617,
|
2365 |
+
"step": 332
|
2366 |
+
},
|
2367 |
+
{
|
2368 |
+
"epoch": 0.004426705128929448,
|
2369 |
+
"grad_norm": 2.2920470237731934,
|
2370 |
+
"learning_rate": 2.700938207760701e-05,
|
2371 |
+
"loss": 0.873,
|
2372 |
+
"step": 333
|
2373 |
+
},
|
2374 |
+
{
|
2375 |
+
"epoch": 0.004439998537725033,
|
2376 |
+
"grad_norm": 2.666349172592163,
|
2377 |
+
"learning_rate": 2.671927398374443e-05,
|
2378 |
+
"loss": 0.9184,
|
2379 |
+
"step": 334
|
2380 |
+
},
|
2381 |
+
{
|
2382 |
+
"epoch": 0.004453291946520616,
|
2383 |
+
"grad_norm": 2.1982336044311523,
|
2384 |
+
"learning_rate": 2.6430163158700115e-05,
|
2385 |
+
"loss": 0.9704,
|
2386 |
+
"step": 335
|
2387 |
+
},
|
2388 |
+
{
|
2389 |
+
"epoch": 0.0044665853553162,
|
2390 |
+
"grad_norm": 2.6879024505615234,
|
2391 |
+
"learning_rate": 2.6142061987019577e-05,
|
2392 |
+
"loss": 0.6279,
|
2393 |
+
"step": 336
|
2394 |
+
},
|
2395 |
+
{
|
2396 |
+
"epoch": 0.004479878764111784,
|
2397 |
+
"grad_norm": 2.241543769836426,
|
2398 |
+
"learning_rate": 2.5854982809998153e-05,
|
2399 |
+
"loss": 0.7901,
|
2400 |
+
"step": 337
|
2401 |
+
},
|
2402 |
+
{
|
2403 |
+
"epoch": 0.004493172172907368,
|
2404 |
+
"grad_norm": 2.5128378868103027,
|
2405 |
+
"learning_rate": 2.556893792515227e-05,
|
2406 |
+
"loss": 0.7943,
|
2407 |
+
"step": 338
|
2408 |
+
},
|
2409 |
+
{
|
2410 |
+
"epoch": 0.004506465581702952,
|
2411 |
+
"grad_norm": 2.971026659011841,
|
2412 |
+
"learning_rate": 2.5283939585692783e-05,
|
2413 |
+
"loss": 0.7436,
|
2414 |
+
"step": 339
|
2415 |
+
},
|
2416 |
+
{
|
2417 |
+
"epoch": 0.004519758990498536,
|
2418 |
+
"grad_norm": 2.6545255184173584,
|
2419 |
+
"learning_rate": 2.500000000000001e-05,
|
2420 |
+
"loss": 0.8035,
|
2421 |
+
"step": 340
|
2422 |
+
},
|
2423 |
+
{
|
2424 |
+
"epoch": 0.00453305239929412,
|
2425 |
+
"grad_norm": 3.0109236240386963,
|
2426 |
+
"learning_rate": 2.471713133110078e-05,
|
2427 |
+
"loss": 0.892,
|
2428 |
+
"step": 341
|
2429 |
+
},
|
2430 |
+
{
|
2431 |
+
"epoch": 0.004546345808089704,
|
2432 |
+
"grad_norm": 2.4946107864379883,
|
2433 |
+
"learning_rate": 2.4435345696147403e-05,
|
2434 |
+
"loss": 0.6965,
|
2435 |
+
"step": 342
|
2436 |
+
},
|
2437 |
+
{
|
2438 |
+
"epoch": 0.004559639216885288,
|
2439 |
+
"grad_norm": 3.106733798980713,
|
2440 |
+
"learning_rate": 2.4154655165898627e-05,
|
2441 |
+
"loss": 0.8695,
|
2442 |
+
"step": 343
|
2443 |
+
},
|
2444 |
+
{
|
2445 |
+
"epoch": 0.004572932625680872,
|
2446 |
+
"grad_norm": 3.4859445095062256,
|
2447 |
+
"learning_rate": 2.3875071764202563e-05,
|
2448 |
+
"loss": 0.7741,
|
2449 |
+
"step": 344
|
2450 |
+
},
|
2451 |
+
{
|
2452 |
+
"epoch": 0.004586226034476456,
|
2453 |
+
"grad_norm": 3.2647345066070557,
|
2454 |
+
"learning_rate": 2.3596607467481603e-05,
|
2455 |
+
"loss": 0.8903,
|
2456 |
+
"step": 345
|
2457 |
+
},
|
2458 |
+
{
|
2459 |
+
"epoch": 0.0045995194432720396,
|
2460 |
+
"grad_norm": 3.4573628902435303,
|
2461 |
+
"learning_rate": 2.3319274204219428e-05,
|
2462 |
+
"loss": 0.8457,
|
2463 |
+
"step": 346
|
2464 |
+
},
|
2465 |
+
{
|
2466 |
+
"epoch": 0.004612812852067624,
|
2467 |
+
"grad_norm": 4.146410942077637,
|
2468 |
+
"learning_rate": 2.3043083854449988e-05,
|
2469 |
+
"loss": 0.8782,
|
2470 |
+
"step": 347
|
2471 |
+
},
|
2472 |
+
{
|
2473 |
+
"epoch": 0.004626106260863207,
|
2474 |
+
"grad_norm": 3.767638683319092,
|
2475 |
+
"learning_rate": 2.2768048249248648e-05,
|
2476 |
+
"loss": 1.0579,
|
2477 |
+
"step": 348
|
2478 |
+
},
|
2479 |
+
{
|
2480 |
+
"epoch": 0.004639399669658792,
|
2481 |
+
"grad_norm": 4.626466751098633,
|
2482 |
+
"learning_rate": 2.2494179170225333e-05,
|
2483 |
+
"loss": 0.6208,
|
2484 |
+
"step": 349
|
2485 |
+
},
|
2486 |
+
{
|
2487 |
+
"epoch": 0.004652693078454375,
|
2488 |
+
"grad_norm": 5.113692760467529,
|
2489 |
+
"learning_rate": 2.2221488349019903e-05,
|
2490 |
+
"loss": 0.8458,
|
2491 |
+
"step": 350
|
2492 |
+
},
|
2493 |
+
{
|
2494 |
+
"epoch": 0.00466598648724996,
|
2495 |
+
"grad_norm": 1.0921778678894043,
|
2496 |
+
"learning_rate": 2.194998746679952e-05,
|
2497 |
+
"loss": 1.0162,
|
2498 |
+
"step": 351
|
2499 |
+
},
|
2500 |
+
{
|
2501 |
+
"epoch": 0.004679279896045543,
|
2502 |
+
"grad_norm": 1.255760669708252,
|
2503 |
+
"learning_rate": 2.167968815375837e-05,
|
2504 |
+
"loss": 0.9845,
|
2505 |
+
"step": 352
|
2506 |
+
},
|
2507 |
+
{
|
2508 |
+
"epoch": 0.004692573304841127,
|
2509 |
+
"grad_norm": 1.47697114944458,
|
2510 |
+
"learning_rate": 2.1410601988619394e-05,
|
2511 |
+
"loss": 1.0847,
|
2512 |
+
"step": 353
|
2513 |
+
},
|
2514 |
+
{
|
2515 |
+
"epoch": 0.004705866713636711,
|
2516 |
+
"grad_norm": 1.4926693439483643,
|
2517 |
+
"learning_rate": 2.1142740498138324e-05,
|
2518 |
+
"loss": 0.89,
|
2519 |
+
"step": 354
|
2520 |
+
},
|
2521 |
+
{
|
2522 |
+
"epoch": 0.004719160122432295,
|
2523 |
+
"grad_norm": 1.4988641738891602,
|
2524 |
+
"learning_rate": 2.08761151566099e-05,
|
2525 |
+
"loss": 0.9026,
|
2526 |
+
"step": 355
|
2527 |
+
},
|
2528 |
+
{
|
2529 |
+
"epoch": 0.004732453531227879,
|
2530 |
+
"grad_norm": 1.8952800035476685,
|
2531 |
+
"learning_rate": 2.061073738537635e-05,
|
2532 |
+
"loss": 0.8742,
|
2533 |
+
"step": 356
|
2534 |
+
},
|
2535 |
+
{
|
2536 |
+
"epoch": 0.004745746940023463,
|
2537 |
+
"grad_norm": 1.5465558767318726,
|
2538 |
+
"learning_rate": 2.034661855233815e-05,
|
2539 |
+
"loss": 0.9852,
|
2540 |
+
"step": 357
|
2541 |
+
},
|
2542 |
+
{
|
2543 |
+
"epoch": 0.0047590403488190465,
|
2544 |
+
"grad_norm": 1.766453504562378,
|
2545 |
+
"learning_rate": 2.008376997146705e-05,
|
2546 |
+
"loss": 0.9855,
|
2547 |
+
"step": 358
|
2548 |
+
},
|
2549 |
+
{
|
2550 |
+
"epoch": 0.004772333757614631,
|
2551 |
+
"grad_norm": 1.4662683010101318,
|
2552 |
+
"learning_rate": 1.982220290232143e-05,
|
2553 |
+
"loss": 1.0793,
|
2554 |
+
"step": 359
|
2555 |
+
},
|
2556 |
+
{
|
2557 |
+
"epoch": 0.004785627166410214,
|
2558 |
+
"grad_norm": 1.4013056755065918,
|
2559 |
+
"learning_rate": 1.9561928549563968e-05,
|
2560 |
+
"loss": 0.9454,
|
2561 |
+
"step": 360
|
2562 |
+
},
|
2563 |
+
{
|
2564 |
+
"epoch": 0.004798920575205799,
|
2565 |
+
"grad_norm": 1.6331448554992676,
|
2566 |
+
"learning_rate": 1.9302958062481673e-05,
|
2567 |
+
"loss": 0.8652,
|
2568 |
+
"step": 361
|
2569 |
+
},
|
2570 |
+
{
|
2571 |
+
"epoch": 0.004812213984001382,
|
2572 |
+
"grad_norm": 1.8215956687927246,
|
2573 |
+
"learning_rate": 1.9045302534508297e-05,
|
2574 |
+
"loss": 0.8735,
|
2575 |
+
"step": 362
|
2576 |
+
},
|
2577 |
+
{
|
2578 |
+
"epoch": 0.0048255073927969665,
|
2579 |
+
"grad_norm": 1.5076251029968262,
|
2580 |
+
"learning_rate": 1.8788973002749112e-05,
|
2581 |
+
"loss": 0.8257,
|
2582 |
+
"step": 363
|
2583 |
+
},
|
2584 |
+
{
|
2585 |
+
"epoch": 0.00483880080159255,
|
2586 |
+
"grad_norm": 1.6483572721481323,
|
2587 |
+
"learning_rate": 1.8533980447508137e-05,
|
2588 |
+
"loss": 0.8008,
|
2589 |
+
"step": 364
|
2590 |
+
},
|
2591 |
+
{
|
2592 |
+
"epoch": 0.004852094210388134,
|
2593 |
+
"grad_norm": 1.7253248691558838,
|
2594 |
+
"learning_rate": 1.8280335791817733e-05,
|
2595 |
+
"loss": 0.8845,
|
2596 |
+
"step": 365
|
2597 |
+
},
|
2598 |
+
{
|
2599 |
+
"epoch": 0.004865387619183718,
|
2600 |
+
"grad_norm": 1.6659144163131714,
|
2601 |
+
"learning_rate": 1.8028049900970767e-05,
|
2602 |
+
"loss": 0.8973,
|
2603 |
+
"step": 366
|
2604 |
+
},
|
2605 |
+
{
|
2606 |
+
"epoch": 0.004878681027979302,
|
2607 |
+
"grad_norm": 1.8108363151550293,
|
2608 |
+
"learning_rate": 1.777713358205514e-05,
|
2609 |
+
"loss": 0.8601,
|
2610 |
+
"step": 367
|
2611 |
+
},
|
2612 |
+
{
|
2613 |
+
"epoch": 0.004891974436774886,
|
2614 |
+
"grad_norm": 1.6906031370162964,
|
2615 |
+
"learning_rate": 1.7527597583490822e-05,
|
2616 |
+
"loss": 0.8842,
|
2617 |
+
"step": 368
|
2618 |
+
},
|
2619 |
+
{
|
2620 |
+
"epoch": 0.00490526784557047,
|
2621 |
+
"grad_norm": 1.9413644075393677,
|
2622 |
+
"learning_rate": 1.7279452594569483e-05,
|
2623 |
+
"loss": 0.8367,
|
2624 |
+
"step": 369
|
2625 |
+
},
|
2626 |
+
{
|
2627 |
+
"epoch": 0.004918561254366054,
|
2628 |
+
"grad_norm": 1.9167553186416626,
|
2629 |
+
"learning_rate": 1.703270924499656e-05,
|
2630 |
+
"loss": 0.8102,
|
2631 |
+
"step": 370
|
2632 |
+
},
|
2633 |
+
{
|
2634 |
+
"epoch": 0.004931854663161638,
|
2635 |
+
"grad_norm": 1.9133857488632202,
|
2636 |
+
"learning_rate": 1.678737810443593e-05,
|
2637 |
+
"loss": 0.9387,
|
2638 |
+
"step": 371
|
2639 |
+
},
|
2640 |
+
{
|
2641 |
+
"epoch": 0.004945148071957222,
|
2642 |
+
"grad_norm": 1.723745584487915,
|
2643 |
+
"learning_rate": 1.6543469682057106e-05,
|
2644 |
+
"loss": 0.8633,
|
2645 |
+
"step": 372
|
2646 |
+
},
|
2647 |
+
{
|
2648 |
+
"epoch": 0.004958441480752806,
|
2649 |
+
"grad_norm": 2.0419116020202637,
|
2650 |
+
"learning_rate": 1.6300994426085103e-05,
|
2651 |
+
"loss": 0.8323,
|
2652 |
+
"step": 373
|
2653 |
+
},
|
2654 |
+
{
|
2655 |
+
"epoch": 0.00497173488954839,
|
2656 |
+
"grad_norm": 1.9794385433197021,
|
2657 |
+
"learning_rate": 1.605996272335291e-05,
|
2658 |
+
"loss": 0.8905,
|
2659 |
+
"step": 374
|
2660 |
+
},
|
2661 |
+
{
|
2662 |
+
"epoch": 0.0049850282983439735,
|
2663 |
+
"grad_norm": 2.3014209270477295,
|
2664 |
+
"learning_rate": 1.5820384898856434e-05,
|
2665 |
+
"loss": 0.9045,
|
2666 |
+
"step": 375
|
2667 |
+
},
|
2668 |
+
{
|
2669 |
+
"epoch": 0.004998321707139558,
|
2670 |
+
"grad_norm": 1.863885521888733,
|
2671 |
+
"learning_rate": 1.5582271215312294e-05,
|
2672 |
+
"loss": 0.7435,
|
2673 |
+
"step": 376
|
2674 |
+
},
|
2675 |
+
{
|
2676 |
+
"epoch": 0.005011615115935141,
|
2677 |
+
"grad_norm": 2.2184691429138184,
|
2678 |
+
"learning_rate": 1.5345631872718214e-05,
|
2679 |
+
"loss": 0.8444,
|
2680 |
+
"step": 377
|
2681 |
+
},
|
2682 |
+
{
|
2683 |
+
"epoch": 0.005024908524730726,
|
2684 |
+
"grad_norm": 2.174875497817993,
|
2685 |
+
"learning_rate": 1.5110477007916001e-05,
|
2686 |
+
"loss": 0.9467,
|
2687 |
+
"step": 378
|
2688 |
+
},
|
2689 |
+
{
|
2690 |
+
"epoch": 0.005038201933526309,
|
2691 |
+
"grad_norm": 2.121537446975708,
|
2692 |
+
"learning_rate": 1.4876816694157419e-05,
|
2693 |
+
"loss": 0.9535,
|
2694 |
+
"step": 379
|
2695 |
+
},
|
2696 |
+
{
|
2697 |
+
"epoch": 0.0050514953423218935,
|
2698 |
+
"grad_norm": 2.0780320167541504,
|
2699 |
+
"learning_rate": 1.4644660940672627e-05,
|
2700 |
+
"loss": 0.8615,
|
2701 |
+
"step": 380
|
2702 |
+
},
|
2703 |
+
{
|
2704 |
+
"epoch": 0.005064788751117477,
|
2705 |
+
"grad_norm": 1.9616137742996216,
|
2706 |
+
"learning_rate": 1.4414019692241437e-05,
|
2707 |
+
"loss": 0.8279,
|
2708 |
+
"step": 381
|
2709 |
+
},
|
2710 |
+
{
|
2711 |
+
"epoch": 0.005078082159913061,
|
2712 |
+
"grad_norm": 2.3762927055358887,
|
2713 |
+
"learning_rate": 1.4184902828767287e-05,
|
2714 |
+
"loss": 0.7334,
|
2715 |
+
"step": 382
|
2716 |
+
},
|
2717 |
+
{
|
2718 |
+
"epoch": 0.005091375568708645,
|
2719 |
+
"grad_norm": 3.85578989982605,
|
2720 |
+
"learning_rate": 1.3957320164854059e-05,
|
2721 |
+
"loss": 0.9445,
|
2722 |
+
"step": 383
|
2723 |
+
},
|
2724 |
+
{
|
2725 |
+
"epoch": 0.005104668977504229,
|
2726 |
+
"grad_norm": 2.578687906265259,
|
2727 |
+
"learning_rate": 1.373128144938563e-05,
|
2728 |
+
"loss": 0.7616,
|
2729 |
+
"step": 384
|
2730 |
+
},
|
2731 |
+
{
|
2732 |
+
"epoch": 0.005117962386299813,
|
2733 |
+
"grad_norm": 2.467428684234619,
|
2734 |
+
"learning_rate": 1.3506796365108232e-05,
|
2735 |
+
"loss": 0.8458,
|
2736 |
+
"step": 385
|
2737 |
+
},
|
2738 |
+
{
|
2739 |
+
"epoch": 0.005131255795095397,
|
2740 |
+
"grad_norm": 2.394502639770508,
|
2741 |
+
"learning_rate": 1.3283874528215733e-05,
|
2742 |
+
"loss": 0.7064,
|
2743 |
+
"step": 386
|
2744 |
+
},
|
2745 |
+
{
|
2746 |
+
"epoch": 0.0051445492038909805,
|
2747 |
+
"grad_norm": 2.236889600753784,
|
2748 |
+
"learning_rate": 1.3062525487937699e-05,
|
2749 |
+
"loss": 0.8134,
|
2750 |
+
"step": 387
|
2751 |
+
},
|
2752 |
+
{
|
2753 |
+
"epoch": 0.005157842612686565,
|
2754 |
+
"grad_norm": 2.4321963787078857,
|
2755 |
+
"learning_rate": 1.2842758726130283e-05,
|
2756 |
+
"loss": 0.7657,
|
2757 |
+
"step": 388
|
2758 |
+
},
|
2759 |
+
{
|
2760 |
+
"epoch": 0.005171136021482148,
|
2761 |
+
"grad_norm": 2.509622097015381,
|
2762 |
+
"learning_rate": 1.2624583656870154e-05,
|
2763 |
+
"loss": 0.8433,
|
2764 |
+
"step": 389
|
2765 |
+
},
|
2766 |
+
{
|
2767 |
+
"epoch": 0.005184429430277733,
|
2768 |
+
"grad_norm": 2.68182373046875,
|
2769 |
+
"learning_rate": 1.2408009626051137e-05,
|
2770 |
+
"loss": 0.7196,
|
2771 |
+
"step": 390
|
2772 |
+
},
|
2773 |
+
{
|
2774 |
+
"epoch": 0.005197722839073316,
|
2775 |
+
"grad_norm": 2.600435495376587,
|
2776 |
+
"learning_rate": 1.2193045910983863e-05,
|
2777 |
+
"loss": 0.8124,
|
2778 |
+
"step": 391
|
2779 |
+
},
|
2780 |
+
{
|
2781 |
+
"epoch": 0.0052110162478689005,
|
2782 |
+
"grad_norm": 3.4181008338928223,
|
2783 |
+
"learning_rate": 1.1979701719998453e-05,
|
2784 |
+
"loss": 0.8008,
|
2785 |
+
"step": 392
|
2786 |
+
},
|
2787 |
+
{
|
2788 |
+
"epoch": 0.005224309656664484,
|
2789 |
+
"grad_norm": 2.620757579803467,
|
2790 |
+
"learning_rate": 1.1767986192049984e-05,
|
2791 |
+
"loss": 0.6854,
|
2792 |
+
"step": 393
|
2793 |
+
},
|
2794 |
+
{
|
2795 |
+
"epoch": 0.005237603065460068,
|
2796 |
+
"grad_norm": 2.952314853668213,
|
2797 |
+
"learning_rate": 1.1557908396327028e-05,
|
2798 |
+
"loss": 0.8588,
|
2799 |
+
"step": 394
|
2800 |
+
},
|
2801 |
+
{
|
2802 |
+
"epoch": 0.005250896474255652,
|
2803 |
+
"grad_norm": 3.42287015914917,
|
2804 |
+
"learning_rate": 1.134947733186315e-05,
|
2805 |
+
"loss": 0.7585,
|
2806 |
+
"step": 395
|
2807 |
+
},
|
2808 |
+
{
|
2809 |
+
"epoch": 0.005264189883051236,
|
2810 |
+
"grad_norm": 4.159125328063965,
|
2811 |
+
"learning_rate": 1.1142701927151456e-05,
|
2812 |
+
"loss": 0.823,
|
2813 |
+
"step": 396
|
2814 |
+
},
|
2815 |
+
{
|
2816 |
+
"epoch": 0.00527748329184682,
|
2817 |
+
"grad_norm": 2.9435524940490723,
|
2818 |
+
"learning_rate": 1.0937591039762085e-05,
|
2819 |
+
"loss": 0.6073,
|
2820 |
+
"step": 397
|
2821 |
+
},
|
2822 |
+
{
|
2823 |
+
"epoch": 0.005290776700642404,
|
2824 |
+
"grad_norm": 3.4450318813323975,
|
2825 |
+
"learning_rate": 1.0734153455962765e-05,
|
2826 |
+
"loss": 0.8219,
|
2827 |
+
"step": 398
|
2828 |
+
},
|
2829 |
+
{
|
2830 |
+
"epoch": 0.005304070109437988,
|
2831 |
+
"grad_norm": 4.6743621826171875,
|
2832 |
+
"learning_rate": 1.0532397890342505e-05,
|
2833 |
+
"loss": 0.8998,
|
2834 |
+
"step": 399
|
2835 |
+
},
|
2836 |
+
{
|
2837 |
+
"epoch": 0.005317363518233572,
|
2838 |
+
"grad_norm": 6.201481819152832,
|
2839 |
+
"learning_rate": 1.0332332985438248e-05,
|
2840 |
+
"loss": 0.673,
|
2841 |
+
"step": 400
|
2842 |
+
},
|
2843 |
+
{
|
2844 |
+
"epoch": 0.005317363518233572,
|
2845 |
+
"eval_loss": 0.8248791694641113,
|
2846 |
+
"eval_runtime": 6450.9082,
|
2847 |
+
"eval_samples_per_second": 19.64,
|
2848 |
+
"eval_steps_per_second": 4.91,
|
2849 |
+
"step": 400
|
2850 |
}
|
2851 |
],
|
2852 |
"logging_steps": 1,
|
|
|
2875 |
"attributes": {}
|
2876 |
}
|
2877 |
},
|
2878 |
+
"total_flos": 2.506940810621092e+17,
|
2879 |
"train_batch_size": 8,
|
2880 |
"trial_name": null,
|
2881 |
"trial_params": null
|