|
{ |
|
"best_metric": 0.9638691322901849, |
|
"best_model_checkpoint": "mobilenet_v2_1.0_224-plant-disease2/checkpoint-948", |
|
"epoch": 5.990521327014218, |
|
"eval_steps": 500, |
|
"global_step": 948, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 5.263157894736842e-06, |
|
"loss": 3.7263, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.0526315789473684e-05, |
|
"loss": 3.7035, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.5789473684210526e-05, |
|
"loss": 3.6641, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 2.105263157894737e-05, |
|
"loss": 3.5615, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 2.6315789473684212e-05, |
|
"loss": 3.4717, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 3.157894736842105e-05, |
|
"loss": 3.3354, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 3.6842105263157895e-05, |
|
"loss": 3.1731, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.210526315789474e-05, |
|
"loss": 2.9675, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.736842105263158e-05, |
|
"loss": 2.6944, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.970691676436108e-05, |
|
"loss": 2.427, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.912075029308324e-05, |
|
"loss": 2.1254, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.8534583821805394e-05, |
|
"loss": 1.8635, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.794841735052755e-05, |
|
"loss": 1.6268, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 4.736225087924971e-05, |
|
"loss": 1.4245, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.6776084407971864e-05, |
|
"loss": 1.2171, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.8187766714082504, |
|
"eval_loss": 1.059464693069458, |
|
"eval_runtime": 29.4475, |
|
"eval_samples_per_second": 238.73, |
|
"eval_steps_per_second": 2.411, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.6189917936694024e-05, |
|
"loss": 1.1133, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.5603751465416176e-05, |
|
"loss": 1.012, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 4.501758499413834e-05, |
|
"loss": 0.8743, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 4.4431418522860494e-05, |
|
"loss": 0.8055, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 4.3845252051582654e-05, |
|
"loss": 0.7388, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 4.3259085580304806e-05, |
|
"loss": 0.6954, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.2672919109026965e-05, |
|
"loss": 0.6384, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 4.2086752637749124e-05, |
|
"loss": 0.5831, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.150058616647128e-05, |
|
"loss": 0.5561, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.0914419695193436e-05, |
|
"loss": 0.5289, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.0328253223915595e-05, |
|
"loss": 0.5063, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 3.9742086752637754e-05, |
|
"loss": 0.4788, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.915592028135991e-05, |
|
"loss": 0.4552, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.8569753810082066e-05, |
|
"loss": 0.4645, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 3.7983587338804225e-05, |
|
"loss": 0.4267, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 3.739742086752638e-05, |
|
"loss": 0.4082, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9386913229018492, |
|
"eval_loss": 0.3154248893260956, |
|
"eval_runtime": 29.2493, |
|
"eval_samples_per_second": 240.347, |
|
"eval_steps_per_second": 2.427, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 3.681125439624854e-05, |
|
"loss": 0.4, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 3.622508792497069e-05, |
|
"loss": 0.3832, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.5638921453692855e-05, |
|
"loss": 0.3888, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.505275498241501e-05, |
|
"loss": 0.3595, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 3.446658851113717e-05, |
|
"loss": 0.3454, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 3.388042203985932e-05, |
|
"loss": 0.3392, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.329425556858148e-05, |
|
"loss": 0.3466, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.270808909730364e-05, |
|
"loss": 0.345, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.212192262602579e-05, |
|
"loss": 0.3102, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 3.153575615474795e-05, |
|
"loss": 0.3045, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.094958968347011e-05, |
|
"loss": 0.3242, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 3.0363423212192264e-05, |
|
"loss": 0.3022, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.9777256740914423e-05, |
|
"loss": 0.3098, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.919109026963658e-05, |
|
"loss": 0.2845, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 2.8604923798358735e-05, |
|
"loss": 0.2807, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.801875732708089e-05, |
|
"loss": 0.295, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.9554765291607397, |
|
"eval_loss": 0.21907542645931244, |
|
"eval_runtime": 28.5527, |
|
"eval_samples_per_second": 246.211, |
|
"eval_steps_per_second": 2.487, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.7432590855803047e-05, |
|
"loss": 0.2667, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 2.6846424384525202e-05, |
|
"loss": 0.281, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 2.6260257913247365e-05, |
|
"loss": 0.268, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.567409144196952e-05, |
|
"loss": 0.2458, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.508792497069168e-05, |
|
"loss": 0.2493, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 2.4501758499413836e-05, |
|
"loss": 0.2607, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 2.391559202813599e-05, |
|
"loss": 0.2553, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 2.3329425556858147e-05, |
|
"loss": 0.2481, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 2.2743259085580307e-05, |
|
"loss": 0.2537, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 2.2157092614302462e-05, |
|
"loss": 0.2407, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 2.1570926143024618e-05, |
|
"loss": 0.2465, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 2.0984759671746777e-05, |
|
"loss": 0.2562, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 2.0398593200468937e-05, |
|
"loss": 0.2471, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 1.9812426729191092e-05, |
|
"loss": 0.2453, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.9226260257913248e-05, |
|
"loss": 0.2292, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 1.8640093786635404e-05, |
|
"loss": 0.2266, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.9594594594594594, |
|
"eval_loss": 0.1747303009033203, |
|
"eval_runtime": 28.6742, |
|
"eval_samples_per_second": 245.168, |
|
"eval_steps_per_second": 2.476, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 1.8053927315357563e-05, |
|
"loss": 0.2208, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 1.746776084407972e-05, |
|
"loss": 0.2339, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 1.6881594372801875e-05, |
|
"loss": 0.2395, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 1.6295427901524034e-05, |
|
"loss": 0.2314, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 1.5709261430246193e-05, |
|
"loss": 0.2208, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 1.5123094958968347e-05, |
|
"loss": 0.229, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 1.4536928487690505e-05, |
|
"loss": 0.2251, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 1.395076201641266e-05, |
|
"loss": 0.232, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 1.336459554513482e-05, |
|
"loss": 0.2342, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 1.2778429073856976e-05, |
|
"loss": 0.1995, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 1.2192262602579133e-05, |
|
"loss": 0.2119, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 1.160609613130129e-05, |
|
"loss": 0.2185, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 1.1019929660023446e-05, |
|
"loss": 0.2079, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 1.0433763188745604e-05, |
|
"loss": 0.2113, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 9.847596717467761e-06, |
|
"loss": 0.2177, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 9.261430246189919e-06, |
|
"loss": 0.2168, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.9499288762446657, |
|
"eval_loss": 0.21349148452281952, |
|
"eval_runtime": 28.3255, |
|
"eval_samples_per_second": 248.186, |
|
"eval_steps_per_second": 2.507, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 8.675263774912075e-06, |
|
"loss": 0.2072, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 8.089097303634232e-06, |
|
"loss": 0.2083, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 7.502930832356389e-06, |
|
"loss": 0.1997, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 6.916764361078546e-06, |
|
"loss": 0.1917, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 6.330597889800703e-06, |
|
"loss": 0.2098, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 5.7444314185228606e-06, |
|
"loss": 0.2001, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 5.158264947245018e-06, |
|
"loss": 0.2105, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 4.572098475967175e-06, |
|
"loss": 0.2168, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 3.985932004689332e-06, |
|
"loss": 0.2111, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 3.3997655334114893e-06, |
|
"loss": 0.1999, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 2.8135990621336464e-06, |
|
"loss": 0.2069, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 2.2274325908558035e-06, |
|
"loss": 0.2302, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 1.6412661195779603e-06, |
|
"loss": 0.2072, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 1.0550996483001172e-06, |
|
"loss": 0.2042, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 4.689331770222743e-07, |
|
"loss": 0.2091, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"eval_accuracy": 0.9638691322901849, |
|
"eval_loss": 0.15102514624595642, |
|
"eval_runtime": 27.9653, |
|
"eval_samples_per_second": 251.383, |
|
"eval_steps_per_second": 2.539, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"step": 948, |
|
"total_flos": 1.01609365487616e+18, |
|
"train_loss": 0.711250480468766, |
|
"train_runtime": 3427.9798, |
|
"train_samples_per_second": 110.733, |
|
"train_steps_per_second": 0.277 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 948, |
|
"num_train_epochs": 6, |
|
"save_steps": 500, |
|
"total_flos": 1.01609365487616e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|