|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.03576266594418857, |
|
"eval_steps": 50, |
|
"global_step": 99, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00036123904994129867, |
|
"grad_norm": 11.258881568908691, |
|
"learning_rate": 5e-05, |
|
"loss": 4.9805, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00036123904994129867, |
|
"eval_loss": 1.3654987812042236, |
|
"eval_runtime": 65.6871, |
|
"eval_samples_per_second": 17.751, |
|
"eval_steps_per_second": 8.875, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0007224780998825973, |
|
"grad_norm": 11.392064094543457, |
|
"learning_rate": 0.0001, |
|
"loss": 5.6516, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.001083717149823896, |
|
"grad_norm": 10.067161560058594, |
|
"learning_rate": 0.00015, |
|
"loss": 5.1962, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0014449561997651947, |
|
"grad_norm": 8.352842330932617, |
|
"learning_rate": 0.0002, |
|
"loss": 4.3996, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0018061952497064934, |
|
"grad_norm": 6.142587661743164, |
|
"learning_rate": 0.00025, |
|
"loss": 3.9857, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.002167434299647792, |
|
"grad_norm": 12.07518196105957, |
|
"learning_rate": 0.0003, |
|
"loss": 5.3352, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0025286733495890907, |
|
"grad_norm": 7.874082088470459, |
|
"learning_rate": 0.00035, |
|
"loss": 4.9328, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0028899123995303894, |
|
"grad_norm": 4.830751895904541, |
|
"learning_rate": 0.0004, |
|
"loss": 3.4086, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.003251151449471688, |
|
"grad_norm": 4.93334436416626, |
|
"learning_rate": 0.00045000000000000004, |
|
"loss": 3.7905, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0036123904994129867, |
|
"grad_norm": 5.134774684906006, |
|
"learning_rate": 0.0005, |
|
"loss": 4.1506, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.003973629549354285, |
|
"grad_norm": 5.855077266693115, |
|
"learning_rate": 0.0004998442655654946, |
|
"loss": 3.7257, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.004334868599295584, |
|
"grad_norm": 5.537413120269775, |
|
"learning_rate": 0.0004993772562876909, |
|
"loss": 4.3003, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.004696107649236883, |
|
"grad_norm": 5.094138145446777, |
|
"learning_rate": 0.0004985995540019955, |
|
"loss": 4.8794, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.005057346699178181, |
|
"grad_norm": 4.8982930183410645, |
|
"learning_rate": 0.0004975121276286136, |
|
"loss": 4.6231, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.00541858574911948, |
|
"grad_norm": 4.569957733154297, |
|
"learning_rate": 0.0004961163319653958, |
|
"loss": 4.0061, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.005779824799060779, |
|
"grad_norm": 7.189935207366943, |
|
"learning_rate": 0.0004944139059999286, |
|
"loss": 5.0499, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.006141063849002077, |
|
"grad_norm": 4.611532688140869, |
|
"learning_rate": 0.000492406970742972, |
|
"loss": 3.2843, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.006502302898943376, |
|
"grad_norm": 4.90513277053833, |
|
"learning_rate": 0.0004900980265859448, |
|
"loss": 3.4266, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.006863541948884675, |
|
"grad_norm": 6.561873435974121, |
|
"learning_rate": 0.0004874899501857477, |
|
"loss": 4.2261, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.007224780998825973, |
|
"grad_norm": 11.763420104980469, |
|
"learning_rate": 0.00048458599088080736, |
|
"loss": 3.9256, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.007586020048767272, |
|
"grad_norm": 5.913724422454834, |
|
"learning_rate": 0.0004813897666428053, |
|
"loss": 3.4543, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.00794725909870857, |
|
"grad_norm": 8.865498542785645, |
|
"learning_rate": 0.00047790525956913543, |
|
"loss": 5.0105, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.00830849814864987, |
|
"grad_norm": 6.202934265136719, |
|
"learning_rate": 0.0004741368109217071, |
|
"loss": 4.0654, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.008669737198591168, |
|
"grad_norm": 9.578519821166992, |
|
"learning_rate": 0.00047008911571827283, |
|
"loss": 5.3396, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.009030976248532467, |
|
"grad_norm": 7.616293430328369, |
|
"learning_rate": 0.00046576721688302105, |
|
"loss": 3.7428, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.009392215298473765, |
|
"grad_norm": 5.799994945526123, |
|
"learning_rate": 0.0004611764989637205, |
|
"loss": 4.0877, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.009753454348415064, |
|
"grad_norm": 5.58920955657959, |
|
"learning_rate": 0.0004563226814232444, |
|
"loss": 4.2916, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.010114693398356363, |
|
"grad_norm": 4.770939826965332, |
|
"learning_rate": 0.0004512118115138315, |
|
"loss": 4.2956, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.010475932448297661, |
|
"grad_norm": 4.9692063331604, |
|
"learning_rate": 0.0004458502567429631, |
|
"loss": 3.542, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.01083717149823896, |
|
"grad_norm": 4.753838539123535, |
|
"learning_rate": 0.00044024469694024196, |
|
"loss": 3.0264, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.011198410548180259, |
|
"grad_norm": 5.1347575187683105, |
|
"learning_rate": 0.00043440211593515554, |
|
"loss": 3.1049, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.011559649598121557, |
|
"grad_norm": 5.5637006759643555, |
|
"learning_rate": 0.0004283297928560951, |
|
"loss": 3.8163, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.011920888648062856, |
|
"grad_norm": 5.401363849639893, |
|
"learning_rate": 0.0004220352930614672, |
|
"loss": 3.8318, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.012282127698004155, |
|
"grad_norm": 5.22984504699707, |
|
"learning_rate": 0.00041552645871420013, |
|
"loss": 3.8126, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.012643366747945453, |
|
"grad_norm": 4.754082679748535, |
|
"learning_rate": 0.00040881139901138467, |
|
"loss": 4.0395, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.013004605797886752, |
|
"grad_norm": 4.820829391479492, |
|
"learning_rate": 0.00040189848008122475, |
|
"loss": 3.1665, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.01336584484782805, |
|
"grad_norm": 3.9806621074676514, |
|
"learning_rate": 0.00039479631455988334, |
|
"loss": 2.9394, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.01372708389776935, |
|
"grad_norm": 4.494911193847656, |
|
"learning_rate": 0.0003875137508612103, |
|
"loss": 3.1021, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.014088322947710648, |
|
"grad_norm": 4.835848331451416, |
|
"learning_rate": 0.00038005986215272055, |
|
"loss": 3.6138, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.014449561997651947, |
|
"grad_norm": 5.597693920135498, |
|
"learning_rate": 0.0003724439350515571, |
|
"loss": 4.5553, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.014810801047593245, |
|
"grad_norm": 4.290703773498535, |
|
"learning_rate": 0.0003646754580545226, |
|
"loss": 3.271, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.015172040097534544, |
|
"grad_norm": 4.479038238525391, |
|
"learning_rate": 0.000356764109716594, |
|
"loss": 3.1925, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.015533279147475843, |
|
"grad_norm": 4.6170654296875, |
|
"learning_rate": 0.00034871974659264783, |
|
"loss": 2.8589, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.01589451819741714, |
|
"grad_norm": 5.0329060554504395, |
|
"learning_rate": 0.0003405523909574206, |
|
"loss": 3.1158, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.01625575724735844, |
|
"grad_norm": 5.719956398010254, |
|
"learning_rate": 0.0003322722183190025, |
|
"loss": 3.856, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.01661699629729974, |
|
"grad_norm": 6.794640064239502, |
|
"learning_rate": 0.0003238895447414211, |
|
"loss": 4.4511, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.016978235347241036, |
|
"grad_norm": 3.9683847427368164, |
|
"learning_rate": 0.0003154148139921102, |
|
"loss": 2.8105, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.017339474397182336, |
|
"grad_norm": 5.173192024230957, |
|
"learning_rate": 0.00030685858453027663, |
|
"loss": 2.7872, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.017700713447123633, |
|
"grad_norm": 3.741673469543457, |
|
"learning_rate": 0.0002982315163523742, |
|
"loss": 2.6865, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.018061952497064934, |
|
"grad_norm": 3.9549710750579834, |
|
"learning_rate": 0.000289544357711076, |
|
"loss": 3.0372, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.018061952497064934, |
|
"eval_loss": 0.8259496092796326, |
|
"eval_runtime": 66.5192, |
|
"eval_samples_per_second": 17.529, |
|
"eval_steps_per_second": 8.764, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01842319154700623, |
|
"grad_norm": 5.567710876464844, |
|
"learning_rate": 0.0002808079317242896, |
|
"loss": 3.7515, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.01878443059694753, |
|
"grad_norm": 7.243776321411133, |
|
"learning_rate": 0.0002720331228909005, |
|
"loss": 2.6164, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.019145669646888828, |
|
"grad_norm": 4.8548431396484375, |
|
"learning_rate": 0.00026323086353004075, |
|
"loss": 3.2759, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.019506908696830128, |
|
"grad_norm": 3.919628858566284, |
|
"learning_rate": 0.0002544121201607822, |
|
"loss": 2.9809, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.019868147746771425, |
|
"grad_norm": 5.447530269622803, |
|
"learning_rate": 0.00024558787983921783, |
|
"loss": 3.5247, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.020229386796712726, |
|
"grad_norm": 5.87801456451416, |
|
"learning_rate": 0.0002367691364699592, |
|
"loss": 2.7521, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.020590625846654022, |
|
"grad_norm": 4.117478847503662, |
|
"learning_rate": 0.00022796687710909964, |
|
"loss": 3.7246, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.020951864896595323, |
|
"grad_norm": 3.548586845397949, |
|
"learning_rate": 0.00021919206827571036, |
|
"loss": 3.4873, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.02131310394653662, |
|
"grad_norm": 4.6662702560424805, |
|
"learning_rate": 0.00021045564228892402, |
|
"loss": 3.6514, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.02167434299647792, |
|
"grad_norm": 3.70032000541687, |
|
"learning_rate": 0.00020176848364762578, |
|
"loss": 3.4641, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.022035582046419217, |
|
"grad_norm": 5.118409633636475, |
|
"learning_rate": 0.00019314141546972343, |
|
"loss": 3.7131, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.022396821096360518, |
|
"grad_norm": 5.562695026397705, |
|
"learning_rate": 0.00018458518600788986, |
|
"loss": 3.2622, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.022758060146301815, |
|
"grad_norm": 6.675144195556641, |
|
"learning_rate": 0.00017611045525857898, |
|
"loss": 3.6492, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.023119299196243115, |
|
"grad_norm": 4.012448310852051, |
|
"learning_rate": 0.0001677277816809975, |
|
"loss": 2.7895, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.023480538246184412, |
|
"grad_norm": 3.4179670810699463, |
|
"learning_rate": 0.00015944760904257942, |
|
"loss": 2.6486, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.023841777296125712, |
|
"grad_norm": 5.845145225524902, |
|
"learning_rate": 0.0001512802534073522, |
|
"loss": 4.1475, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.02420301634606701, |
|
"grad_norm": 3.7966604232788086, |
|
"learning_rate": 0.00014323589028340596, |
|
"loss": 3.6578, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.02456425539600831, |
|
"grad_norm": 4.147678852081299, |
|
"learning_rate": 0.00013532454194547733, |
|
"loss": 3.165, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.024925494445949607, |
|
"grad_norm": 4.22127628326416, |
|
"learning_rate": 0.00012755606494844294, |
|
"loss": 3.6101, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.025286733495890907, |
|
"grad_norm": 3.980015516281128, |
|
"learning_rate": 0.00011994013784727947, |
|
"loss": 2.9809, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.025647972545832204, |
|
"grad_norm": 3.8896799087524414, |
|
"learning_rate": 0.00011248624913878966, |
|
"loss": 2.7593, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.026009211595773504, |
|
"grad_norm": 3.7454440593719482, |
|
"learning_rate": 0.0001052036854401166, |
|
"loss": 2.7478, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.0263704506457148, |
|
"grad_norm": 3.633699417114258, |
|
"learning_rate": 9.810151991877531e-05, |
|
"loss": 3.4857, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.0267316896956561, |
|
"grad_norm": 3.71138334274292, |
|
"learning_rate": 9.118860098861537e-05, |
|
"loss": 2.4003, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.0270929287455974, |
|
"grad_norm": 3.327180862426758, |
|
"learning_rate": 8.44735412857999e-05, |
|
"loss": 2.5128, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.0274541677955387, |
|
"grad_norm": 3.146794080734253, |
|
"learning_rate": 7.79647069385328e-05, |
|
"loss": 2.2593, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.027815406845479996, |
|
"grad_norm": 3.6162102222442627, |
|
"learning_rate": 7.167020714390501e-05, |
|
"loss": 2.586, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.028176645895421296, |
|
"grad_norm": 3.828462600708008, |
|
"learning_rate": 6.559788406484446e-05, |
|
"loss": 2.3276, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.028537884945362593, |
|
"grad_norm": 3.7188730239868164, |
|
"learning_rate": 5.975530305975807e-05, |
|
"loss": 3.0949, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.028899123995303894, |
|
"grad_norm": 4.314111709594727, |
|
"learning_rate": 5.414974325703686e-05, |
|
"loss": 2.537, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.02926036304524519, |
|
"grad_norm": 4.148589134216309, |
|
"learning_rate": 4.8788188486168616e-05, |
|
"loss": 2.833, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.02962160209518649, |
|
"grad_norm": 3.602301836013794, |
|
"learning_rate": 4.367731857675569e-05, |
|
"loss": 2.4284, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.029982841145127788, |
|
"grad_norm": 5.501028537750244, |
|
"learning_rate": 3.882350103627952e-05, |
|
"loss": 2.9707, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.03034408019506909, |
|
"grad_norm": 4.275007247924805, |
|
"learning_rate": 3.423278311697897e-05, |
|
"loss": 2.373, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.030705319245010385, |
|
"grad_norm": 3.906873941421509, |
|
"learning_rate": 2.9910884281727225e-05, |
|
"loss": 3.7784, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.031066558294951686, |
|
"grad_norm": 3.9153549671173096, |
|
"learning_rate": 2.586318907829291e-05, |
|
"loss": 3.0894, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.03142779734489298, |
|
"grad_norm": 4.450911521911621, |
|
"learning_rate": 2.209474043086457e-05, |
|
"loss": 3.2057, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.03178903639483428, |
|
"grad_norm": 3.1464052200317383, |
|
"learning_rate": 1.861023335719475e-05, |
|
"loss": 2.3528, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.03215027544477558, |
|
"grad_norm": 3.700227737426758, |
|
"learning_rate": 1.5414009119192633e-05, |
|
"loss": 2.8889, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.03251151449471688, |
|
"grad_norm": 3.8259425163269043, |
|
"learning_rate": 1.25100498142523e-05, |
|
"loss": 2.3349, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03287275354465818, |
|
"grad_norm": 4.0107316970825195, |
|
"learning_rate": 9.901973414055187e-06, |
|
"loss": 2.3647, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.03323399259459948, |
|
"grad_norm": 3.5053553581237793, |
|
"learning_rate": 7.593029257027956e-06, |
|
"loss": 3.2543, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.03359523164454078, |
|
"grad_norm": 3.6908042430877686, |
|
"learning_rate": 5.5860940000714015e-06, |
|
"loss": 2.9373, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.03395647069448207, |
|
"grad_norm": 3.165818214416504, |
|
"learning_rate": 3.8836680346041594e-06, |
|
"loss": 2.6173, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.03431770974442337, |
|
"grad_norm": 4.046987056732178, |
|
"learning_rate": 2.487872371386424e-06, |
|
"loss": 3.5724, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.03467894879436467, |
|
"grad_norm": 4.770341873168945, |
|
"learning_rate": 1.4004459980045125e-06, |
|
"loss": 3.2451, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.03504018784430597, |
|
"grad_norm": 3.3098256587982178, |
|
"learning_rate": 6.22743712309054e-07, |
|
"loss": 2.3355, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.035401426894247266, |
|
"grad_norm": 3.6350321769714355, |
|
"learning_rate": 1.557344345054501e-07, |
|
"loss": 3.2009, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.03576266594418857, |
|
"grad_norm": 4.967100620269775, |
|
"learning_rate": 0.0, |
|
"loss": 3.182, |
|
"step": 99 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 99, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 300, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.308168021016576e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|