| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0775354794046383, | |
| "eval_steps": 500, | |
| "global_step": 2250, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.009230414214837892, | |
| "grad_norm": 0.0994892492890358, | |
| "learning_rate": 0.0002, | |
| "loss": 1.5722, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.018460828429675783, | |
| "grad_norm": 0.062255557626485825, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2813, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.027691242644513673, | |
| "grad_norm": 0.12040963768959045, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3037, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.036921656859351566, | |
| "grad_norm": 0.04763123765587807, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3666, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.046152071074189456, | |
| "grad_norm": 0.12753669917583466, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1438, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.055382485289027346, | |
| "grad_norm": 0.10537329316139221, | |
| "learning_rate": 0.0002, | |
| "loss": 1.4241, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.06461289950386524, | |
| "grad_norm": 0.03980427607893944, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2929, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.07384331371870313, | |
| "grad_norm": 0.12119753658771515, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2938, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.08307372793354102, | |
| "grad_norm": 0.062459882348775864, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3273, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.09230414214837891, | |
| "grad_norm": 0.15507832169532776, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1885, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.1015345563632168, | |
| "grad_norm": 0.09834089875221252, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3693, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.11076497057805469, | |
| "grad_norm": 0.05203542485833168, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2587, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.11999538479289258, | |
| "grad_norm": 0.12015814334154129, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2815, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.12922579900773049, | |
| "grad_norm": 0.06197419390082359, | |
| "learning_rate": 0.0002, | |
| "loss": 1.429, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.13845621322256838, | |
| "grad_norm": 0.17958974838256836, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0941, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.14768662743740626, | |
| "grad_norm": 0.09950366616249084, | |
| "learning_rate": 0.0002, | |
| "loss": 1.4066, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.15691704165224415, | |
| "grad_norm": 0.050984274595975876, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2131, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.16614745586708204, | |
| "grad_norm": 0.150551900267601, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2422, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.17537787008191993, | |
| "grad_norm": 0.06124914437532425, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3434, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.18460828429675782, | |
| "grad_norm": 0.17724090814590454, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0898, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.19383869851159571, | |
| "grad_norm": 0.0937797948718071, | |
| "learning_rate": 0.0002, | |
| "loss": 1.432, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.2030691127264336, | |
| "grad_norm": 0.059550438076257706, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1605, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.2122995269412715, | |
| "grad_norm": 0.10598563402891159, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2004, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.22152994115610938, | |
| "grad_norm": 0.059932854026556015, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2944, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.23076035537094727, | |
| "grad_norm": 0.16503147780895233, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1301, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.23999076958578516, | |
| "grad_norm": 0.08444368839263916, | |
| "learning_rate": 0.0002, | |
| "loss": 1.4195, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.24922118380062305, | |
| "grad_norm": 0.05936718359589577, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2805, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.25845159801546097, | |
| "grad_norm": 0.12579703330993652, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2394, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.26768201223029886, | |
| "grad_norm": 0.06689989566802979, | |
| "learning_rate": 0.0002, | |
| "loss": 1.298, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.27691242644513675, | |
| "grad_norm": 0.17793583869934082, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0807, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.28614284065997464, | |
| "grad_norm": 0.08275260776281357, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3433, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.29537325487481253, | |
| "grad_norm": 0.05578906834125519, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2197, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.3046036690896504, | |
| "grad_norm": 0.13691303133964539, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2487, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.3138340833044883, | |
| "grad_norm": 0.055021870881319046, | |
| "learning_rate": 0.0002, | |
| "loss": 1.275, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.3230644975193262, | |
| "grad_norm": 0.1495254933834076, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0445, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3322949117341641, | |
| "grad_norm": 0.09207426011562347, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3997, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.341525325949002, | |
| "grad_norm": 0.05421067774295807, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1986, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.35075574016383987, | |
| "grad_norm": 0.12197154760360718, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2589, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.35998615437867776, | |
| "grad_norm": 0.06101464852690697, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3766, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.36921656859351565, | |
| "grad_norm": 0.19063565135002136, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0677, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.37844698280835354, | |
| "grad_norm": 0.09075415134429932, | |
| "learning_rate": 0.0002, | |
| "loss": 1.4032, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.38767739702319143, | |
| "grad_norm": 0.053780850023031235, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2328, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.3969078112380293, | |
| "grad_norm": 0.11555945128202438, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2205, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.4061382254528672, | |
| "grad_norm": 0.0680965855717659, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2906, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.4153686396677051, | |
| "grad_norm": 0.17800922691822052, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0823, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.424599053882543, | |
| "grad_norm": 0.10128472000360489, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3619, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.4338294680973809, | |
| "grad_norm": 0.049957141280174255, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1817, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.44305988231221877, | |
| "grad_norm": 0.1359386295080185, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2062, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.45229029652705666, | |
| "grad_norm": 0.0684947818517685, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3084, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.46152071074189455, | |
| "grad_norm": 0.1941768378019333, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0832, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.47075112495673244, | |
| "grad_norm": 0.09089575707912445, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3617, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.4799815391715703, | |
| "grad_norm": 0.06254442036151886, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1961, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.4892119533864082, | |
| "grad_norm": 0.15948618948459625, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2078, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.4984423676012461, | |
| "grad_norm": 0.06899358332157135, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3514, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.507672781816084, | |
| "grad_norm": 0.16271327435970306, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0975, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.5169031960309219, | |
| "grad_norm": 0.0963057279586792, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3987, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.5261336102457598, | |
| "grad_norm": 0.05728481337428093, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1942, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.5353640244605977, | |
| "grad_norm": 0.1385851353406906, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1708, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5445944386754356, | |
| "grad_norm": 0.06655001640319824, | |
| "learning_rate": 0.0002, | |
| "loss": 1.339, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.5538248528902735, | |
| "grad_norm": 0.1913049817085266, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0967, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.5630552671051113, | |
| "grad_norm": 0.11070458590984344, | |
| "learning_rate": 0.0002, | |
| "loss": 1.4031, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5722856813199493, | |
| "grad_norm": 0.05270432308316231, | |
| "learning_rate": 0.0002, | |
| "loss": 1.203, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.5815160955347871, | |
| "grad_norm": 0.12132929265499115, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1632, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.5907465097496251, | |
| "grad_norm": 0.06843800842761993, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3555, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.5999769239644629, | |
| "grad_norm": 0.1670321673154831, | |
| "learning_rate": 0.0002, | |
| "loss": 1.059, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.6092073381793008, | |
| "grad_norm": 0.10655465722084045, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3544, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.6184377523941387, | |
| "grad_norm": 0.05795924738049507, | |
| "learning_rate": 0.0002, | |
| "loss": 1.25, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.6276681666089766, | |
| "grad_norm": 0.1298709660768509, | |
| "learning_rate": 0.0002, | |
| "loss": 1.222, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.6368985808238145, | |
| "grad_norm": 0.0654703825712204, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3189, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.6461289950386524, | |
| "grad_norm": 0.18943524360656738, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0885, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.6553594092534902, | |
| "grad_norm": 0.11117199063301086, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3343, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.6645898234683282, | |
| "grad_norm": 0.05172109976410866, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1328, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.673820237683166, | |
| "grad_norm": 0.13359065353870392, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2447, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.683050651898004, | |
| "grad_norm": 0.059676457196474075, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2718, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.6922810661128418, | |
| "grad_norm": 0.1960563063621521, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0669, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.7015114803276797, | |
| "grad_norm": 0.11250808089971542, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3465, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.7107418945425176, | |
| "grad_norm": 0.05576665699481964, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1737, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.7199723087573555, | |
| "grad_norm": 0.16185137629508972, | |
| "learning_rate": 0.0002, | |
| "loss": 1.187, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.7292027229721934, | |
| "grad_norm": 0.06587795913219452, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3014, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.7384331371870313, | |
| "grad_norm": 0.21672724187374115, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1078, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.7476635514018691, | |
| "grad_norm": 0.09454522281885147, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3883, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.7568939656167071, | |
| "grad_norm": 0.05657172575592995, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2014, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.7661243798315449, | |
| "grad_norm": 0.12518398463726044, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1624, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.7753547940463829, | |
| "grad_norm": 0.0718185231089592, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2899, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.7845852082612207, | |
| "grad_norm": 0.1621280312538147, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0833, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.7938156224760586, | |
| "grad_norm": 0.1027815043926239, | |
| "learning_rate": 0.0002, | |
| "loss": 1.352, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.8030460366908965, | |
| "grad_norm": 0.056900862604379654, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2106, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.8122764509057344, | |
| "grad_norm": 0.138154998421669, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1345, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.8215068651205722, | |
| "grad_norm": 0.06931914389133453, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3355, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.8307372793354102, | |
| "grad_norm": 0.185394287109375, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0777, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.839967693550248, | |
| "grad_norm": 0.09037897735834122, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3078, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.849198107765086, | |
| "grad_norm": 0.05863342061638832, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2147, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.8584285219799238, | |
| "grad_norm": 0.14246703684329987, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1953, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.8676589361947618, | |
| "grad_norm": 0.07498980313539505, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2837, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.8768893504095996, | |
| "grad_norm": 0.23275145888328552, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0615, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.8861197646244375, | |
| "grad_norm": 0.09213528782129288, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3883, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.8953501788392754, | |
| "grad_norm": 0.05749181658029556, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2223, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.9045805930541133, | |
| "grad_norm": 0.12675738334655762, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1542, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.9138110072689511, | |
| "grad_norm": 0.06891526281833649, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2883, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.9230414214837891, | |
| "grad_norm": 0.1506439447402954, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1152, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.9322718356986269, | |
| "grad_norm": 0.08551948517560959, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3746, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.9415022499134649, | |
| "grad_norm": 0.05816769599914551, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2117, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.9507326641283027, | |
| "grad_norm": 0.15679936110973358, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2107, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.9599630783431407, | |
| "grad_norm": 0.06274525076150894, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2783, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.9691934925579785, | |
| "grad_norm": 0.1663607805967331, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0381, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.9784239067728164, | |
| "grad_norm": 0.10284286737442017, | |
| "learning_rate": 0.0002, | |
| "loss": 1.4281, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.9876543209876543, | |
| "grad_norm": 0.06172878295183182, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1913, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.9968847352024922, | |
| "grad_norm": 0.11540690809488297, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2357, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.0064612899503864, | |
| "grad_norm": 0.07405360043048859, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3859, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.0156917041652245, | |
| "grad_norm": 0.07363928109407425, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9267, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.0249221183800623, | |
| "grad_norm": 0.10718333721160889, | |
| "learning_rate": 0.0002, | |
| "loss": 1.242, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.0341525325949001, | |
| "grad_norm": 0.06949968636035919, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0087, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.043382946809738, | |
| "grad_norm": 0.15124961733818054, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0951, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.052613361024576, | |
| "grad_norm": 0.08505109697580338, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2338, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.0618437752394139, | |
| "grad_norm": 0.0844084769487381, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9387, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.0710741894542517, | |
| "grad_norm": 0.12017443031072617, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2559, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.0803046036690898, | |
| "grad_norm": 0.07678249478340149, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0786, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.0895350178839276, | |
| "grad_norm": 0.18587274849414825, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0891, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.0987654320987654, | |
| "grad_norm": 0.087877057492733, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2243, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.1079958463136033, | |
| "grad_norm": 0.08151056617498398, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9319, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.1172262605284413, | |
| "grad_norm": 0.13637030124664307, | |
| "learning_rate": 0.0002, | |
| "loss": 1.342, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.1264566747432792, | |
| "grad_norm": 0.07987112551927567, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0845, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.135687088958117, | |
| "grad_norm": 0.17300938069820404, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1104, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.1449175031729548, | |
| "grad_norm": 0.0821269229054451, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2084, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.1541479173877929, | |
| "grad_norm": 0.08363176882266998, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9193, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.1633783316026307, | |
| "grad_norm": 0.12400569021701813, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3398, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.1726087458174685, | |
| "grad_norm": 0.0741545781493187, | |
| "learning_rate": 0.0002, | |
| "loss": 1.07, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.1818391600323064, | |
| "grad_norm": 0.18392737209796906, | |
| "learning_rate": 0.0002, | |
| "loss": 1.102, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.1910695742471444, | |
| "grad_norm": 0.10395547747612, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1766, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.2002999884619823, | |
| "grad_norm": 0.08065596967935562, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9593, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.20953040267682, | |
| "grad_norm": 0.13076524436473846, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2912, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.218760816891658, | |
| "grad_norm": 0.08445240557193756, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0545, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.227991231106496, | |
| "grad_norm": 0.20568707585334778, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0472, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.2372216453213338, | |
| "grad_norm": 0.0978812500834465, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1746, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.2464520595361717, | |
| "grad_norm": 0.08013073354959488, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9361, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.2556824737510095, | |
| "grad_norm": 0.15785863995552063, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3058, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.2649128879658473, | |
| "grad_norm": 0.08266527205705643, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0807, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.2741433021806854, | |
| "grad_norm": 0.18238036334514618, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0587, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.2833737163955232, | |
| "grad_norm": 0.09454452991485596, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2299, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.292604130610361, | |
| "grad_norm": 0.09221120178699493, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9915, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.3018345448251991, | |
| "grad_norm": 0.13429689407348633, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3109, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.311064959040037, | |
| "grad_norm": 0.08239381015300751, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0279, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.3202953732548748, | |
| "grad_norm": 0.20243394374847412, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1165, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.3295257874697128, | |
| "grad_norm": 0.10623496025800705, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1892, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.3387562016845507, | |
| "grad_norm": 0.08493519574403763, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9719, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.3479866158993885, | |
| "grad_norm": 0.16793687641620636, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2933, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.3572170301142263, | |
| "grad_norm": 0.0803951844573021, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0914, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.3664474443290642, | |
| "grad_norm": 0.18061946332454681, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0951, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.3756778585439022, | |
| "grad_norm": 0.09481924027204514, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2137, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.38490827275874, | |
| "grad_norm": 0.08988731354475021, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0406, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.394138686973578, | |
| "grad_norm": 0.16403962671756744, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2664, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.403369101188416, | |
| "grad_norm": 0.0818110927939415, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0402, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.4125995154032538, | |
| "grad_norm": 0.16447734832763672, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1246, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.4218299296180916, | |
| "grad_norm": 0.10654182732105255, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1963, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.4310603438329295, | |
| "grad_norm": 0.09120084345340729, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0389, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.4402907580477673, | |
| "grad_norm": 0.14979740977287292, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2998, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.4495211722626054, | |
| "grad_norm": 0.07872021943330765, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0553, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.4587515864774432, | |
| "grad_norm": 0.1971125304698944, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0648, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.467982000692281, | |
| "grad_norm": 0.10692698508501053, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2246, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.477212414907119, | |
| "grad_norm": 0.0899726077914238, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8801, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.486442829121957, | |
| "grad_norm": 0.14624738693237305, | |
| "learning_rate": 0.0002, | |
| "loss": 1.268, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.4956732433367947, | |
| "grad_norm": 0.07690660655498505, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0909, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.5049036575516326, | |
| "grad_norm": 0.18284741044044495, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0576, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.5141340717664704, | |
| "grad_norm": 0.09398135542869568, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2297, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.5233644859813085, | |
| "grad_norm": 0.08855324983596802, | |
| "learning_rate": 0.0002, | |
| "loss": 0.97, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.5325949001961463, | |
| "grad_norm": 0.15404872596263885, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3044, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.5418253144109841, | |
| "grad_norm": 0.08170903474092484, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0842, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.5510557286258222, | |
| "grad_norm": 0.18026649951934814, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0973, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.56028614284066, | |
| "grad_norm": 0.09690876305103302, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2005, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.5695165570554979, | |
| "grad_norm": 0.09389860183000565, | |
| "learning_rate": 0.0002, | |
| "loss": 0.97, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.578746971270336, | |
| "grad_norm": 0.15237314999103546, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2745, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.5879773854851735, | |
| "grad_norm": 0.09236445277929306, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0959, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.5972077997000116, | |
| "grad_norm": 0.193682461977005, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1163, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.6064382139148494, | |
| "grad_norm": 0.1000017300248146, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2142, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.6156686281296873, | |
| "grad_norm": 0.0881427600979805, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9718, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.6248990423445253, | |
| "grad_norm": 0.16513171792030334, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2881, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.6341294565593631, | |
| "grad_norm": 0.0824236124753952, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0153, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.643359870774201, | |
| "grad_norm": 0.18334250152111053, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0999, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.652590284989039, | |
| "grad_norm": 0.09759881347417831, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2143, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.6618206992038767, | |
| "grad_norm": 0.09835653752088547, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9948, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.6710511134187147, | |
| "grad_norm": 0.1422744244337082, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2753, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.6802815276335525, | |
| "grad_norm": 0.0781414732336998, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0262, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.6895119418483904, | |
| "grad_norm": 0.2046748399734497, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0542, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.6987423560632284, | |
| "grad_norm": 0.09908697754144669, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1942, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.7079727702780663, | |
| "grad_norm": 0.09801312536001205, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9554, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.717203184492904, | |
| "grad_norm": 0.1688520461320877, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2497, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.7264335987077422, | |
| "grad_norm": 0.08527534455060959, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0928, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.7356640129225798, | |
| "grad_norm": 0.19082818925380707, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0729, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.7448944271374178, | |
| "grad_norm": 0.10551278293132782, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2043, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.7541248413522557, | |
| "grad_norm": 0.0871649906039238, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9317, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.7633552555670935, | |
| "grad_norm": 0.13540509343147278, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2648, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.7725856697819315, | |
| "grad_norm": 0.0853731706738472, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0521, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.7818160839967694, | |
| "grad_norm": 0.20368990302085876, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1177, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.7910464982116072, | |
| "grad_norm": 0.09358594566583633, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1872, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.8002769124264453, | |
| "grad_norm": 0.08881039917469025, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9926, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.8095073266412829, | |
| "grad_norm": 0.14714112877845764, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2216, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.818737740856121, | |
| "grad_norm": 0.08117840439081192, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0313, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.8279681550709588, | |
| "grad_norm": 0.2248132824897766, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1024, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.8371985692857966, | |
| "grad_norm": 0.11052978783845901, | |
| "learning_rate": 0.0002, | |
| "loss": 1.18, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.8464289835006347, | |
| "grad_norm": 0.08844051510095596, | |
| "learning_rate": 0.0002, | |
| "loss": 0.99, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.8556593977154725, | |
| "grad_norm": 0.17737261950969696, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3163, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.8648898119303103, | |
| "grad_norm": 0.09116645157337189, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0379, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.8741202261451484, | |
| "grad_norm": 0.16292434930801392, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0473, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.883350640359986, | |
| "grad_norm": 0.09686768054962158, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1816, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.892581054574824, | |
| "grad_norm": 0.09110133349895477, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9796, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.901811468789662, | |
| "grad_norm": 0.1265280693769455, | |
| "learning_rate": 0.0002, | |
| "loss": 1.3238, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.9110418830044997, | |
| "grad_norm": 0.07903092354536057, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0891, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.9202722972193378, | |
| "grad_norm": 0.17232394218444824, | |
| "learning_rate": 0.0002, | |
| "loss": 1.134, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.9295027114341756, | |
| "grad_norm": 0.10463748872280121, | |
| "learning_rate": 0.0002, | |
| "loss": 1.219, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.9387331256490135, | |
| "grad_norm": 0.09100574254989624, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9918, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.9479635398638515, | |
| "grad_norm": 0.16899889707565308, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2809, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 1.9571939540786891, | |
| "grad_norm": 0.08439228683710098, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0566, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.9664243682935272, | |
| "grad_norm": 0.1993444412946701, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0572, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.975654782508365, | |
| "grad_norm": 0.1062462106347084, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2122, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.9848851967232028, | |
| "grad_norm": 0.08482355624437332, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9347, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.994115610938041, | |
| "grad_norm": 0.13126371800899506, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2604, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 2.0036921656859352, | |
| "grad_norm": 0.09182780981063843, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1431, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 2.012922579900773, | |
| "grad_norm": 0.1420363485813141, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8912, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 2.022152994115611, | |
| "grad_norm": 0.1376965492963791, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1347, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 2.031383408330449, | |
| "grad_norm": 0.07587343454360962, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7146, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 2.0406138225452866, | |
| "grad_norm": 0.20123699307441711, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1783, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 2.0498442367601246, | |
| "grad_norm": 0.10732077807188034, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8584, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 2.0590746509749627, | |
| "grad_norm": 0.19853752851486206, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9121, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 2.0683050651898003, | |
| "grad_norm": 0.13605843484401703, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1323, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 2.0775354794046383, | |
| "grad_norm": 0.07920292764902115, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6811, | |
| "step": 2250 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 10000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 250, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.194849501524787e+17, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |