| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 31.446190102120973, | |
| "global_step": 15000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002094789211835559, | |
| "learning_rate": 1.0206207261596573e-07, | |
| "loss": 42.596397399902344, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.020947892118355592, | |
| "learning_rate": 1.0206207261596575e-06, | |
| "loss": 41.090047200520836, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.041895784236711184, | |
| "learning_rate": 2.041241452319315e-06, | |
| "loss": 40.72335205078125, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.06284367635506677, | |
| "learning_rate": 3.0618621784789722e-06, | |
| "loss": 40.80942077636719, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.08379156847342237, | |
| "learning_rate": 4.08248290463863e-06, | |
| "loss": 40.0015869140625, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.10473946059177795, | |
| "learning_rate": 5.103103630798286e-06, | |
| "loss": 39.082891845703124, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.12568735271013354, | |
| "learning_rate": 6.1237243569579445e-06, | |
| "loss": 39.320306396484376, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.14663524482848914, | |
| "learning_rate": 7.144345083117603e-06, | |
| "loss": 38.66647033691406, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.16758313694684474, | |
| "learning_rate": 8.16496580927726e-06, | |
| "loss": 38.0055908203125, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.1885310290652003, | |
| "learning_rate": 9.185586535436916e-06, | |
| "loss": 37.521505737304686, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2094789211835559, | |
| "learning_rate": 1.0206207261596573e-05, | |
| "loss": 35.75002746582031, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2304268133019115, | |
| "learning_rate": 1.1226827987756233e-05, | |
| "loss": 35.38407592773437, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2513747054202671, | |
| "learning_rate": 1.2247448713915889e-05, | |
| "loss": 34.455230712890625, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.2723225975386227, | |
| "learning_rate": 1.3268069440075545e-05, | |
| "loss": 33.48695373535156, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2932704896569783, | |
| "learning_rate": 1.4288690166235205e-05, | |
| "loss": 33.0624755859375, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.31421838177533384, | |
| "learning_rate": 1.530931089239486e-05, | |
| "loss": 31.633297729492188, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.33516627389368947, | |
| "learning_rate": 1.632993161855452e-05, | |
| "loss": 30.392620849609376, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.35611416601204504, | |
| "learning_rate": 1.7350552344714174e-05, | |
| "loss": 30.184588623046874, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.3770620581304006, | |
| "learning_rate": 1.8371173070873833e-05, | |
| "loss": 29.326535034179688, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.39800995024875624, | |
| "learning_rate": 1.939179379703349e-05, | |
| "loss": 28.23697509765625, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4189578423671118, | |
| "learning_rate": 2.0412414523193145e-05, | |
| "loss": 27.301419067382813, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4399057344854674, | |
| "learning_rate": 2.1433035249352804e-05, | |
| "loss": 26.5281494140625, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.460853626603823, | |
| "learning_rate": 2.2453655975512465e-05, | |
| "loss": 25.992547607421876, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.4818015187221786, | |
| "learning_rate": 2.347427670167212e-05, | |
| "loss": 24.90663299560547, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5027494108405341, | |
| "learning_rate": 2.4494897427831778e-05, | |
| "loss": 23.896534729003907, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5236973029588897, | |
| "learning_rate": 2.5515518153991436e-05, | |
| "loss": 22.488262939453126, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5446451950772454, | |
| "learning_rate": 2.653613888015109e-05, | |
| "loss": 21.760206604003905, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.565593087195601, | |
| "learning_rate": 2.755675960631075e-05, | |
| "loss": 20.808561706542967, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.5865409793139565, | |
| "learning_rate": 2.857738033247041e-05, | |
| "loss": 20.107774353027345, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6074888714323121, | |
| "learning_rate": 2.9598001058630065e-05, | |
| "loss": 19.469386291503906, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.6284367635506677, | |
| "learning_rate": 3.061862178478972e-05, | |
| "loss": 18.442234802246094, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6493846556690233, | |
| "learning_rate": 3.163924251094938e-05, | |
| "loss": 18.132992553710938, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.6703325477873789, | |
| "learning_rate": 3.265986323710904e-05, | |
| "loss": 17.425698852539064, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.6912804399057345, | |
| "learning_rate": 3.3680483963268694e-05, | |
| "loss": 16.683474731445312, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7122283320240901, | |
| "learning_rate": 3.470110468942835e-05, | |
| "loss": 16.511445617675783, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7331762241424457, | |
| "learning_rate": 3.5721725415588004e-05, | |
| "loss": 15.8478759765625, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.7541241162608012, | |
| "learning_rate": 3.6742346141747665e-05, | |
| "loss": 15.450515747070312, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.7750720083791568, | |
| "learning_rate": 3.7762966867907327e-05, | |
| "loss": 14.916090393066407, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.7960199004975125, | |
| "learning_rate": 3.878358759406698e-05, | |
| "loss": 14.495413208007813, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.816967792615868, | |
| "learning_rate": 3.980420832022664e-05, | |
| "loss": 13.862504577636718, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.8379156847342236, | |
| "learning_rate": 4.082482904638629e-05, | |
| "loss": 13.526719665527343, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8588635768525792, | |
| "learning_rate": 4.184544977254595e-05, | |
| "loss": 13.241981506347656, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.8798114689709348, | |
| "learning_rate": 4.286607049870561e-05, | |
| "loss": 12.791949462890624, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.9007593610892904, | |
| "learning_rate": 4.388669122486527e-05, | |
| "loss": 12.353260040283203, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.921707253207646, | |
| "learning_rate": 4.490731195102493e-05, | |
| "loss": 11.939605712890625, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.9426551453260016, | |
| "learning_rate": 4.5927932677184585e-05, | |
| "loss": 11.615445709228515, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.9636030374443572, | |
| "learning_rate": 4.694855340334424e-05, | |
| "loss": 11.29063720703125, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.9845509295627127, | |
| "learning_rate": 4.7969174129503894e-05, | |
| "loss": 11.051673889160156, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.0062843676355067, | |
| "learning_rate": 4.8989794855663556e-05, | |
| "loss": 11.174005889892578, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.0272322597538623, | |
| "learning_rate": 5.001041558182322e-05, | |
| "loss": 10.351375579833984, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.0481801518722178, | |
| "learning_rate": 5.103103630798287e-05, | |
| "loss": 10.082479858398438, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.0691280439905735, | |
| "learning_rate": 5.205165703414253e-05, | |
| "loss": 9.848101806640624, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.090075936108929, | |
| "learning_rate": 5.307227776030218e-05, | |
| "loss": 9.558543395996093, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.1110238282272846, | |
| "learning_rate": 5.409289848646184e-05, | |
| "loss": 9.263871765136718, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.1319717203456403, | |
| "learning_rate": 5.51135192126215e-05, | |
| "loss": 9.101696014404297, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.1529196124639958, | |
| "learning_rate": 5.613413993878116e-05, | |
| "loss": 8.784052276611328, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.1738675045823515, | |
| "learning_rate": 5.715476066494082e-05, | |
| "loss": 8.466715240478516, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.194815396700707, | |
| "learning_rate": 5.817538139110047e-05, | |
| "loss": 8.419536590576172, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.2157632888190626, | |
| "learning_rate": 5.919600211726013e-05, | |
| "loss": 8.445430755615234, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.236711180937418, | |
| "learning_rate": 6.0216622843419785e-05, | |
| "loss": 7.745582580566406, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.2576590730557737, | |
| "learning_rate": 6.123724356957945e-05, | |
| "loss": 7.821333312988282, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.2786069651741294, | |
| "learning_rate": 6.22578642957391e-05, | |
| "loss": 7.500454711914062, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.2995548572924849, | |
| "learning_rate": 6.327848502189876e-05, | |
| "loss": 7.44578857421875, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.3205027494108406, | |
| "learning_rate": 6.429910574805841e-05, | |
| "loss": 7.184627532958984, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.341450641529196, | |
| "learning_rate": 6.531972647421808e-05, | |
| "loss": 6.880846405029297, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.3623985336475517, | |
| "learning_rate": 6.634034720037773e-05, | |
| "loss": 6.639464569091797, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.3833464257659074, | |
| "learning_rate": 6.736096792653739e-05, | |
| "loss": 6.543840789794922, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.4042943178842628, | |
| "learning_rate": 6.838158865269704e-05, | |
| "loss": 6.245196914672851, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.4252422100026185, | |
| "learning_rate": 6.94022093788567e-05, | |
| "loss": 6.316292572021484, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.446190102120974, | |
| "learning_rate": 7.042283010501637e-05, | |
| "loss": 6.2703697204589846, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.4671379942393297, | |
| "learning_rate": 7.144345083117601e-05, | |
| "loss": 6.09345703125, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.4880858863576854, | |
| "learning_rate": 7.246407155733568e-05, | |
| "loss": 5.9814506530761715, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.5090337784760408, | |
| "learning_rate": 7.348469228349533e-05, | |
| "loss": 5.9337646484375, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.5299816705943965, | |
| "learning_rate": 7.450531300965498e-05, | |
| "loss": 5.937409591674805, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.550929562712752, | |
| "learning_rate": 7.552593373581465e-05, | |
| "loss": 5.64327392578125, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.5718774548311076, | |
| "learning_rate": 7.654655446197431e-05, | |
| "loss": 5.371760559082031, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.5928253469494633, | |
| "learning_rate": 7.756717518813396e-05, | |
| "loss": 5.309605407714844, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.6137732390678188, | |
| "learning_rate": 7.858779591429362e-05, | |
| "loss": 5.3783222198486325, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.6347211311861742, | |
| "learning_rate": 7.960841664045329e-05, | |
| "loss": 5.400894546508789, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.65566902330453, | |
| "learning_rate": 8.062903736661294e-05, | |
| "loss": 5.137008285522461, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.6766169154228856, | |
| "learning_rate": 8.164965809277258e-05, | |
| "loss": 5.365228271484375, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.6975648075412413, | |
| "learning_rate": 8.267027881893225e-05, | |
| "loss": 5.210577392578125, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.7185126996595967, | |
| "learning_rate": 8.36908995450919e-05, | |
| "loss": 5.149754333496094, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.7394605917779522, | |
| "learning_rate": 8.471152027125156e-05, | |
| "loss": 5.063209915161133, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.7604084838963079, | |
| "learning_rate": 8.573214099741121e-05, | |
| "loss": 4.89969482421875, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.7813563760146636, | |
| "learning_rate": 8.675276172357088e-05, | |
| "loss": 4.8380378723144535, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.8023042681330192, | |
| "learning_rate": 8.777338244973054e-05, | |
| "loss": 4.826490783691407, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.8232521602513747, | |
| "learning_rate": 8.879400317589019e-05, | |
| "loss": 4.733642578125, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.8442000523697302, | |
| "learning_rate": 8.981462390204986e-05, | |
| "loss": 4.840193557739258, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.8651479444880859, | |
| "learning_rate": 9.083524462820951e-05, | |
| "loss": 4.841461944580078, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.8860958366064415, | |
| "learning_rate": 9.185586535436917e-05, | |
| "loss": 4.591343688964844, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.9070437287247972, | |
| "learning_rate": 9.287648608052881e-05, | |
| "loss": 4.796835708618164, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.9279916208431527, | |
| "learning_rate": 9.389710680668848e-05, | |
| "loss": 4.928312301635742, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.9489395129615081, | |
| "learning_rate": 9.491772753284813e-05, | |
| "loss": 4.6936603546142575, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.9698874050798638, | |
| "learning_rate": 9.593834825900779e-05, | |
| "loss": 4.678832626342773, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.9908352971982195, | |
| "learning_rate": 9.695896898516746e-05, | |
| "loss": 4.801474380493164, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 2.0125687352710133, | |
| "learning_rate": 9.797958971132711e-05, | |
| "loss": 4.615579986572266, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.033516627389369, | |
| "learning_rate": 9.900021043748677e-05, | |
| "loss": 4.685293197631836, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.0544645195077247, | |
| "learning_rate": 0.00010002083116364643, | |
| "loss": 4.529672622680664, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.07541241162608, | |
| "learning_rate": 0.00010104145188980609, | |
| "loss": 4.6519828796386715, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.0963603037444356, | |
| "learning_rate": 0.00010206207261596574, | |
| "loss": 4.200122451782226, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.1173081958627913, | |
| "learning_rate": 0.0001030826933421254, | |
| "loss": 4.467970275878907, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 2.138256087981147, | |
| "learning_rate": 0.00010410331406828505, | |
| "loss": 4.364266586303711, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 2.1592039800995027, | |
| "learning_rate": 0.00010512393479444471, | |
| "loss": 4.321992874145508, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 2.180151872217858, | |
| "learning_rate": 0.00010614455552060436, | |
| "loss": 4.208817672729492, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 2.2010997643362136, | |
| "learning_rate": 0.00010716517624676403, | |
| "loss": 4.414374923706054, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.2220476564545693, | |
| "learning_rate": 0.00010818579697292369, | |
| "loss": 4.222100067138672, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 2.242995548572925, | |
| "learning_rate": 0.00010920641769908334, | |
| "loss": 4.392937850952149, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 2.2639434406912806, | |
| "learning_rate": 0.000110227038425243, | |
| "loss": 4.361217498779297, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 2.284891332809636, | |
| "learning_rate": 0.00011124765915140266, | |
| "loss": 4.305131912231445, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 2.3058392249279915, | |
| "learning_rate": 0.00011226827987756232, | |
| "loss": 4.322722244262695, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.326787117046347, | |
| "learning_rate": 0.00011328890060372197, | |
| "loss": 4.250308990478516, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 2.347735009164703, | |
| "learning_rate": 0.00011430952132988164, | |
| "loss": 4.366016006469726, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 2.3686829012830586, | |
| "learning_rate": 0.00011533014205604128, | |
| "loss": 4.271330642700195, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 2.389630793401414, | |
| "learning_rate": 0.00011635076278220094, | |
| "loss": 4.35943489074707, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.4105786855197695, | |
| "learning_rate": 0.00011737138350836059, | |
| "loss": 4.14013442993164, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.431526577638125, | |
| "learning_rate": 0.00011839200423452026, | |
| "loss": 4.263423156738281, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.452474469756481, | |
| "learning_rate": 0.00011941262496067991, | |
| "loss": 4.046255874633789, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.473422361874836, | |
| "learning_rate": 0.00012043324568683957, | |
| "loss": 4.267144775390625, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.494370253993192, | |
| "learning_rate": 0.00012145386641299924, | |
| "loss": 4.446724319458008, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 2.5153181461115475, | |
| "learning_rate": 0.0001224744871391589, | |
| "loss": 4.310148239135742, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.536266038229903, | |
| "learning_rate": 0.00012349510786531856, | |
| "loss": 4.145759963989258, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 2.557213930348259, | |
| "learning_rate": 0.0001245157285914782, | |
| "loss": 3.9344154357910157, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.5781618224666145, | |
| "learning_rate": 0.00012553634931763784, | |
| "loss": 4.1616455078125, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 2.5991097145849698, | |
| "learning_rate": 0.0001265569700437975, | |
| "loss": 4.117146682739258, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.6200576067033254, | |
| "learning_rate": 0.00012757759076995718, | |
| "loss": 4.102180480957031, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.641005498821681, | |
| "learning_rate": 0.00012859821149611682, | |
| "loss": 4.15636100769043, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.661953390940037, | |
| "learning_rate": 0.0001296188322222765, | |
| "loss": 4.069457626342773, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.682901283058392, | |
| "learning_rate": 0.00013063945294843616, | |
| "loss": 4.0670215606689455, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.7038491751767477, | |
| "learning_rate": 0.0001316600736745958, | |
| "loss": 3.985906219482422, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.7247970672951034, | |
| "learning_rate": 0.00013268069440075547, | |
| "loss": 4.272599792480468, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.745744959413459, | |
| "learning_rate": 0.00013370131512691514, | |
| "loss": 3.9709007263183596, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.7666928515318148, | |
| "learning_rate": 0.00013472193585307478, | |
| "loss": 4.1308135986328125, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.7876407436501704, | |
| "learning_rate": 0.00013574255657923444, | |
| "loss": 4.175233840942383, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.8085886357685257, | |
| "learning_rate": 0.00013676317730539409, | |
| "loss": 3.9187103271484376, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.8295365278868814, | |
| "learning_rate": 0.00013778379803155375, | |
| "loss": 3.8264163970947265, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.850484420005237, | |
| "learning_rate": 0.0001388044187577134, | |
| "loss": 4.124664306640625, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.8714323121235923, | |
| "learning_rate": 0.00013982503948387306, | |
| "loss": 3.952465057373047, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.892380204241948, | |
| "learning_rate": 0.00014084566021003273, | |
| "loss": 3.987406921386719, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.9133280963603037, | |
| "learning_rate": 0.00014186628093619237, | |
| "loss": 3.92406005859375, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.9342759884786593, | |
| "learning_rate": 0.00014288690166235201, | |
| "loss": 3.932135009765625, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.955223880597015, | |
| "learning_rate": 0.00014390752238851168, | |
| "loss": 4.134164428710937, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.9761717727153707, | |
| "learning_rate": 0.00014492814311467135, | |
| "loss": 4.045958709716797, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.9971196648337264, | |
| "learning_rate": 0.000145948763840831, | |
| "loss": 3.9462562561035157, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 3.01885310290652, | |
| "learning_rate": 0.00014696938456699066, | |
| "loss": 4.254601669311524, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 3.0398009950248754, | |
| "learning_rate": 0.00014799000529315033, | |
| "loss": 3.9297733306884766, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 3.060748887143231, | |
| "learning_rate": 0.00014901062601930997, | |
| "loss": 3.8294136047363283, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 3.081696779261587, | |
| "learning_rate": 0.00015003124674546964, | |
| "loss": 4.05113525390625, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 3.1026446713799425, | |
| "learning_rate": 0.0001510518674716293, | |
| "loss": 3.975564956665039, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 3.123592563498298, | |
| "learning_rate": 0.00015207248819778895, | |
| "loss": 3.9852630615234377, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 3.1445404556166534, | |
| "learning_rate": 0.00015309310892394862, | |
| "loss": 4.0287940979003904, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 3.165488347735009, | |
| "learning_rate": 0.00015411372965010828, | |
| "loss": 3.9631397247314455, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 3.1864362398533648, | |
| "learning_rate": 0.00015513435037626793, | |
| "loss": 3.9254386901855467, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 3.2073841319717205, | |
| "learning_rate": 0.0001561549711024276, | |
| "loss": 3.82324104309082, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 3.228332024090076, | |
| "learning_rate": 0.00015717559182858723, | |
| "loss": 3.9215030670166016, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 3.2492799162084314, | |
| "learning_rate": 0.0001581962125547469, | |
| "loss": 3.8164131164550783, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 3.270227808326787, | |
| "learning_rate": 0.00015921683328090657, | |
| "loss": 3.7764801025390624, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 3.2911757004451427, | |
| "learning_rate": 0.0001602374540070662, | |
| "loss": 3.9773059844970704, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 3.3121235925634984, | |
| "learning_rate": 0.00016125807473322588, | |
| "loss": 3.8427078247070314, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 3.333071484681854, | |
| "learning_rate": 0.00016227869545938555, | |
| "loss": 3.8103118896484376, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 3.3540193768002093, | |
| "learning_rate": 0.00016329931618554516, | |
| "loss": 3.8182399749755858, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 3.374967268918565, | |
| "learning_rate": 0.00016431993691170483, | |
| "loss": 3.750722122192383, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 3.3959151610369207, | |
| "learning_rate": 0.0001653405576378645, | |
| "loss": 3.9241371154785156, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 3.4168630531552764, | |
| "learning_rate": 0.00016636117836402414, | |
| "loss": 3.827128219604492, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 3.4378109452736316, | |
| "learning_rate": 0.0001673817990901838, | |
| "loss": 3.750970458984375, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 3.4587588373919873, | |
| "learning_rate": 0.00016840241981634345, | |
| "loss": 3.827236557006836, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 3.479706729510343, | |
| "learning_rate": 0.00016942304054250312, | |
| "loss": 3.862264633178711, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 3.5006546216286987, | |
| "learning_rate": 0.0001704436612686628, | |
| "loss": 3.9794548034667967, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 3.5216025137470544, | |
| "learning_rate": 0.00017146428199482243, | |
| "loss": 3.8642444610595703, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 3.54255040586541, | |
| "learning_rate": 0.0001724849027209821, | |
| "loss": 3.8710708618164062, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 3.5634982979837653, | |
| "learning_rate": 0.00017350552344714176, | |
| "loss": 3.7993534088134764, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 3.584446190102121, | |
| "learning_rate": 0.0001745261441733014, | |
| "loss": 3.8951980590820314, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 3.6053940822204766, | |
| "learning_rate": 0.00017554676489946107, | |
| "loss": 3.728221893310547, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 3.6263419743388323, | |
| "learning_rate": 0.00017656738562562074, | |
| "loss": 3.7405670166015623, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 3.6472898664571876, | |
| "learning_rate": 0.00017758800635178038, | |
| "loss": 3.7010948181152346, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 3.6682377585755432, | |
| "learning_rate": 0.00017860862707794005, | |
| "loss": 3.7746726989746096, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 3.689185650693899, | |
| "learning_rate": 0.00017962924780409972, | |
| "loss": 3.842145538330078, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 3.7101335428122546, | |
| "learning_rate": 0.00018064986853025936, | |
| "loss": 3.817458724975586, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 3.7310814349306103, | |
| "learning_rate": 0.00018167048925641903, | |
| "loss": 3.7765247344970705, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 3.752029327048966, | |
| "learning_rate": 0.0001826911099825787, | |
| "loss": 3.6870758056640627, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 3.772977219167321, | |
| "learning_rate": 0.00018371173070873834, | |
| "loss": 3.8823310852050783, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 3.793925111285677, | |
| "learning_rate": 0.000184732351434898, | |
| "loss": 3.7599208831787108, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 3.8148730034040326, | |
| "learning_rate": 0.00018575297216105762, | |
| "loss": 3.8205623626708984, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 3.835820895522388, | |
| "learning_rate": 0.0001867735928872173, | |
| "loss": 3.840204620361328, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 3.8567687876407435, | |
| "learning_rate": 0.00018779421361337696, | |
| "loss": 3.8769672393798826, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 3.877716679759099, | |
| "learning_rate": 0.0001888148343395366, | |
| "loss": 3.7292160034179687, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 3.898664571877455, | |
| "learning_rate": 0.00018983545506569627, | |
| "loss": 3.785298156738281, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 3.9196124639958105, | |
| "learning_rate": 0.00019085607579185594, | |
| "loss": 3.691172790527344, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 3.940560356114166, | |
| "learning_rate": 0.00019187669651801558, | |
| "loss": 3.7312793731689453, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 3.9615082482325215, | |
| "learning_rate": 0.00019289731724417525, | |
| "loss": 3.7419872283935547, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 3.982456140350877, | |
| "learning_rate": 0.0001939179379703349, | |
| "loss": 3.587678909301758, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 4.004189578423671, | |
| "learning_rate": 0.00019493855869649455, | |
| "loss": 3.9609317779541016, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 4.025137470542027, | |
| "learning_rate": 0.00019595917942265422, | |
| "loss": 3.7329071044921873, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 4.046085362660382, | |
| "learning_rate": 0.0001969798001488139, | |
| "loss": 3.7702545166015624, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 4.067033254778738, | |
| "learning_rate": 0.00019800042087497353, | |
| "loss": 3.796523666381836, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 4.087981146897094, | |
| "learning_rate": 0.0001990210416011332, | |
| "loss": 3.643301773071289, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 4.108929039015449, | |
| "learning_rate": 0.00020004166232729287, | |
| "loss": 3.705374526977539, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 4.129876931133805, | |
| "learning_rate": 0.0002010622830534525, | |
| "loss": 3.619226837158203, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 4.15082482325216, | |
| "learning_rate": 0.00020208290377961218, | |
| "loss": 3.862563705444336, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 4.1717727153705155, | |
| "learning_rate": 0.00020310352450577185, | |
| "loss": 3.6324195861816406, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 4.192720607488871, | |
| "learning_rate": 0.0002041241452319315, | |
| "loss": 3.674951171875, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 4.213668499607227, | |
| "learning_rate": 0.00020514476595809116, | |
| "loss": 3.59210205078125, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 4.234616391725583, | |
| "learning_rate": 0.0002061653866842508, | |
| "loss": 3.8746570587158202, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 4.255564283843938, | |
| "learning_rate": 0.00020718600741041044, | |
| "loss": 3.682146453857422, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 4.276512175962294, | |
| "learning_rate": 0.0002082066281365701, | |
| "loss": 3.8217212677001955, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 4.29746006808065, | |
| "learning_rate": 0.00020922724886272975, | |
| "loss": 3.691872787475586, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 4.318407960199005, | |
| "learning_rate": 0.00021024786958888942, | |
| "loss": 3.7160354614257813, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 4.339355852317361, | |
| "learning_rate": 0.00021126849031504908, | |
| "loss": 3.5512325286865236, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 4.360303744435716, | |
| "learning_rate": 0.00021228911104120873, | |
| "loss": 3.623905563354492, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 4.3812516365540715, | |
| "learning_rate": 0.0002133097317673684, | |
| "loss": 3.756671905517578, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 4.402199528672427, | |
| "learning_rate": 0.00021433035249352806, | |
| "loss": 3.71322021484375, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 4.423147420790783, | |
| "learning_rate": 0.0002153509732196877, | |
| "loss": 3.588302993774414, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 4.4440953129091385, | |
| "learning_rate": 0.00021637159394584737, | |
| "loss": 3.666096496582031, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 4.465043205027494, | |
| "learning_rate": 0.00021739221467200704, | |
| "loss": 3.660139465332031, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 4.48599109714585, | |
| "learning_rate": 0.00021841283539816668, | |
| "loss": 3.5749874114990234, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 4.5069389892642056, | |
| "learning_rate": 0.00021943345612432635, | |
| "loss": 3.621977615356445, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 4.527886881382561, | |
| "learning_rate": 0.000220454076850486, | |
| "loss": 3.7179306030273436, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 4.548834773500916, | |
| "learning_rate": 0.00022147469757664566, | |
| "loss": 3.5558433532714844, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 4.569782665619272, | |
| "learning_rate": 0.00022249531830280533, | |
| "loss": 3.754520034790039, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 4.590730557737627, | |
| "learning_rate": 0.00022351593902896497, | |
| "loss": 3.734426498413086, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 4.611678449855983, | |
| "learning_rate": 0.00022453655975512464, | |
| "loss": 3.6508132934570314, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 4.632626341974339, | |
| "learning_rate": 0.0002255571804812843, | |
| "loss": 3.7082672119140625, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 4.653574234092694, | |
| "learning_rate": 0.00022657780120744395, | |
| "loss": 3.600681686401367, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 4.67452212621105, | |
| "learning_rate": 0.00022759842193360361, | |
| "loss": 3.6909461975097657, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 4.695470018329406, | |
| "learning_rate": 0.00022861904265976328, | |
| "loss": 3.557560348510742, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 4.7164179104477615, | |
| "learning_rate": 0.0002296396633859229, | |
| "loss": 3.6415851593017576, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 4.737365802566117, | |
| "learning_rate": 0.00023066028411208256, | |
| "loss": 3.6133026123046874, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 4.758313694684473, | |
| "learning_rate": 0.00023168090483824223, | |
| "loss": 3.5348537445068358, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 4.779261586802828, | |
| "learning_rate": 0.00023270152556440187, | |
| "loss": 3.6589839935302733, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 4.800209478921183, | |
| "learning_rate": 0.00023372214629056154, | |
| "loss": 3.5354270935058594, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 4.821157371039539, | |
| "learning_rate": 0.00023474276701672118, | |
| "loss": 3.536578369140625, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 4.842105263157895, | |
| "learning_rate": 0.00023576338774288085, | |
| "loss": 3.5566326141357423, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 4.86305315527625, | |
| "learning_rate": 0.00023678400846904052, | |
| "loss": 3.475338363647461, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 4.884001047394606, | |
| "learning_rate": 0.00023780462919520016, | |
| "loss": 3.6042369842529296, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 4.904948939512962, | |
| "learning_rate": 0.00023882524992135983, | |
| "loss": 3.621173095703125, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 4.925896831631317, | |
| "learning_rate": 0.0002398458706475195, | |
| "loss": 3.640410232543945, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 4.946844723749672, | |
| "learning_rate": 0.00024086649137367914, | |
| "loss": 3.488922882080078, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 4.967792615868028, | |
| "learning_rate": 0.0002418871120998388, | |
| "loss": 3.4922332763671875, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 4.988740507986384, | |
| "learning_rate": 0.00024290773282599848, | |
| "loss": 3.590007019042969, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 5.010473946059178, | |
| "learning_rate": 0.00024392835355215812, | |
| "loss": 3.6829368591308596, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 5.0314218381775335, | |
| "learning_rate": 0.0002449489742783178, | |
| "loss": 3.651840591430664, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 5.052369730295889, | |
| "learning_rate": 0.00024596959500447745, | |
| "loss": 3.7254043579101563, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 5.073317622414245, | |
| "learning_rate": 0.0002469902157306371, | |
| "loss": 3.594907760620117, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 5.094265514532601, | |
| "learning_rate": 0.00024801083645679674, | |
| "loss": 3.4854148864746093, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 5.115213406650955, | |
| "learning_rate": 0.0002490314571829564, | |
| "loss": 3.58482666015625, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 5.136161298769311, | |
| "learning_rate": 0.00025005207790911607, | |
| "loss": 3.5847278594970704, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 5.157109190887667, | |
| "learning_rate": 0.0002510726986352757, | |
| "loss": 3.5045509338378906, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 5.178057083006022, | |
| "learning_rate": 0.00025209331936143535, | |
| "loss": 3.6352733612060546, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 5.199004975124378, | |
| "learning_rate": 0.000253113940087595, | |
| "loss": 3.555766296386719, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 5.219952867242734, | |
| "learning_rate": 0.0002541345608137547, | |
| "loss": 3.541688919067383, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 5.2409007593610895, | |
| "learning_rate": 0.00025515518153991436, | |
| "loss": 3.5400638580322266, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 5.261848651479445, | |
| "learning_rate": 0.00025617580226607403, | |
| "loss": 3.4929561614990234, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 5.282796543597801, | |
| "learning_rate": 0.00025719642299223364, | |
| "loss": 3.5611968994140626, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 5.3037444357161565, | |
| "learning_rate": 0.0002582170437183933, | |
| "loss": 3.4528472900390623, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 5.324692327834511, | |
| "learning_rate": 0.000259237664444553, | |
| "loss": 3.474958801269531, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 5.345640219952867, | |
| "learning_rate": 0.00026025828517071265, | |
| "loss": 3.6800113677978517, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 5.366588112071223, | |
| "learning_rate": 0.0002612789058968723, | |
| "loss": 3.524998092651367, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 5.387536004189578, | |
| "learning_rate": 0.00026229952662303193, | |
| "loss": 3.3865074157714843, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 5.408483896307934, | |
| "learning_rate": 0.0002633201473491916, | |
| "loss": 3.416782760620117, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 5.42943178842629, | |
| "learning_rate": 0.00026434076807535127, | |
| "loss": 3.4291786193847655, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 5.450379680544645, | |
| "learning_rate": 0.00026536138880151093, | |
| "loss": 3.487574005126953, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 5.471327572663001, | |
| "learning_rate": 0.0002663820095276706, | |
| "loss": 3.5091732025146483, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 5.492275464781357, | |
| "learning_rate": 0.00026740263025383027, | |
| "loss": 4.268975448608399, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 5.5132233568997115, | |
| "learning_rate": 0.0002684232509799899, | |
| "loss": 3.732823944091797, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 5.534171249018067, | |
| "learning_rate": 0.00026944387170614955, | |
| "loss": 3.650152587890625, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 5.555119141136423, | |
| "learning_rate": 0.0002704644924323092, | |
| "loss": 3.6992671966552733, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 5.576067033254779, | |
| "learning_rate": 0.0002714851131584689, | |
| "loss": 3.5627864837646483, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 5.597014925373134, | |
| "learning_rate": 0.0002725057338846285, | |
| "loss": 3.5473575592041016, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 5.61796281749149, | |
| "learning_rate": 0.00027352635461078817, | |
| "loss": 3.7565258026123045, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 5.638910709609846, | |
| "learning_rate": 0.00027454697533694784, | |
| "loss": 3.4811996459960937, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 5.659858601728201, | |
| "learning_rate": 0.0002755675960631075, | |
| "loss": 3.5412361145019533, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 5.680806493846557, | |
| "learning_rate": 0.0002765882167892671, | |
| "loss": 3.6894275665283205, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 5.701754385964913, | |
| "learning_rate": 0.0002776088375154268, | |
| "loss": 3.532870864868164, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 5.722702278083268, | |
| "learning_rate": 0.00027862945824158646, | |
| "loss": 3.4718368530273436, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 5.743650170201623, | |
| "learning_rate": 0.00027965007896774613, | |
| "loss": 3.5320533752441405, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 5.764598062319979, | |
| "learning_rate": 0.0002806706996939058, | |
| "loss": 3.6331645965576174, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 5.7855459544383345, | |
| "learning_rate": 0.00028169132042006546, | |
| "loss": 3.50958137512207, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 5.80649384655669, | |
| "learning_rate": 0.00028271194114622513, | |
| "loss": 3.5480377197265627, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 5.827441738675046, | |
| "learning_rate": 0.00028373256187238475, | |
| "loss": 3.5677505493164063, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 5.848389630793402, | |
| "learning_rate": 0.0002847531825985444, | |
| "loss": 3.5496990203857424, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 5.869337522911757, | |
| "learning_rate": 0.00028577380332470403, | |
| "loss": 3.4990489959716795, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 5.890285415030113, | |
| "learning_rate": 0.0002867944240508637, | |
| "loss": 3.4764991760253907, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 5.911233307148468, | |
| "learning_rate": 0.00028781504477702337, | |
| "loss": 3.4782173156738283, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 5.932181199266823, | |
| "learning_rate": 0.00028883566550318303, | |
| "loss": 3.5007530212402345, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 5.953129091385179, | |
| "learning_rate": 0.0002898562862293427, | |
| "loss": 3.4596179962158202, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 5.974076983503535, | |
| "learning_rate": 0.0002908769069555023, | |
| "loss": 3.4080764770507814, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 5.9950248756218905, | |
| "learning_rate": 0.000291897527681662, | |
| "loss": 3.5865558624267577, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 6.016758313694685, | |
| "learning_rate": 0.00029291814840782165, | |
| "loss": 3.539356231689453, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 6.03770620581304, | |
| "learning_rate": 0.0002939387691339813, | |
| "loss": 3.5746910095214846, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 6.058654097931396, | |
| "learning_rate": 0.000294959389860141, | |
| "loss": 3.477669906616211, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 6.079601990049751, | |
| "learning_rate": 0.00029598001058630066, | |
| "loss": 3.385912322998047, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 6.100549882168107, | |
| "learning_rate": 0.00029700063131246027, | |
| "loss": 3.566743850708008, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 6.121497774286462, | |
| "learning_rate": 0.00029802125203861994, | |
| "loss": 3.4617984771728514, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 6.142445666404818, | |
| "learning_rate": 0.0002990418727647796, | |
| "loss": 3.46124382019043, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 6.163393558523174, | |
| "learning_rate": 0.0003000624934909393, | |
| "loss": 3.5851741790771485, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 6.184341450641529, | |
| "learning_rate": 0.00030108311421709894, | |
| "loss": 3.5333206176757814, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 6.205289342759885, | |
| "learning_rate": 0.0003021037349432586, | |
| "loss": 3.4518871307373047, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 6.226237234878241, | |
| "learning_rate": 0.0003031243556694182, | |
| "loss": 3.5048519134521485, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 6.247185126996596, | |
| "learning_rate": 0.0003041449763955779, | |
| "loss": 3.386810302734375, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 6.268133019114952, | |
| "learning_rate": 0.00030516559712173756, | |
| "loss": 3.3472484588623046, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 6.289080911233307, | |
| "learning_rate": 0.00030618621784789723, | |
| "loss": 3.578289794921875, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 6.3100288033516625, | |
| "learning_rate": 0.0003072068385740569, | |
| "loss": 3.4641948699951173, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 6.330976695470018, | |
| "learning_rate": 0.00030822745930021657, | |
| "loss": 3.42608757019043, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 6.351924587588374, | |
| "learning_rate": 0.0003092480800263762, | |
| "loss": 3.4154186248779297, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 6.3728724797067295, | |
| "learning_rate": 0.00031026870075253585, | |
| "loss": 3.5270923614501952, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 6.393820371825085, | |
| "learning_rate": 0.0003112893214786955, | |
| "loss": 3.4780391693115233, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 6.414768263943441, | |
| "learning_rate": 0.0003123099422048552, | |
| "loss": 3.379390335083008, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 6.435716156061797, | |
| "learning_rate": 0.00031333056293101486, | |
| "loss": 3.405352020263672, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 6.456664048180152, | |
| "learning_rate": 0.00031435118365717447, | |
| "loss": 3.523044204711914, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 6.477611940298507, | |
| "learning_rate": 0.00031537180438333414, | |
| "loss": 3.488709259033203, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 6.498559832416863, | |
| "learning_rate": 0.0003163924251094938, | |
| "loss": 3.3736576080322265, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 6.519507724535218, | |
| "learning_rate": 0.0003174130458356535, | |
| "loss": 3.5457527160644533, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 6.540455616653574, | |
| "learning_rate": 0.00031843366656181314, | |
| "loss": 3.38338623046875, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 6.56140350877193, | |
| "learning_rate": 0.0003194542872879728, | |
| "loss": 3.4261444091796873, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 6.5823514008902855, | |
| "learning_rate": 0.0003204749080141324, | |
| "loss": 3.365514373779297, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 6.603299293008641, | |
| "learning_rate": 0.0003214955287402921, | |
| "loss": 3.3958808898925783, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 6.624247185126997, | |
| "learning_rate": 0.00032251614946645176, | |
| "loss": 3.4493579864501953, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 6.6451950772453525, | |
| "learning_rate": 0.00032353677019261143, | |
| "loss": 3.509814453125, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 6.666142969363708, | |
| "learning_rate": 0.0003245573909187711, | |
| "loss": 3.3270954132080077, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 6.687090861482063, | |
| "learning_rate": 0.00032557801164493077, | |
| "loss": 3.407461929321289, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 6.708038753600419, | |
| "learning_rate": 0.0003265986323710903, | |
| "loss": 3.3440326690673827, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 6.728986645718774, | |
| "learning_rate": 0.00032761925309725, | |
| "loss": 3.3119239807128906, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 6.74993453783713, | |
| "learning_rate": 0.00032863987382340966, | |
| "loss": 3.441579818725586, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 6.770882429955486, | |
| "learning_rate": 0.00032966049454956933, | |
| "loss": 3.448055648803711, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 6.791830322073841, | |
| "learning_rate": 0.000330681115275729, | |
| "loss": 3.2744544982910155, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 6.812778214192197, | |
| "learning_rate": 0.0003317017360018886, | |
| "loss": 3.4412532806396485, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 6.833726106310553, | |
| "learning_rate": 0.0003327223567280483, | |
| "loss": 3.459817123413086, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 6.8546739984289085, | |
| "learning_rate": 0.00033374297745420795, | |
| "loss": 3.3376232147216798, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 6.875621890547263, | |
| "learning_rate": 0.0003347635981803676, | |
| "loss": 3.557674789428711, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 6.896569782665619, | |
| "learning_rate": 0.0003357842189065273, | |
| "loss": 3.290896987915039, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 6.917517674783975, | |
| "learning_rate": 0.0003368048396326869, | |
| "loss": 3.49056396484375, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 6.93846556690233, | |
| "learning_rate": 0.00033782546035884657, | |
| "loss": 3.3487998962402346, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 6.959413459020686, | |
| "learning_rate": 0.00033884608108500624, | |
| "loss": 3.3251983642578127, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 6.980361351139042, | |
| "learning_rate": 0.0003398667018111659, | |
| "loss": 3.463846206665039, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 7.002094789211836, | |
| "learning_rate": 0.0003408873225373256, | |
| "loss": 3.5404449462890626, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 7.023042681330191, | |
| "learning_rate": 0.00034190794326348524, | |
| "loss": 3.317121887207031, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 7.043990573448546, | |
| "learning_rate": 0.00034292856398964486, | |
| "loss": 3.446538543701172, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 7.064938465566902, | |
| "learning_rate": 0.0003439491847158045, | |
| "loss": 3.410959243774414, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 7.085886357685258, | |
| "learning_rate": 0.0003449698054419642, | |
| "loss": 3.4593124389648438, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 7.1068342498036134, | |
| "learning_rate": 0.00034599042616812386, | |
| "loss": 3.4331336975097657, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 7.127782141921969, | |
| "learning_rate": 0.00034701104689428353, | |
| "loss": 3.415497970581055, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 7.148730034040325, | |
| "learning_rate": 0.0003480316676204432, | |
| "loss": 3.358320617675781, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 7.1696779261586805, | |
| "learning_rate": 0.0003490522883466028, | |
| "loss": 3.369782257080078, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 7.190625818277036, | |
| "learning_rate": 0.0003500729090727625, | |
| "loss": 3.3603092193603517, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 7.211573710395392, | |
| "learning_rate": 0.00035109352979892215, | |
| "loss": 3.4065528869628907, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 7.232521602513747, | |
| "learning_rate": 0.0003521141505250818, | |
| "loss": 3.368368148803711, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 7.253469494632102, | |
| "learning_rate": 0.0003531347712512415, | |
| "loss": 3.3687610626220703, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 7.274417386750458, | |
| "learning_rate": 0.00035415539197740115, | |
| "loss": 3.397439956665039, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 7.295365278868814, | |
| "learning_rate": 0.00035517601270356077, | |
| "loss": 3.329518508911133, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 7.316313170987169, | |
| "learning_rate": 0.00035619663342972044, | |
| "loss": 3.3606395721435547, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 7.337261063105525, | |
| "learning_rate": 0.0003572172541558801, | |
| "loss": 3.373159408569336, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 7.358208955223881, | |
| "learning_rate": 0.00035823787488203977, | |
| "loss": 3.460713195800781, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 7.379156847342236, | |
| "learning_rate": 0.00035925849560819944, | |
| "loss": 3.3600276947021483, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 7.400104739460592, | |
| "learning_rate": 0.0003602791163343591, | |
| "loss": 3.3381488800048826, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 7.421052631578947, | |
| "learning_rate": 0.0003612997370605187, | |
| "loss": 3.5098854064941407, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 7.442000523697303, | |
| "learning_rate": 0.0003623203577866784, | |
| "loss": 3.358294677734375, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 7.462948415815658, | |
| "learning_rate": 0.00036334097851283806, | |
| "loss": 3.361553955078125, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 7.483896307934014, | |
| "learning_rate": 0.00036436159923899773, | |
| "loss": 3.3908660888671873, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 7.50484420005237, | |
| "learning_rate": 0.0003653822199651574, | |
| "loss": 3.305834197998047, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 7.525792092170725, | |
| "learning_rate": 0.000366402840691317, | |
| "loss": 3.3551094055175783, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 7.546739984289081, | |
| "learning_rate": 0.0003674234614174767, | |
| "loss": 3.336803436279297, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 7.567687876407437, | |
| "learning_rate": 0.00036844408214363635, | |
| "loss": 3.402811050415039, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 7.588635768525792, | |
| "learning_rate": 0.000369464702869796, | |
| "loss": 3.3823123931884767, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 7.609583660644148, | |
| "learning_rate": 0.00037048532359595563, | |
| "loss": 3.3616653442382813, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 7.630531552762504, | |
| "learning_rate": 0.00037150594432211524, | |
| "loss": 3.409253692626953, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 7.6514794448808585, | |
| "learning_rate": 0.0003725265650482749, | |
| "loss": 3.1992008209228517, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 7.672427336999214, | |
| "learning_rate": 0.0003735471857744346, | |
| "loss": 3.2639488220214843, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 7.69337522911757, | |
| "learning_rate": 0.00037456780650059425, | |
| "loss": 3.4038814544677733, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 7.714323121235926, | |
| "learning_rate": 0.0003755884272267539, | |
| "loss": 3.382715606689453, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 7.735271013354281, | |
| "learning_rate": 0.0003766090479529136, | |
| "loss": 3.380691146850586, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 7.756218905472637, | |
| "learning_rate": 0.0003776296686790732, | |
| "loss": 3.3293548583984376, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 7.777166797590993, | |
| "learning_rate": 0.00037865028940523287, | |
| "loss": 3.2755306243896483, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 7.798114689709348, | |
| "learning_rate": 0.00037967091013139253, | |
| "loss": 3.3747108459472654, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 7.819062581827704, | |
| "learning_rate": 0.0003806915308575522, | |
| "loss": 3.325161361694336, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 7.840010473946059, | |
| "learning_rate": 0.00038171215158371187, | |
| "loss": 3.3385120391845704, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 7.8609583660644144, | |
| "learning_rate": 0.00038273277230987154, | |
| "loss": 3.365946960449219, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 7.88190625818277, | |
| "learning_rate": 0.00038375339303603115, | |
| "loss": 3.2765518188476563, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 7.902854150301126, | |
| "learning_rate": 0.0003847740137621908, | |
| "loss": 3.4135189056396484, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 7.9238020424194815, | |
| "learning_rate": 0.0003857946344883505, | |
| "loss": 3.3337100982666015, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 7.944749934537837, | |
| "learning_rate": 0.00038681525521451016, | |
| "loss": 3.469867706298828, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 7.965697826656193, | |
| "learning_rate": 0.0003878358759406698, | |
| "loss": 3.4340835571289063, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 7.9866457187745485, | |
| "learning_rate": 0.00038885649666682944, | |
| "loss": 3.3163192749023436, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 8.008379156847342, | |
| "learning_rate": 0.0003898771173929891, | |
| "loss": 3.4269264221191404, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 8.029327048965698, | |
| "learning_rate": 0.0003908977381191488, | |
| "loss": 3.366756057739258, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 8.050274941084053, | |
| "learning_rate": 0.00039191835884530845, | |
| "loss": 3.5194732666015627, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 8.071222833202409, | |
| "learning_rate": 0.0003929389795714681, | |
| "loss": 3.2884559631347656, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 8.092170725320765, | |
| "learning_rate": 0.0003939596002976278, | |
| "loss": 3.3115074157714846, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 8.11311861743912, | |
| "learning_rate": 0.0003949802210237874, | |
| "loss": 3.342890167236328, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 8.134066509557476, | |
| "learning_rate": 0.00039600084174994706, | |
| "loss": 3.337261962890625, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 8.155014401675832, | |
| "learning_rate": 0.00039702146247610673, | |
| "loss": 3.3896888732910155, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 8.175962293794187, | |
| "learning_rate": 0.0003980420832022664, | |
| "loss": 3.314004898071289, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 8.196910185912543, | |
| "learning_rate": 0.00039906270392842607, | |
| "loss": 3.253472900390625, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 8.217858078030899, | |
| "learning_rate": 0.00040008332465458574, | |
| "loss": 3.2534595489501954, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 8.238805970149254, | |
| "learning_rate": 0.00040110394538074535, | |
| "loss": 3.3420372009277344, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 8.25975386226761, | |
| "learning_rate": 0.000402124566106905, | |
| "loss": 3.3110313415527344, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 8.280701754385966, | |
| "learning_rate": 0.0004031451868330647, | |
| "loss": 3.3015865325927733, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 8.30164964650432, | |
| "learning_rate": 0.00040416580755922436, | |
| "loss": 3.207544708251953, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 8.322597538622675, | |
| "learning_rate": 0.000405186428285384, | |
| "loss": 3.2784183502197264, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 8.343545430741031, | |
| "learning_rate": 0.0004062070490115437, | |
| "loss": 3.29312744140625, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 8.364493322859387, | |
| "learning_rate": 0.0004072276697377033, | |
| "loss": 3.349509048461914, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 8.385441214977742, | |
| "learning_rate": 0.000408248290463863, | |
| "loss": 3.311314010620117, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 8.406389107096098, | |
| "learning_rate": 0.00040926891119002264, | |
| "loss": 3.2704097747802736, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 8.427336999214454, | |
| "learning_rate": 0.0004102895319161823, | |
| "loss": 3.3431529998779297, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 8.44828489133281, | |
| "learning_rate": 0.000411310152642342, | |
| "loss": 3.2444534301757812, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 8.469232783451165, | |
| "learning_rate": 0.0004123307733685016, | |
| "loss": 3.2499820709228517, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 8.49018067556952, | |
| "learning_rate": 0.00041335139409466126, | |
| "loss": 3.4333778381347657, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 8.511128567687877, | |
| "learning_rate": 0.0004143720148208209, | |
| "loss": 3.598118209838867, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 8.532076459806232, | |
| "learning_rate": 0.00041539263554698055, | |
| "loss": 3.1703567504882812, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 8.553024351924588, | |
| "learning_rate": 0.0004164132562731402, | |
| "loss": 3.308456802368164, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 8.573972244042944, | |
| "learning_rate": 0.00041743387699929983, | |
| "loss": 3.262325668334961, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 8.5949201361613, | |
| "learning_rate": 0.0004184544977254595, | |
| "loss": 3.277301788330078, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 8.615868028279655, | |
| "learning_rate": 0.00041947511845161916, | |
| "loss": 3.245453643798828, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 8.63681592039801, | |
| "learning_rate": 0.00042049573917777883, | |
| "loss": 3.246595764160156, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 8.657763812516366, | |
| "learning_rate": 0.0004215163599039385, | |
| "loss": 3.310033416748047, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 8.678711704634722, | |
| "learning_rate": 0.00042253698063009817, | |
| "loss": 3.301158905029297, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 8.699659596753076, | |
| "learning_rate": 0.0004235576013562578, | |
| "loss": 3.2295578002929686, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 8.720607488871432, | |
| "learning_rate": 0.00042457822208241745, | |
| "loss": 3.26501579284668, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 8.741555380989787, | |
| "learning_rate": 0.0004255988428085771, | |
| "loss": 3.346723937988281, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 8.762503273108143, | |
| "learning_rate": 0.0004266194635347368, | |
| "loss": 3.2818328857421877, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 8.783451165226499, | |
| "learning_rate": 0.00042764008426089646, | |
| "loss": 3.3082435607910154, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 8.804399057344854, | |
| "learning_rate": 0.0004286607049870561, | |
| "loss": 3.2165481567382814, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 8.82534694946321, | |
| "learning_rate": 0.00042968132571321574, | |
| "loss": 3.217595672607422, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 8.846294841581566, | |
| "learning_rate": 0.0004307019464393754, | |
| "loss": 3.2202774047851563, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 8.867242733699921, | |
| "learning_rate": 0.0004317225671655351, | |
| "loss": 3.3156604766845703, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 8.888190625818277, | |
| "learning_rate": 0.00043274318789169474, | |
| "loss": 3.2440589904785155, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 8.909138517936633, | |
| "learning_rate": 0.0004337638086178544, | |
| "loss": 3.354948043823242, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 8.930086410054988, | |
| "learning_rate": 0.0004347844293440141, | |
| "loss": 3.2932735443115235, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 8.951034302173344, | |
| "learning_rate": 0.0004358050500701737, | |
| "loss": 3.1642740249633787, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 8.9719821942917, | |
| "learning_rate": 0.00043682567079633336, | |
| "loss": 3.1953109741210937, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 8.992930086410055, | |
| "learning_rate": 0.00043784629152249303, | |
| "loss": 3.1374893188476562, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 9.014663524482849, | |
| "learning_rate": 0.0004388669122486527, | |
| "loss": 3.4447471618652346, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 9.035611416601204, | |
| "learning_rate": 0.00043988753297481237, | |
| "loss": 3.278203582763672, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 9.05655930871956, | |
| "learning_rate": 0.000440908153700972, | |
| "loss": 3.2842811584472655, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 9.077507200837916, | |
| "learning_rate": 0.00044192877442713165, | |
| "loss": 3.2456493377685547, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 9.098455092956272, | |
| "learning_rate": 0.0004429493951532913, | |
| "loss": 3.18524112701416, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 9.119402985074627, | |
| "learning_rate": 0.000443970015879451, | |
| "loss": 3.2943866729736326, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 9.140350877192983, | |
| "learning_rate": 0.00044499063660561065, | |
| "loss": 3.325389862060547, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 9.161298769311339, | |
| "learning_rate": 0.0004460112573317703, | |
| "loss": 3.1400611877441404, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 9.182246661429694, | |
| "learning_rate": 0.00044703187805792994, | |
| "loss": 3.272686004638672, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 9.20319455354805, | |
| "learning_rate": 0.0004480524987840896, | |
| "loss": 3.284004974365234, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 9.224142445666406, | |
| "learning_rate": 0.0004490731195102493, | |
| "loss": 3.2489898681640623, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 9.24509033778476, | |
| "learning_rate": 0.00045009374023640894, | |
| "loss": 3.2408329010009767, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 9.266038229903115, | |
| "learning_rate": 0.0004511143609625686, | |
| "loss": 3.2516738891601564, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 9.286986122021471, | |
| "learning_rate": 0.0004521349816887283, | |
| "loss": 3.1824373245239257, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 9.307934014139827, | |
| "learning_rate": 0.0004531556024148879, | |
| "loss": 3.2249637603759767, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 9.328881906258182, | |
| "learning_rate": 0.00045417622314104756, | |
| "loss": 3.174178695678711, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 9.349829798376538, | |
| "learning_rate": 0.00045519684386720723, | |
| "loss": 3.188156318664551, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 9.370777690494894, | |
| "learning_rate": 0.0004562174645933669, | |
| "loss": 3.228466796875, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 9.39172558261325, | |
| "learning_rate": 0.00045723808531952657, | |
| "loss": 3.2763172149658204, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 9.412673474731605, | |
| "learning_rate": 0.0004582587060456861, | |
| "loss": 3.266025161743164, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 9.43362136684996, | |
| "learning_rate": 0.0004592793267718458, | |
| "loss": 3.2390396118164064, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 9.454569258968316, | |
| "learning_rate": 0.00046029994749800546, | |
| "loss": 3.23939208984375, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 9.475517151086672, | |
| "learning_rate": 0.00046132056822416513, | |
| "loss": 3.219116973876953, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 9.496465043205028, | |
| "learning_rate": 0.0004623411889503248, | |
| "loss": 3.277789306640625, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 9.517412935323383, | |
| "learning_rate": 0.00046336180967648447, | |
| "loss": 3.284175491333008, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 9.538360827441739, | |
| "learning_rate": 0.0004643824304026441, | |
| "loss": 3.208120346069336, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 9.559308719560095, | |
| "learning_rate": 0.00046540305112880375, | |
| "loss": 3.271435546875, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 9.58025661167845, | |
| "learning_rate": 0.0004664236718549634, | |
| "loss": 3.2989322662353517, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 9.601204503796806, | |
| "learning_rate": 0.0004674442925811231, | |
| "loss": 3.1721576690673827, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 9.622152395915162, | |
| "learning_rate": 0.00046846491330728275, | |
| "loss": 3.1907968521118164, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 9.643100288033516, | |
| "learning_rate": 0.00046948553403344237, | |
| "loss": 3.3379592895507812, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 9.664048180151871, | |
| "learning_rate": 0.00047050615475960204, | |
| "loss": 3.2489646911621093, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 9.684996072270227, | |
| "learning_rate": 0.0004715267754857617, | |
| "loss": 3.3012271881103517, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 9.705943964388583, | |
| "learning_rate": 0.00047254739621192137, | |
| "loss": 3.235492706298828, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 9.726891856506938, | |
| "learning_rate": 0.00047356801693808104, | |
| "loss": 3.2170711517333985, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 9.747839748625294, | |
| "learning_rate": 0.0004745886376642407, | |
| "loss": 3.1358510971069338, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 9.76878764074365, | |
| "learning_rate": 0.0004756092583904003, | |
| "loss": 3.154219055175781, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 9.789735532862005, | |
| "learning_rate": 0.00047662987911656, | |
| "loss": 3.3397506713867187, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 9.810683424980361, | |
| "learning_rate": 0.00047765049984271966, | |
| "loss": 3.2498783111572265, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 9.831631317098717, | |
| "learning_rate": 0.00047867112056887933, | |
| "loss": 3.2211677551269533, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 9.852579209217073, | |
| "learning_rate": 0.000479691741295039, | |
| "loss": 3.2689888000488283, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 9.873527101335428, | |
| "learning_rate": 0.00048071236202119866, | |
| "loss": 3.1732282638549805, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 9.894474993453784, | |
| "learning_rate": 0.0004817329827473583, | |
| "loss": 3.1548320770263674, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 9.91542288557214, | |
| "learning_rate": 0.00048275360347351795, | |
| "loss": 3.2713703155517577, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 9.936370777690495, | |
| "learning_rate": 0.0004837742241996776, | |
| "loss": 3.1592134475708007, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 9.957318669808851, | |
| "learning_rate": 0.0004847948449258373, | |
| "loss": 3.213180923461914, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 9.978266561927207, | |
| "learning_rate": 0.00048581546565199695, | |
| "loss": 3.2513309478759767, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 9.999214454045562, | |
| "learning_rate": 0.0004868360863781566, | |
| "loss": 3.3612499237060547, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 10.020947892118356, | |
| "learning_rate": 0.00048785670710431623, | |
| "loss": 3.2473423004150392, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 10.041895784236711, | |
| "learning_rate": 0.000488877327830476, | |
| "loss": 3.3008792877197264, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 10.062843676355067, | |
| "learning_rate": 0.0004898979485566356, | |
| "loss": 3.286971664428711, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 10.083791568473423, | |
| "learning_rate": 0.0004909185692827952, | |
| "loss": 3.300416946411133, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 10.104739460591778, | |
| "learning_rate": 0.0004919391900089549, | |
| "loss": 3.1919363021850584, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 10.125687352710134, | |
| "learning_rate": 0.0004929598107351145, | |
| "loss": 3.2064422607421874, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 10.14663524482849, | |
| "learning_rate": 0.0004939804314612742, | |
| "loss": 3.1432363510131838, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 10.167583136946845, | |
| "learning_rate": 0.0004950010521874339, | |
| "loss": 3.17150764465332, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 10.188531029065201, | |
| "learning_rate": 0.0004960216729135935, | |
| "loss": 3.2499244689941404, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 10.209478921183555, | |
| "learning_rate": 0.0004970422936397532, | |
| "loss": 3.1550519943237303, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 10.23042681330191, | |
| "learning_rate": 0.0004980629143659128, | |
| "loss": 3.1220830917358398, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 10.251374705420266, | |
| "learning_rate": 0.0004990835350920725, | |
| "loss": 3.2074295043945313, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 10.272322597538622, | |
| "learning_rate": 0.0005001041558182321, | |
| "loss": 3.198388862609863, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 10.293270489656978, | |
| "learning_rate": 0.0005011247765443918, | |
| "loss": 3.218120574951172, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 10.314218381775333, | |
| "learning_rate": 0.0005021453972705514, | |
| "loss": 3.2131175994873047, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 10.33516627389369, | |
| "learning_rate": 0.0005031660179967111, | |
| "loss": 3.1987491607666017, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 10.356114166012045, | |
| "learning_rate": 0.0005041866387228707, | |
| "loss": 3.175269889831543, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 10.3770620581304, | |
| "learning_rate": 0.0005052072594490304, | |
| "loss": 3.183473014831543, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 10.398009950248756, | |
| "learning_rate": 0.00050622788017519, | |
| "loss": 3.0571062088012697, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 10.418957842367112, | |
| "learning_rate": 0.0005072485009013497, | |
| "loss": 3.325624465942383, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 10.439905734485468, | |
| "learning_rate": 0.0005082691216275094, | |
| "loss": 3.249886703491211, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 10.460853626603823, | |
| "learning_rate": 0.000509289742353669, | |
| "loss": 3.18145866394043, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 10.481801518722179, | |
| "learning_rate": 0.0005103103630798287, | |
| "loss": 3.1750720977783202, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 10.502749410840535, | |
| "learning_rate": 0.0005098008169087462, | |
| "loss": 3.298409271240234, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 10.52369730295889, | |
| "learning_rate": 0.0005092927940452339, | |
| "loss": 3.0969064712524412, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 10.544645195077246, | |
| "learning_rate": 0.0005087862869144114, | |
| "loss": 3.3151134490966796, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 10.565593087195602, | |
| "learning_rate": 0.0005082812879940277, | |
| "loss": 3.2290550231933595, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 10.586540979313957, | |
| "learning_rate": 0.0005077777898139921, | |
| "loss": 3.140799713134766, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 10.607488871432313, | |
| "learning_rate": 0.0005072757849559103, | |
| "loss": 3.1083478927612305, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 10.628436763550667, | |
| "learning_rate": 0.0005067752660526248, | |
| "loss": 3.1976173400878904, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 10.649384655669023, | |
| "learning_rate": 0.0005062762257877613, | |
| "loss": 3.2580982208251954, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 10.670332547787378, | |
| "learning_rate": 0.0005057786568952791, | |
| "loss": 3.280201721191406, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 10.691280439905734, | |
| "learning_rate": 0.000505282552159027, | |
| "loss": 3.141094779968262, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 10.71222833202409, | |
| "learning_rate": 0.000504787904412304, | |
| "loss": 3.095392608642578, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 10.733176224142445, | |
| "learning_rate": 0.000504294706537424, | |
| "loss": 3.2083145141601563, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 10.754124116260801, | |
| "learning_rate": 0.0005038029514652858, | |
| "loss": 3.289701461791992, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 10.775072008379157, | |
| "learning_rate": 0.0005033126321749477, | |
| "loss": 3.2425827026367187, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 10.796019900497512, | |
| "learning_rate": 0.000502823741693206, | |
| "loss": 3.277665710449219, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 10.816967792615868, | |
| "learning_rate": 0.0005023362730941793, | |
| "loss": 3.1909582138061525, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 10.837915684734224, | |
| "learning_rate": 0.0005018502194988955, | |
| "loss": 3.170912170410156, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 10.85886357685258, | |
| "learning_rate": 0.0005013655740748848, | |
| "loss": 3.2589969635009766, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 10.879811468970935, | |
| "learning_rate": 0.0005008823300357761, | |
| "loss": 3.1849817276000976, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 10.90075936108929, | |
| "learning_rate": 0.0005004004806408972, | |
| "loss": 3.1448366165161135, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 10.921707253207646, | |
| "learning_rate": 0.0004999200191948814, | |
| "loss": 3.122829055786133, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 10.942655145326002, | |
| "learning_rate": 0.0004994409390472751, | |
| "loss": 3.1754734039306642, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 10.963603037444358, | |
| "learning_rate": 0.0004989632335921523, | |
| "loss": 3.1149194717407225, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 10.984550929562714, | |
| "learning_rate": 0.0004984868962677315, | |
| "loss": 3.100501823425293, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 11.006284367635507, | |
| "learning_rate": 0.0004980119205559973, | |
| "loss": 3.330778121948242, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 11.027232259753863, | |
| "learning_rate": 0.0004975382999823259, | |
| "loss": 3.1559074401855467, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 11.048180151872218, | |
| "learning_rate": 0.0004970660281151141, | |
| "loss": 3.172486114501953, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 11.069128043990574, | |
| "learning_rate": 0.0004965950985654126, | |
| "loss": 3.1997749328613283, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 11.09007593610893, | |
| "learning_rate": 0.0004961255049865635, | |
| "loss": 3.2853694915771485, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 11.111023828227285, | |
| "learning_rate": 0.0004956572410738401, | |
| "loss": 3.145161819458008, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 11.131971720345641, | |
| "learning_rate": 0.000495190300564092, | |
| "loss": 3.126105308532715, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 11.152919612463997, | |
| "learning_rate": 0.0004947246772353933, | |
| "loss": 3.2012374877929686, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 11.17386750458235, | |
| "learning_rate": 0.0004942603649066942, | |
| "loss": 3.1729455947875977, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 11.194815396700706, | |
| "learning_rate": 0.0004937973574374762, | |
| "loss": 3.148386001586914, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 11.215763288819062, | |
| "learning_rate": 0.0004933356487274114, | |
| "loss": 3.065207290649414, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 11.236711180937418, | |
| "learning_rate": 0.0004928752327160248, | |
| "loss": 3.150010108947754, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 11.257659073055773, | |
| "learning_rate": 0.0004924161033823598, | |
| "loss": 3.2157524108886717, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 11.278606965174129, | |
| "learning_rate": 0.0004919582547446482, | |
| "loss": 3.0866676330566407, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 11.299554857292485, | |
| "learning_rate": 0.0004915016808599824, | |
| "loss": 3.2173648834228517, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 11.32050274941084, | |
| "learning_rate": 0.0004910463758239914, | |
| "loss": 3.1369649887084963, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 11.341450641529196, | |
| "learning_rate": 0.0004905923337705201, | |
| "loss": 3.0868097305297852, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 11.362398533647552, | |
| "learning_rate": 0.0004901395488713123, | |
| "loss": 3.2488777160644533, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 11.383346425765907, | |
| "learning_rate": 0.0004896880153356963, | |
| "loss": 3.121846008300781, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 11.404294317884263, | |
| "learning_rate": 0.000489237727410273, | |
| "loss": 3.1137548446655274, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 11.425242210002619, | |
| "learning_rate": 0.0004887886793786093, | |
| "loss": 3.0799299240112306, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 11.446190102120974, | |
| "learning_rate": 0.0004883408655609327, | |
| "loss": 3.0961063385009764, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 11.46713799423933, | |
| "learning_rate": 0.0004878942803138293, | |
| "loss": 3.0329910278320313, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 11.488085886357686, | |
| "learning_rate": 0.0004874489180299454, | |
| "loss": 3.173397445678711, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 11.509033778476041, | |
| "learning_rate": 0.00048700477313769213, | |
| "loss": 3.1852407455444336, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 11.529981670594397, | |
| "learning_rate": 0.00048656184010095185, | |
| "loss": 3.2040431976318358, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 11.550929562712753, | |
| "learning_rate": 0.00048612011341878916, | |
| "loss": 3.189468002319336, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 11.571877454831107, | |
| "learning_rate": 0.0004856795876251634, | |
| "loss": 3.2286914825439452, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 11.592825346949462, | |
| "learning_rate": 0.00048524025728864493, | |
| "loss": 3.034438896179199, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 11.613773239067818, | |
| "learning_rate": 0.0004848021170121335, | |
| "loss": 3.085980987548828, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 11.634721131186174, | |
| "learning_rate": 0.0004843651614325803, | |
| "loss": 3.1111934661865233, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 11.65566902330453, | |
| "learning_rate": 0.00048392938522071163, | |
| "loss": 3.159061050415039, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 11.676616915422885, | |
| "learning_rate": 0.0004834947830807563, | |
| "loss": 3.0602264404296875, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 11.69756480754124, | |
| "learning_rate": 0.00048306134975017523, | |
| "loss": 3.120003890991211, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 11.718512699659597, | |
| "learning_rate": 0.0004826290799993939, | |
| "loss": 3.1456703186035155, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 11.739460591777952, | |
| "learning_rate": 0.0004821979686315372, | |
| "loss": 3.112548828125, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 11.760408483896308, | |
| "learning_rate": 0.00048176801048216693, | |
| "loss": 3.1137924194335938, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 11.781356376014664, | |
| "learning_rate": 0.0004813392004190223, | |
| "loss": 3.146605110168457, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 11.80230426813302, | |
| "learning_rate": 0.00048091153334176224, | |
| "loss": 3.1099647521972655, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 11.823252160251375, | |
| "learning_rate": 0.00048048500418171097, | |
| "loss": 3.099277305603027, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 11.84420005236973, | |
| "learning_rate": 0.0004800596079016053, | |
| "loss": 3.085763931274414, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 11.865147944488086, | |
| "learning_rate": 0.0004796353394953452, | |
| "loss": 3.188782501220703, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 11.886095836606442, | |
| "learning_rate": 0.0004792121939877459, | |
| "loss": 3.176821708679199, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 11.907043728724798, | |
| "learning_rate": 0.00047879016643429336, | |
| "loss": 3.067020797729492, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 11.927991620843153, | |
| "learning_rate": 0.00047836925192090116, | |
| "loss": 3.1511611938476562, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 11.948939512961509, | |
| "learning_rate": 0.0004779494455636703, | |
| "loss": 3.0738733291625975, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 11.969887405079863, | |
| "learning_rate": 0.00047753074250865145, | |
| "loss": 3.237213897705078, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 11.990835297198219, | |
| "learning_rate": 0.00047711313793160877, | |
| "loss": 3.052178382873535, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 12.012568735271014, | |
| "learning_rate": 0.000476696627037787, | |
| "loss": 3.2651294708251952, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 12.03351662738937, | |
| "learning_rate": 0.0004762812050616797, | |
| "loss": 3.162643241882324, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 12.054464519507725, | |
| "learning_rate": 0.0004758668672668006, | |
| "loss": 3.1709291458129885, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 12.07541241162608, | |
| "learning_rate": 0.00047545360894545664, | |
| "loss": 3.063345527648926, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 12.096360303744436, | |
| "learning_rate": 0.0004750414254185235, | |
| "loss": 3.093794250488281, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 12.117308195862792, | |
| "learning_rate": 0.0004746303120352226, | |
| "loss": 3.1082719802856444, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 12.138256087981146, | |
| "learning_rate": 0.00047422026417290146, | |
| "loss": 3.1271081924438477, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 12.159203980099502, | |
| "learning_rate": 0.0004738112772368146, | |
| "loss": 3.141692543029785, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 12.180151872217857, | |
| "learning_rate": 0.00047340334665990787, | |
| "loss": 3.1134639739990235, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 12.201099764336213, | |
| "learning_rate": 0.0004729964679026039, | |
| "loss": 3.03677921295166, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 12.222047656454569, | |
| "learning_rate": 0.0004725906364525903, | |
| "loss": 3.2071063995361326, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 12.242995548572924, | |
| "learning_rate": 0.0004721858478246089, | |
| "loss": 3.173069953918457, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 12.26394344069128, | |
| "learning_rate": 0.0004717820975602482, | |
| "loss": 3.048240089416504, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 12.284891332809636, | |
| "learning_rate": 0.0004713793812277367, | |
| "loss": 3.041463088989258, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 12.305839224927992, | |
| "learning_rate": 0.00047097769442173856, | |
| "loss": 3.0645767211914063, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 12.326787117046347, | |
| "learning_rate": 0.00047057703276315164, | |
| "loss": 2.982158088684082, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 12.347735009164703, | |
| "learning_rate": 0.0004701773918989065, | |
| "loss": 3.111321449279785, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 12.368682901283059, | |
| "learning_rate": 0.00046977876750176805, | |
| "loss": 3.141143798828125, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 12.389630793401414, | |
| "learning_rate": 0.0004693811552701385, | |
| "loss": 3.0916566848754883, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 12.41057868551977, | |
| "learning_rate": 0.0004689845509278626, | |
| "loss": 3.0807928085327148, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 12.431526577638126, | |
| "learning_rate": 0.00046858895022403474, | |
| "loss": 3.0439529418945312, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 12.452474469756481, | |
| "learning_rate": 0.000468194348932807, | |
| "loss": 3.0425508499145506, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 12.473422361874837, | |
| "learning_rate": 0.00046780074285319984, | |
| "loss": 3.187800407409668, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 12.494370253993193, | |
| "learning_rate": 0.0004674081278089144, | |
| "loss": 3.0460309982299805, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 12.515318146111547, | |
| "learning_rate": 0.00046701649964814616, | |
| "loss": 3.2187931060791017, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 12.536266038229904, | |
| "learning_rate": 0.0004666258542434007, | |
| "loss": 3.109378433227539, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 12.557213930348258, | |
| "learning_rate": 0.000466236187491311, | |
| "loss": 3.0655149459838866, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 12.578161822466614, | |
| "learning_rate": 0.00046584749531245617, | |
| "loss": 3.132980728149414, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 12.59910971458497, | |
| "learning_rate": 0.0004654597736511823, | |
| "loss": 3.0032047271728515, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 12.620057606703325, | |
| "learning_rate": 0.0004650730184754247, | |
| "loss": 3.0569095611572266, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 12.64100549882168, | |
| "learning_rate": 0.0004646872257765318, | |
| "loss": 3.1891340255737304, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 12.661953390940036, | |
| "learning_rate": 0.00046430239156909045, | |
| "loss": 3.036951446533203, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 12.682901283058392, | |
| "learning_rate": 0.00046391851189075343, | |
| "loss": 3.077804374694824, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 12.703849175176748, | |
| "learning_rate": 0.00046353558280206746, | |
| "loss": 3.1198028564453124, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 12.724797067295103, | |
| "learning_rate": 0.00046315360038630404, | |
| "loss": 3.1289579391479494, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 12.745744959413459, | |
| "learning_rate": 0.0004627725607492909, | |
| "loss": 3.047295570373535, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 12.766692851531815, | |
| "learning_rate": 0.00046239246001924503, | |
| "loss": 3.071992111206055, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 12.78764074365017, | |
| "learning_rate": 0.000462013294346608, | |
| "loss": 3.050577735900879, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 12.808588635768526, | |
| "learning_rate": 0.00046163505990388167, | |
| "loss": 3.0774341583251954, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 12.829536527886882, | |
| "learning_rate": 0.00046125775288546623, | |
| "loss": 3.120297431945801, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 12.850484420005237, | |
| "learning_rate": 0.00046088136950749937, | |
| "loss": 3.1301042556762697, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 12.871432312123593, | |
| "learning_rate": 0.0004605059060076967, | |
| "loss": 3.0827388763427734, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 12.892380204241949, | |
| "learning_rate": 0.0004601313586451939, | |
| "loss": 3.136738967895508, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 12.913328096360305, | |
| "learning_rate": 0.00045975772370039034, | |
| "loss": 3.0567752838134767, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 12.93427598847866, | |
| "learning_rate": 0.0004593849974747937, | |
| "loss": 3.053047752380371, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 12.955223880597014, | |
| "learning_rate": 0.0004590131762908664, | |
| "loss": 3.0607650756835936, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 12.97617177271537, | |
| "learning_rate": 0.00045864225649187287, | |
| "loss": 3.1021827697753905, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 12.997119664833725, | |
| "learning_rate": 0.000458272234441729, | |
| "loss": 3.016301727294922, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 13.01885310290652, | |
| "learning_rate": 0.00045790310652485205, | |
| "loss": 3.1855663299560546, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 13.039800995024876, | |
| "learning_rate": 0.0004575348691460124, | |
| "loss": 3.0596897125244142, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 13.060748887143232, | |
| "learning_rate": 0.00045716751873018654, | |
| "loss": 3.061813735961914, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 13.081696779261588, | |
| "learning_rate": 0.00045680105172241103, | |
| "loss": 3.106767463684082, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 13.102644671379942, | |
| "learning_rate": 0.0004564354645876384, | |
| "loss": 3.101357269287109, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 13.123592563498297, | |
| "learning_rate": 0.00045607075381059363, | |
| "loss": 3.085792350769043, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 13.144540455616653, | |
| "learning_rate": 0.00045570691589563234, | |
| "loss": 3.071797752380371, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 13.165488347735009, | |
| "learning_rate": 0.0004553439473666, | |
| "loss": 3.087900161743164, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 13.186436239853364, | |
| "learning_rate": 0.0004549818447666924, | |
| "loss": 3.0981624603271483, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 13.20738413197172, | |
| "learning_rate": 0.00045462060465831743, | |
| "loss": 2.995559501647949, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 13.228332024090076, | |
| "learning_rate": 0.0004542602236229581, | |
| "loss": 3.0986444473266603, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 13.249279916208431, | |
| "learning_rate": 0.00045390069826103653, | |
| "loss": 3.1423923492431642, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 13.270227808326787, | |
| "learning_rate": 0.00045354202519177925, | |
| "loss": 2.981964111328125, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 13.291175700445143, | |
| "learning_rate": 0.0004531842010530839, | |
| "loss": 3.062668800354004, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 13.312123592563498, | |
| "learning_rate": 0.0004528272225013865, | |
| "loss": 3.1759321212768556, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 13.333071484681854, | |
| "learning_rate": 0.00045247108621153056, | |
| "loss": 3.0892358779907227, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 13.35401937680021, | |
| "learning_rate": 0.0004521157888766368, | |
| "loss": 3.0303468704223633, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 13.374967268918565, | |
| "learning_rate": 0.00045176132720797443, | |
| "loss": 3.042502021789551, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 13.395915161036921, | |
| "learning_rate": 0.0004514076979348328, | |
| "loss": 3.109409713745117, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 13.416863053155277, | |
| "learning_rate": 0.0004510548978043951, | |
| "loss": 3.0352380752563475, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 13.437810945273633, | |
| "learning_rate": 0.00045070292358161265, | |
| "loss": 2.9740083694458006, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 13.458758837391988, | |
| "learning_rate": 0.0004503517720490801, | |
| "loss": 3.1181098937988283, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 13.479706729510344, | |
| "learning_rate": 0.000450001440006912, | |
| "loss": 3.089175987243652, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 13.500654621628698, | |
| "learning_rate": 0.00044965192427262043, | |
| "loss": 3.0885658264160156, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 13.521602513747053, | |
| "learning_rate": 0.0004493032216809934, | |
| "loss": 3.0794023513793944, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 13.54255040586541, | |
| "learning_rate": 0.00044895532908397455, | |
| "loss": 3.0824106216430662, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 13.563498297983765, | |
| "learning_rate": 0.00044860824335054384, | |
| "loss": 3.0646196365356446, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 13.58444619010212, | |
| "learning_rate": 0.00044826196136659916, | |
| "loss": 3.044062614440918, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 13.605394082220476, | |
| "learning_rate": 0.00044791648003483884, | |
| "loss": 3.0133747100830077, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 13.626341974338832, | |
| "learning_rate": 0.0004475717962746455, | |
| "loss": 3.070328712463379, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 13.647289866457188, | |
| "learning_rate": 0.0004472279070219706, | |
| "loss": 3.0279052734375, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 13.668237758575543, | |
| "learning_rate": 0.00044688480922922, | |
| "loss": 3.147620964050293, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 13.689185650693899, | |
| "learning_rate": 0.00044654249986514057, | |
| "loss": 3.1151987075805665, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 13.710133542812255, | |
| "learning_rate": 0.0004462009759147076, | |
| "loss": 3.072108268737793, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 13.73108143493061, | |
| "learning_rate": 0.0004458602343790135, | |
| "loss": 3.134627342224121, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 13.752029327048966, | |
| "learning_rate": 0.00044552027227515704, | |
| "loss": 3.032268524169922, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 13.772977219167322, | |
| "learning_rate": 0.00044518108663613355, | |
| "loss": 3.065017509460449, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 13.793925111285677, | |
| "learning_rate": 0.00044484267451072644, | |
| "loss": 3.0611106872558596, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 13.814873003404033, | |
| "learning_rate": 0.0004445050329633992, | |
| "loss": 3.2091243743896483, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 13.835820895522389, | |
| "learning_rate": 0.0004441681590741884, | |
| "loss": 3.058238983154297, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 13.856768787640744, | |
| "learning_rate": 0.0004438320499385977, | |
| "loss": 3.073333168029785, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 13.8777166797591, | |
| "learning_rate": 0.00044349670266749286, | |
| "loss": 3.037291145324707, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 13.898664571877454, | |
| "learning_rate": 0.0004431621143869969, | |
| "loss": 3.0411745071411134, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 13.91961246399581, | |
| "learning_rate": 0.00044282828223838727, | |
| "loss": 3.148990440368652, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 13.940560356114165, | |
| "learning_rate": 0.0004424952033779929, | |
| "loss": 3.014286994934082, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 13.961508248232521, | |
| "learning_rate": 0.00044216287497709253, | |
| "loss": 3.0138343811035155, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 13.982456140350877, | |
| "learning_rate": 0.0004418312942218139, | |
| "loss": 2.9974302291870116, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 14.004189578423672, | |
| "learning_rate": 0.0004415004583130336, | |
| "loss": 3.1527809143066405, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 14.025137470542028, | |
| "learning_rate": 0.0004411703644662778, | |
| "loss": 2.9581697463989256, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 14.046085362660381, | |
| "learning_rate": 0.00044084100991162385, | |
| "loss": 3.0720396041870117, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 14.067033254778737, | |
| "learning_rate": 0.00044051239189360286, | |
| "loss": 3.0470098495483398, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 14.087981146897093, | |
| "learning_rate": 0.00044018450767110235, | |
| "loss": 3.0677566528320312, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 14.108929039015448, | |
| "learning_rate": 0.0004398573545172709, | |
| "loss": 3.110503005981445, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 14.129876931133804, | |
| "learning_rate": 0.0004395309297194223, | |
| "loss": 2.9687520980834963, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 14.15082482325216, | |
| "learning_rate": 0.0004392052305789416, | |
| "loss": 2.985172080993652, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 14.171772715370516, | |
| "learning_rate": 0.0004388802544111908, | |
| "loss": 3.0017110824584963, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 14.192720607488871, | |
| "learning_rate": 0.0004385559985454165, | |
| "loss": 3.041835403442383, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 14.213668499607227, | |
| "learning_rate": 0.0004382324603246575, | |
| "loss": 3.0984907150268555, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 14.234616391725583, | |
| "learning_rate": 0.0004379096371056532, | |
| "loss": 3.0736331939697266, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 14.255564283843938, | |
| "learning_rate": 0.000437587526258753, | |
| "loss": 2.9994585037231447, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 14.276512175962294, | |
| "learning_rate": 0.0004372661251678265, | |
| "loss": 3.1013252258300783, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 14.29746006808065, | |
| "learning_rate": 0.00043694543123017407, | |
| "loss": 3.102655220031738, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 14.318407960199005, | |
| "learning_rate": 0.0004366254418564382, | |
| "loss": 3.1105621337890623, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 14.339355852317361, | |
| "learning_rate": 0.0004363061544705161, | |
| "loss": 3.1368709564208985, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 14.360303744435717, | |
| "learning_rate": 0.0004359875665094723, | |
| "loss": 3.052720069885254, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 14.381251636554072, | |
| "learning_rate": 0.00043566967542345227, | |
| "loss": 2.977310562133789, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 14.402199528672428, | |
| "learning_rate": 0.00043535247867559673, | |
| "loss": 3.111159896850586, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 14.423147420790784, | |
| "learning_rate": 0.00043503597374195665, | |
| "loss": 2.959975814819336, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 14.444095312909138, | |
| "learning_rate": 0.0004347201581114088, | |
| "loss": 3.125636100769043, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 14.465043205027493, | |
| "learning_rate": 0.0004344050292855724, | |
| "loss": 2.9577571868896486, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 14.485991097145849, | |
| "learning_rate": 0.00043409058477872554, | |
| "loss": 3.101388931274414, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 14.506938989264205, | |
| "learning_rate": 0.00043377682211772343, | |
| "loss": 3.0661073684692384, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 14.52788688138256, | |
| "learning_rate": 0.0004334637388419161, | |
| "loss": 2.974909019470215, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 14.548834773500916, | |
| "learning_rate": 0.0004331513325030681, | |
| "loss": 3.0314458847045898, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 14.569782665619272, | |
| "learning_rate": 0.0004328396006652773, | |
| "loss": 3.1744915008544923, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 14.590730557737627, | |
| "learning_rate": 0.00043252854090489564, | |
| "loss": 3.06768741607666, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 14.611678449855983, | |
| "learning_rate": 0.00043221815081044985, | |
| "loss": 3.161996269226074, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 14.632626341974339, | |
| "learning_rate": 0.00043190842798256285, | |
| "loss": 3.102631378173828, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 14.653574234092694, | |
| "learning_rate": 0.00043159937003387584, | |
| "loss": 3.004058074951172, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 14.67452212621105, | |
| "learning_rate": 0.00043129097458897135, | |
| "loss": 3.135270118713379, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 14.695470018329406, | |
| "learning_rate": 0.000430983239284296, | |
| "loss": 3.0084003448486327, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 14.716417910447761, | |
| "learning_rate": 0.0004306761617680849, | |
| "loss": 2.9995773315429686, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 14.737365802566117, | |
| "learning_rate": 0.00043036973970028583, | |
| "loss": 3.000468444824219, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 14.758313694684473, | |
| "learning_rate": 0.00043006397075248464, | |
| "loss": 3.0801364898681642, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 14.779261586802829, | |
| "learning_rate": 0.00042975885260783056, | |
| "loss": 3.0909229278564454, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 14.800209478921184, | |
| "learning_rate": 0.00042945438296096303, | |
| "loss": 2.9928516387939452, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 14.82115737103954, | |
| "learning_rate": 0.0004291505595179379, | |
| "loss": 3.036148262023926, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 14.842105263157894, | |
| "learning_rate": 0.0004288473799961553, | |
| "loss": 3.0693193435668946, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 14.863053155276251, | |
| "learning_rate": 0.0004285448421242875, | |
| "loss": 3.0427278518676757, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 14.884001047394605, | |
| "learning_rate": 0.00042824294364220724, | |
| "loss": 2.9749155044555664, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 14.90494893951296, | |
| "learning_rate": 0.0004279416823009172, | |
| "loss": 2.9990673065185547, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 14.925896831631317, | |
| "learning_rate": 0.0004276410558624791, | |
| "loss": 3.091754913330078, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 14.946844723749672, | |
| "learning_rate": 0.0004273410620999446, | |
| "loss": 3.003107452392578, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 14.967792615868028, | |
| "learning_rate": 0.0004270416987972853, | |
| "loss": 3.023390007019043, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 14.988740507986384, | |
| "learning_rate": 0.00042674296374932424, | |
| "loss": 3.068536376953125, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 15.010473946059177, | |
| "learning_rate": 0.0004264448547616681, | |
| "loss": 3.245321273803711, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 15.031421838177533, | |
| "learning_rate": 0.00042614736965063864, | |
| "loss": 2.9358680725097654, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 15.052369730295888, | |
| "learning_rate": 0.0004258505062432064, | |
| "loss": 3.0279872894287108, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 15.073317622414244, | |
| "learning_rate": 0.0004255542623769234, | |
| "loss": 2.96344108581543, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 15.0942655145326, | |
| "learning_rate": 0.00042525863589985727, | |
| "loss": 3.1603927612304688, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 15.115213406650955, | |
| "learning_rate": 0.00042496362467052564, | |
| "loss": 3.0409677505493162, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 15.136161298769311, | |
| "learning_rate": 0.00042466922655783073, | |
| "loss": 3.154404067993164, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 15.157109190887667, | |
| "learning_rate": 0.00042437543944099504, | |
| "loss": 2.999993324279785, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 15.178057083006022, | |
| "learning_rate": 0.00042408226120949674, | |
| "loss": 2.962456703186035, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 15.199004975124378, | |
| "learning_rate": 0.00042378968976300647, | |
| "loss": 3.050062561035156, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 15.219952867242734, | |
| "learning_rate": 0.00042349772301132377, | |
| "loss": 3.058196258544922, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 15.24090075936109, | |
| "learning_rate": 0.0004232063588743146, | |
| "loss": 3.0242469787597654, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 15.261848651479445, | |
| "learning_rate": 0.00042291559528184904, | |
| "loss": 2.929056930541992, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 15.2827965435978, | |
| "learning_rate": 0.0004226254301737393, | |
| "loss": 3.0346649169921873, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 15.303744435716157, | |
| "learning_rate": 0.0004223358614996787, | |
| "loss": 3.079379081726074, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 15.324692327834512, | |
| "learning_rate": 0.00042204688721918075, | |
| "loss": 2.984081268310547, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 15.345640219952868, | |
| "learning_rate": 0.0004217585053015187, | |
| "loss": 3.025343322753906, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 15.366588112071224, | |
| "learning_rate": 0.0004214707137256656, | |
| "loss": 3.016037940979004, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 15.38753600418958, | |
| "learning_rate": 0.0004211835104802349, | |
| "loss": 3.1470058441162108, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 15.408483896307935, | |
| "learning_rate": 0.00042089689356342115, | |
| "loss": 3.007353591918945, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 15.429431788426289, | |
| "learning_rate": 0.0004206108609829418, | |
| "loss": 3.0402362823486326, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 15.450379680544645, | |
| "learning_rate": 0.00042032541075597875, | |
| "loss": 3.018893241882324, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 15.471327572663, | |
| "learning_rate": 0.0004200405409091207, | |
| "loss": 3.0513105392456055, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 15.492275464781356, | |
| "learning_rate": 0.00041975624947830593, | |
| "loss": 3.0438756942749023, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 15.513223356899712, | |
| "learning_rate": 0.00041947253450876515, | |
| "loss": 2.977249526977539, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 15.534171249018067, | |
| "learning_rate": 0.00041918939405496546, | |
| "loss": 2.956187629699707, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 15.555119141136423, | |
| "learning_rate": 0.00041890682618055396, | |
| "loss": 3.0582775115966796, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 15.576067033254779, | |
| "learning_rate": 0.0004186248289583023, | |
| "loss": 2.9948537826538084, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 15.597014925373134, | |
| "learning_rate": 0.00041834340047005144, | |
| "loss": 2.9837194442749024, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 15.61796281749149, | |
| "learning_rate": 0.0004180625388066569, | |
| "loss": 3.0729391098022463, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 15.638910709609846, | |
| "learning_rate": 0.00041778224206793433, | |
| "loss": 3.054386329650879, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 15.659858601728201, | |
| "learning_rate": 0.00041750250836260536, | |
| "loss": 3.102676582336426, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 15.680806493846557, | |
| "learning_rate": 0.0004172233358082443, | |
| "loss": 2.986006164550781, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 15.701754385964913, | |
| "learning_rate": 0.00041694472253122467, | |
| "loss": 3.0711380004882813, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 15.722702278083268, | |
| "learning_rate": 0.00041666666666666664, | |
| "loss": 2.9737503051757814, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 15.743650170201624, | |
| "learning_rate": 0.0004163891663583843, | |
| "loss": 3.030619812011719, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 15.76459806231998, | |
| "learning_rate": 0.00041611221975883396, | |
| "loss": 3.0626684188842774, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 15.785545954438335, | |
| "learning_rate": 0.00041583582502906203, | |
| "loss": 2.9612255096435547, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 15.806493846556691, | |
| "learning_rate": 0.0004155599803386543, | |
| "loss": 2.9540287017822267, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 15.827441738675045, | |
| "learning_rate": 0.0004152846838656846, | |
| "loss": 2.9890960693359374, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 15.8483896307934, | |
| "learning_rate": 0.00041500993379666443, | |
| "loss": 2.998134803771973, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 15.869337522911756, | |
| "learning_rate": 0.0004147357283264927, | |
| "loss": 3.046440315246582, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 15.890285415030112, | |
| "learning_rate": 0.000414462065658406, | |
| "loss": 2.9803043365478517, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 15.911233307148468, | |
| "learning_rate": 0.0004141889440039292, | |
| "loss": 3.0623497009277343, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 15.932181199266823, | |
| "learning_rate": 0.00041391636158282614, | |
| "loss": 2.926837921142578, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 15.953129091385179, | |
| "learning_rate": 0.00041364431662305114, | |
| "loss": 2.932399368286133, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 15.974076983503535, | |
| "learning_rate": 0.0004133728073607005, | |
| "loss": 2.996663284301758, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 15.99502487562189, | |
| "learning_rate": 0.00041310183203996446, | |
| "loss": 3.0678241729736326, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 16.016758313694684, | |
| "learning_rate": 0.0004128313889130795, | |
| "loss": 3.262166213989258, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 16.03770620581304, | |
| "learning_rate": 0.0004125614762402809, | |
| "loss": 3.072698402404785, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 16.058654097931395, | |
| "learning_rate": 0.00041229209228975627, | |
| "loss": 2.960147476196289, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 16.079601990049753, | |
| "learning_rate": 0.000412023235337598, | |
| "loss": 2.987987518310547, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 16.100549882168107, | |
| "learning_rate": 0.00041175490366775766, | |
| "loss": 2.9958822250366213, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 16.121497774286464, | |
| "learning_rate": 0.0004114870955719997, | |
| "loss": 3.0043949127197265, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 16.142445666404818, | |
| "learning_rate": 0.00041121980934985563, | |
| "loss": 3.013554573059082, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 16.163393558523175, | |
| "learning_rate": 0.000410953043308579, | |
| "loss": 3.0146947860717774, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 16.18434145064153, | |
| "learning_rate": 0.0004106867957631001, | |
| "loss": 2.9756107330322266, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 16.205289342759883, | |
| "learning_rate": 0.00041042106503598165, | |
| "loss": 2.998594284057617, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 16.22623723487824, | |
| "learning_rate": 0.0004101558494573738, | |
| "loss": 2.968126678466797, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 16.247185126996595, | |
| "learning_rate": 0.0004098911473649706, | |
| "loss": 2.938851737976074, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 16.268133019114952, | |
| "learning_rate": 0.0004096269571039658, | |
| "loss": 3.0778596878051756, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 16.289080911233306, | |
| "learning_rate": 0.00040936327702701005, | |
| "loss": 2.871398162841797, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 16.310028803351663, | |
| "learning_rate": 0.00040910010549416687, | |
| "loss": 2.9972572326660156, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 16.330976695470017, | |
| "learning_rate": 0.0004088374408728706, | |
| "loss": 3.1002374649047852, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 16.351924587588375, | |
| "learning_rate": 0.0004085752815378834, | |
| "loss": 2.9723093032836916, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 16.37287247970673, | |
| "learning_rate": 0.0004083136258712532, | |
| "loss": 3.0089197158813477, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 16.393820371825086, | |
| "learning_rate": 0.0004080524722622717, | |
| "loss": 2.9960916519165037, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 16.41476826394344, | |
| "learning_rate": 0.00040779181910743294, | |
| "loss": 2.948496437072754, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 16.435716156061797, | |
| "learning_rate": 0.0004075316648103914, | |
| "loss": 2.986690139770508, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 16.45666404818015, | |
| "learning_rate": 0.0004072720077819216, | |
| "loss": 3.064560317993164, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 16.47761194029851, | |
| "learning_rate": 0.0004070128464398768, | |
| "loss": 3.0366847991943358, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 16.498559832416863, | |
| "learning_rate": 0.0004067541792091489, | |
| "loss": 3.0212535858154297, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 16.51950772453522, | |
| "learning_rate": 0.0004064960045216279, | |
| "loss": 2.966229057312012, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 16.540455616653574, | |
| "learning_rate": 0.0004062383208161624, | |
| "loss": 2.9718713760375977, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 16.56140350877193, | |
| "learning_rate": 0.0004059811265385193, | |
| "loss": 2.938900947570801, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 16.582351400890285, | |
| "learning_rate": 0.00040572442014134516, | |
| "loss": 3.0135732650756837, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 16.60329929300864, | |
| "learning_rate": 0.00040546820008412654, | |
| "loss": 3.016792869567871, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 16.624247185126997, | |
| "learning_rate": 0.0004052124648331515, | |
| "loss": 2.961100387573242, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 16.64519507724535, | |
| "learning_rate": 0.00040495721286147086, | |
| "loss": 2.9855838775634767, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 16.666142969363708, | |
| "learning_rate": 0.00040470244264886006, | |
| "loss": 3.0260711669921876, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 16.687090861482062, | |
| "learning_rate": 0.00040444815268178097, | |
| "loss": 2.9670747756958007, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 16.70803875360042, | |
| "learning_rate": 0.00040419434145334414, | |
| "loss": 2.9868255615234376, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 16.728986645718773, | |
| "learning_rate": 0.00040394100746327154, | |
| "loss": 2.993141746520996, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 16.74993453783713, | |
| "learning_rate": 0.0004036881492178589, | |
| "loss": 2.991754722595215, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 16.770882429955485, | |
| "learning_rate": 0.00040343576522993926, | |
| "loss": 3.0531938552856444, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 16.791830322073842, | |
| "learning_rate": 0.00040318385401884554, | |
| "loss": 2.8399303436279295, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 16.812778214192196, | |
| "learning_rate": 0.00040293241411037484, | |
| "loss": 3.0251434326171873, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 16.833726106310554, | |
| "learning_rate": 0.00040268144403675154, | |
| "loss": 2.9438486099243164, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 16.854673998428908, | |
| "learning_rate": 0.0004024309423365915, | |
| "loss": 2.9782060623168944, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 16.875621890547265, | |
| "learning_rate": 0.0004021809075548668, | |
| "loss": 2.972634696960449, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 16.89656978266562, | |
| "learning_rate": 0.0004019313382428694, | |
| "loss": 2.979868507385254, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 16.917517674783976, | |
| "learning_rate": 0.00040168223295817656, | |
| "loss": 2.990520477294922, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 16.93846556690233, | |
| "learning_rate": 0.00040143359026461554, | |
| "loss": 3.0180835723876953, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 16.959413459020688, | |
| "learning_rate": 0.000401185408732229, | |
| "loss": 3.0260868072509766, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 16.98036135113904, | |
| "learning_rate": 0.0004009376869372401, | |
| "loss": 2.9574857711791993, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 17.002094789211835, | |
| "learning_rate": 0.00040069042346201864, | |
| "loss": 3.0914968490600585, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 17.023042681330192, | |
| "learning_rate": 0.00040044361689504655, | |
| "loss": 3.062566947937012, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 17.043990573448546, | |
| "learning_rate": 0.0004001972658308847, | |
| "loss": 2.976962661743164, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 17.064938465566904, | |
| "learning_rate": 0.0003999513688701383, | |
| "loss": 2.9750572204589845, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 17.085886357685258, | |
| "learning_rate": 0.00039970592461942457, | |
| "loss": 2.987382698059082, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 17.106834249803615, | |
| "learning_rate": 0.00039946093169133874, | |
| "loss": 3.001695442199707, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 17.12778214192197, | |
| "learning_rate": 0.0003992163887044217, | |
| "loss": 2.961598777770996, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 17.148730034040323, | |
| "learning_rate": 0.0003989722942831268, | |
| "loss": 2.957429313659668, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 17.16967792615868, | |
| "learning_rate": 0.0003987286470577879, | |
| "loss": 3.0364120483398436, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 17.190625818277034, | |
| "learning_rate": 0.0003984854456645864, | |
| "loss": 3.0673593521118163, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 17.211573710395392, | |
| "learning_rate": 0.0003982426887455199, | |
| "loss": 2.9361265182495115, | |
| "step": 8210 | |
| }, | |
| { | |
| "epoch": 17.232521602513746, | |
| "learning_rate": 0.00039800037494836985, | |
| "loss": 3.001542854309082, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 17.253469494632103, | |
| "learning_rate": 0.00039775850292667005, | |
| "loss": 2.95641975402832, | |
| "step": 8230 | |
| }, | |
| { | |
| "epoch": 17.274417386750457, | |
| "learning_rate": 0.0003975170713396753, | |
| "loss": 2.8832208633422853, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 17.295365278868815, | |
| "learning_rate": 0.0003972760788523301, | |
| "loss": 2.9211734771728515, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 17.31631317098717, | |
| "learning_rate": 0.0003970355241352378, | |
| "loss": 2.9908830642700197, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 17.337261063105526, | |
| "learning_rate": 0.00039679540586462953, | |
| "loss": 2.991852951049805, | |
| "step": 8270 | |
| }, | |
| { | |
| "epoch": 17.35820895522388, | |
| "learning_rate": 0.00039655572272233384, | |
| "loss": 3.0602521896362305, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 17.379156847342237, | |
| "learning_rate": 0.0003963164733957462, | |
| "loss": 2.975466728210449, | |
| "step": 8290 | |
| }, | |
| { | |
| "epoch": 17.40010473946059, | |
| "learning_rate": 0.00039607765657779864, | |
| "loss": 3.0946418762207033, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 17.42105263157895, | |
| "learning_rate": 0.0003958392709669304, | |
| "loss": 2.993026924133301, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 17.442000523697303, | |
| "learning_rate": 0.00039560131526705723, | |
| "loss": 3.0121936798095703, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 17.46294841581566, | |
| "learning_rate": 0.0003953637881875425, | |
| "loss": 3.0414730072021485, | |
| "step": 8330 | |
| }, | |
| { | |
| "epoch": 17.483896307934014, | |
| "learning_rate": 0.0003951266884431675, | |
| "loss": 3.0235416412353517, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 17.50484420005237, | |
| "learning_rate": 0.00039489001475410214, | |
| "loss": 2.9818603515625, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 17.525792092170725, | |
| "learning_rate": 0.00039465376584587626, | |
| "loss": 2.994624137878418, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 17.54673998428908, | |
| "learning_rate": 0.00039441794044935054, | |
| "loss": 3.0084808349609373, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 17.567687876407437, | |
| "learning_rate": 0.00039418253730068797, | |
| "loss": 2.8967424392700196, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 17.58863576852579, | |
| "learning_rate": 0.0003939475551413253, | |
| "loss": 2.9253704071044924, | |
| "step": 8390 | |
| }, | |
| { | |
| "epoch": 17.609583660644148, | |
| "learning_rate": 0.000393712992717945, | |
| "loss": 3.0752674102783204, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 17.630531552762502, | |
| "learning_rate": 0.0003934788487824469, | |
| "loss": 2.925820159912109, | |
| "step": 8410 | |
| }, | |
| { | |
| "epoch": 17.65147944488086, | |
| "learning_rate": 0.0003932451220919205, | |
| "loss": 2.9983007431030275, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 17.672427336999213, | |
| "learning_rate": 0.0003930118114086172, | |
| "loss": 2.884238624572754, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 17.69337522911757, | |
| "learning_rate": 0.00039277891549992266, | |
| "loss": 2.854781723022461, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 17.714323121235925, | |
| "learning_rate": 0.0003925464331383298, | |
| "loss": 2.9886890411376954, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 17.735271013354282, | |
| "learning_rate": 0.00039231436310141113, | |
| "loss": 3.0236677169799804, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 17.756218905472636, | |
| "learning_rate": 0.00039208270417179214, | |
| "loss": 2.951685905456543, | |
| "step": 8470 | |
| }, | |
| { | |
| "epoch": 17.777166797590993, | |
| "learning_rate": 0.0003918514551371243, | |
| "loss": 2.971786880493164, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 17.798114689709347, | |
| "learning_rate": 0.0003916206147900585, | |
| "loss": 2.983307647705078, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 17.819062581827705, | |
| "learning_rate": 0.00039139018192821845, | |
| "loss": 2.9782459259033205, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 17.84001047394606, | |
| "learning_rate": 0.00039116015535417445, | |
| "loss": 3.02642765045166, | |
| "step": 8510 | |
| }, | |
| { | |
| "epoch": 17.860958366064416, | |
| "learning_rate": 0.00039093053387541745, | |
| "loss": 3.011845588684082, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 17.88190625818277, | |
| "learning_rate": 0.00039070131630433274, | |
| "loss": 2.939919090270996, | |
| "step": 8530 | |
| }, | |
| { | |
| "epoch": 17.902854150301128, | |
| "learning_rate": 0.00039047250145817424, | |
| "loss": 2.996026039123535, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 17.92380204241948, | |
| "learning_rate": 0.00039024408815903914, | |
| "loss": 2.903793716430664, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 17.94474993453784, | |
| "learning_rate": 0.0003900160752338421, | |
| "loss": 2.9599498748779296, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 17.965697826656193, | |
| "learning_rate": 0.00038978846151429, | |
| "loss": 2.841645050048828, | |
| "step": 8570 | |
| }, | |
| { | |
| "epoch": 17.986645718774547, | |
| "learning_rate": 0.0003895612458368572, | |
| "loss": 2.885163497924805, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 18.008379156847344, | |
| "learning_rate": 0.00038933442704275974, | |
| "loss": 2.9911325454711912, | |
| "step": 8590 | |
| }, | |
| { | |
| "epoch": 18.029327048965698, | |
| "learning_rate": 0.0003891080039779314, | |
| "loss": 3.027914619445801, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 18.050274941084055, | |
| "learning_rate": 0.0003888819754929986, | |
| "loss": 2.9694196701049806, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 18.07122283320241, | |
| "learning_rate": 0.0003886563404432558, | |
| "loss": 2.9900096893310546, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 18.092170725320763, | |
| "learning_rate": 0.0003884310976886414, | |
| "loss": 2.987308692932129, | |
| "step": 8630 | |
| }, | |
| { | |
| "epoch": 18.11311861743912, | |
| "learning_rate": 0.0003882062460937135, | |
| "loss": 2.934325408935547, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 18.134066509557474, | |
| "learning_rate": 0.0003879817845276255, | |
| "loss": 2.9511764526367186, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 18.15501440167583, | |
| "learning_rate": 0.0003877577118641029, | |
| "loss": 2.9479455947875977, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 18.175962293794186, | |
| "learning_rate": 0.00038753402698141903, | |
| "loss": 3.0447383880615235, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 18.196910185912543, | |
| "learning_rate": 0.0003873107287623715, | |
| "loss": 2.917817497253418, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 18.217858078030897, | |
| "learning_rate": 0.00038708781609425905, | |
| "loss": 2.8964914321899413, | |
| "step": 8690 | |
| }, | |
| { | |
| "epoch": 18.238805970149254, | |
| "learning_rate": 0.000386865287868858, | |
| "loss": 2.967067527770996, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 18.25975386226761, | |
| "learning_rate": 0.0003866431429823993, | |
| "loss": 2.991856002807617, | |
| "step": 8710 | |
| }, | |
| { | |
| "epoch": 18.280701754385966, | |
| "learning_rate": 0.00038642138033554525, | |
| "loss": 2.96053524017334, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 18.30164964650432, | |
| "learning_rate": 0.00038619999883336703, | |
| "loss": 2.9373369216918945, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 18.322597538622677, | |
| "learning_rate": 0.0003859789973853217, | |
| "loss": 2.9254953384399416, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 18.34354543074103, | |
| "learning_rate": 0.0003857583749052298, | |
| "loss": 3.008597564697266, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 18.36449332285939, | |
| "learning_rate": 0.0003855381303112527, | |
| "loss": 2.9206886291503906, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 18.385441214977742, | |
| "learning_rate": 0.0003853182625258708, | |
| "loss": 2.9910358428955077, | |
| "step": 8770 | |
| }, | |
| { | |
| "epoch": 18.4063891070961, | |
| "learning_rate": 0.0003850987704758608, | |
| "loss": 2.894259452819824, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 18.427336999214454, | |
| "learning_rate": 0.00038487965309227413, | |
| "loss": 3.0239398956298826, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 18.44828489133281, | |
| "learning_rate": 0.0003846609093104148, | |
| "loss": 3.023584747314453, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 18.469232783451165, | |
| "learning_rate": 0.00038444253806981784, | |
| "loss": 2.9302574157714845, | |
| "step": 8810 | |
| }, | |
| { | |
| "epoch": 18.49018067556952, | |
| "learning_rate": 0.00038422453831422784, | |
| "loss": 2.968699264526367, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 18.511128567687877, | |
| "learning_rate": 0.0003840069089915771, | |
| "loss": 2.904973793029785, | |
| "step": 8830 | |
| }, | |
| { | |
| "epoch": 18.53207645980623, | |
| "learning_rate": 0.00038378964905396454, | |
| "loss": 2.9212614059448243, | |
| "step": 8840 | |
| }, | |
| { | |
| "epoch": 18.553024351924588, | |
| "learning_rate": 0.00038357275745763475, | |
| "loss": 2.9412769317626952, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 18.573972244042942, | |
| "learning_rate": 0.0003833562331629563, | |
| "loss": 2.926407814025879, | |
| "step": 8860 | |
| }, | |
| { | |
| "epoch": 18.5949201361613, | |
| "learning_rate": 0.0003831400751344014, | |
| "loss": 2.87393741607666, | |
| "step": 8870 | |
| }, | |
| { | |
| "epoch": 18.615868028279653, | |
| "learning_rate": 0.00038292428234052486, | |
| "loss": 2.93045597076416, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 18.63681592039801, | |
| "learning_rate": 0.0003827088537539434, | |
| "loss": 2.8498041152954103, | |
| "step": 8890 | |
| }, | |
| { | |
| "epoch": 18.657763812516365, | |
| "learning_rate": 0.00038249378835131535, | |
| "loss": 2.980220603942871, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 18.678711704634722, | |
| "learning_rate": 0.0003822790851133196, | |
| "loss": 2.9403018951416016, | |
| "step": 8910 | |
| }, | |
| { | |
| "epoch": 18.699659596753076, | |
| "learning_rate": 0.00038206474302463617, | |
| "loss": 2.932261848449707, | |
| "step": 8920 | |
| }, | |
| { | |
| "epoch": 18.720607488871433, | |
| "learning_rate": 0.00038185076107392544, | |
| "loss": 3.0572792053222657, | |
| "step": 8930 | |
| }, | |
| { | |
| "epoch": 18.741555380989787, | |
| "learning_rate": 0.0003816371382538082, | |
| "loss": 2.8232454299926757, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 18.762503273108145, | |
| "learning_rate": 0.0003814238735608459, | |
| "loss": 2.9247211456298827, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 18.7834511652265, | |
| "learning_rate": 0.0003812109659955207, | |
| "loss": 2.9309356689453123, | |
| "step": 8960 | |
| }, | |
| { | |
| "epoch": 18.804399057344856, | |
| "learning_rate": 0.00038099841456221617, | |
| "loss": 2.9406196594238283, | |
| "step": 8970 | |
| }, | |
| { | |
| "epoch": 18.82534694946321, | |
| "learning_rate": 0.0003807862182691969, | |
| "loss": 2.9863868713378907, | |
| "step": 8980 | |
| }, | |
| { | |
| "epoch": 18.846294841581567, | |
| "learning_rate": 0.00038057437612859003, | |
| "loss": 2.9020156860351562, | |
| "step": 8990 | |
| }, | |
| { | |
| "epoch": 18.86724273369992, | |
| "learning_rate": 0.0003803628871563653, | |
| "loss": 2.8909185409545897, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 18.888190625818275, | |
| "learning_rate": 0.0003801517503723161, | |
| "loss": 2.905278205871582, | |
| "step": 9010 | |
| }, | |
| { | |
| "epoch": 18.909138517936633, | |
| "learning_rate": 0.00037994096480004037, | |
| "loss": 2.971329689025879, | |
| "step": 9020 | |
| }, | |
| { | |
| "epoch": 18.930086410054987, | |
| "learning_rate": 0.0003797305294669214, | |
| "loss": 2.9014846801757814, | |
| "step": 9030 | |
| }, | |
| { | |
| "epoch": 18.951034302173344, | |
| "learning_rate": 0.00037952044340410954, | |
| "loss": 2.919228363037109, | |
| "step": 9040 | |
| }, | |
| { | |
| "epoch": 18.971982194291698, | |
| "learning_rate": 0.00037931070564650276, | |
| "loss": 2.9611515045166015, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 18.992930086410055, | |
| "learning_rate": 0.0003791013152327286, | |
| "loss": 3.06106014251709, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 19.01466352448285, | |
| "learning_rate": 0.00037889227120512545, | |
| "loss": 3.0117502212524414, | |
| "step": 9070 | |
| }, | |
| { | |
| "epoch": 19.035611416601206, | |
| "learning_rate": 0.0003786835726097239, | |
| "loss": 2.908797836303711, | |
| "step": 9080 | |
| }, | |
| { | |
| "epoch": 19.05655930871956, | |
| "learning_rate": 0.00037847521849622895, | |
| "loss": 2.977317047119141, | |
| "step": 9090 | |
| }, | |
| { | |
| "epoch": 19.077507200837914, | |
| "learning_rate": 0.0003782672079180015, | |
| "loss": 2.982106018066406, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 19.09845509295627, | |
| "learning_rate": 0.0003780595399320404, | |
| "loss": 2.9336454391479494, | |
| "step": 9110 | |
| }, | |
| { | |
| "epoch": 19.119402985074625, | |
| "learning_rate": 0.00037785221359896444, | |
| "loss": 2.9511249542236326, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 19.140350877192983, | |
| "learning_rate": 0.00037764522798299443, | |
| "loss": 2.9214210510253906, | |
| "step": 9130 | |
| }, | |
| { | |
| "epoch": 19.161298769311337, | |
| "learning_rate": 0.0003774385821519358, | |
| "loss": 2.9141027450561525, | |
| "step": 9140 | |
| }, | |
| { | |
| "epoch": 19.182246661429694, | |
| "learning_rate": 0.0003772322751771605, | |
| "loss": 3.0709518432617187, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 19.203194553548048, | |
| "learning_rate": 0.00037702630613358986, | |
| "loss": 2.9491186141967773, | |
| "step": 9160 | |
| }, | |
| { | |
| "epoch": 19.224142445666406, | |
| "learning_rate": 0.0003768206740996769, | |
| "loss": 2.9979949951171876, | |
| "step": 9170 | |
| }, | |
| { | |
| "epoch": 19.24509033778476, | |
| "learning_rate": 0.00037661537815738915, | |
| "loss": 2.9300355911254883, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 19.266038229903117, | |
| "learning_rate": 0.00037641041739219143, | |
| "loss": 2.95788631439209, | |
| "step": 9190 | |
| }, | |
| { | |
| "epoch": 19.28698612202147, | |
| "learning_rate": 0.00037620579089302876, | |
| "loss": 2.960785675048828, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 19.30793401413983, | |
| "learning_rate": 0.0003760014977523091, | |
| "loss": 2.9440664291381835, | |
| "step": 9210 | |
| }, | |
| { | |
| "epoch": 19.328881906258182, | |
| "learning_rate": 0.00037579753706588697, | |
| "loss": 2.9864282608032227, | |
| "step": 9220 | |
| }, | |
| { | |
| "epoch": 19.34982979837654, | |
| "learning_rate": 0.00037559390793304604, | |
| "loss": 2.830784225463867, | |
| "step": 9230 | |
| }, | |
| { | |
| "epoch": 19.370777690494894, | |
| "learning_rate": 0.00037539060945648286, | |
| "loss": 2.981433868408203, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 19.39172558261325, | |
| "learning_rate": 0.00037518764074229014, | |
| "loss": 2.8778303146362303, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 19.412673474731605, | |
| "learning_rate": 0.00037498500089994, | |
| "loss": 2.896072006225586, | |
| "step": 9260 | |
| }, | |
| { | |
| "epoch": 19.433621366849962, | |
| "learning_rate": 0.00037478268904226795, | |
| "loss": 2.962319564819336, | |
| "step": 9270 | |
| }, | |
| { | |
| "epoch": 19.454569258968316, | |
| "learning_rate": 0.00037458070428545635, | |
| "loss": 2.9110170364379884, | |
| "step": 9280 | |
| }, | |
| { | |
| "epoch": 19.47551715108667, | |
| "learning_rate": 0.00037437904574901817, | |
| "loss": 2.928105926513672, | |
| "step": 9290 | |
| }, | |
| { | |
| "epoch": 19.496465043205028, | |
| "learning_rate": 0.00037417771255578104, | |
| "loss": 2.91275577545166, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 19.51741293532338, | |
| "learning_rate": 0.00037397670383187097, | |
| "loss": 2.964938163757324, | |
| "step": 9310 | |
| }, | |
| { | |
| "epoch": 19.53836082744174, | |
| "learning_rate": 0.0003737760187066967, | |
| "loss": 2.953006935119629, | |
| "step": 9320 | |
| }, | |
| { | |
| "epoch": 19.559308719560093, | |
| "learning_rate": 0.00037357565631293365, | |
| "loss": 3.108722686767578, | |
| "step": 9330 | |
| }, | |
| { | |
| "epoch": 19.58025661167845, | |
| "learning_rate": 0.00037337561578650833, | |
| "loss": 3.078016471862793, | |
| "step": 9340 | |
| }, | |
| { | |
| "epoch": 19.601204503796804, | |
| "learning_rate": 0.00037317589626658255, | |
| "loss": 3.006759262084961, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 19.622152395915162, | |
| "learning_rate": 0.0003729764968955379, | |
| "loss": 2.9581228256225587, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 19.643100288033516, | |
| "learning_rate": 0.00037277741681896045, | |
| "loss": 3.199405860900879, | |
| "step": 9370 | |
| }, | |
| { | |
| "epoch": 19.664048180151873, | |
| "learning_rate": 0.0003725786551856251, | |
| "loss": 3.039951133728027, | |
| "step": 9380 | |
| }, | |
| { | |
| "epoch": 19.684996072270227, | |
| "learning_rate": 0.0003723802111474804, | |
| "loss": 2.99322566986084, | |
| "step": 9390 | |
| }, | |
| { | |
| "epoch": 19.705943964388585, | |
| "learning_rate": 0.0003721820838596335, | |
| "loss": 2.9919605255126953, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 19.72689185650694, | |
| "learning_rate": 0.00037198427248033485, | |
| "loss": 2.9185922622680662, | |
| "step": 9410 | |
| }, | |
| { | |
| "epoch": 19.747839748625296, | |
| "learning_rate": 0.00037178677617096337, | |
| "loss": 2.927593994140625, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 19.76878764074365, | |
| "learning_rate": 0.0003715895940960111, | |
| "loss": 2.9274175643920897, | |
| "step": 9430 | |
| }, | |
| { | |
| "epoch": 19.789735532862007, | |
| "learning_rate": 0.000371392725423069, | |
| "loss": 2.9191694259643555, | |
| "step": 9440 | |
| }, | |
| { | |
| "epoch": 19.81068342498036, | |
| "learning_rate": 0.00037119616932281165, | |
| "loss": 3.002157974243164, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 19.83163131709872, | |
| "learning_rate": 0.00037099992496898276, | |
| "loss": 2.9765802383422852, | |
| "step": 9460 | |
| }, | |
| { | |
| "epoch": 19.852579209217073, | |
| "learning_rate": 0.00037080399153838065, | |
| "loss": 2.983877182006836, | |
| "step": 9470 | |
| }, | |
| { | |
| "epoch": 19.873527101335426, | |
| "learning_rate": 0.00037060836821084373, | |
| "loss": 2.9292572021484373, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 19.894474993453784, | |
| "learning_rate": 0.00037041305416923604, | |
| "loss": 2.9650571823120115, | |
| "step": 9490 | |
| }, | |
| { | |
| "epoch": 19.915422885572138, | |
| "learning_rate": 0.0003702180485994327, | |
| "loss": 2.9420921325683596, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 19.936370777690495, | |
| "learning_rate": 0.00037002335069030614, | |
| "loss": 2.8849225997924806, | |
| "step": 9510 | |
| }, | |
| { | |
| "epoch": 19.95731866980885, | |
| "learning_rate": 0.0003698289596337116, | |
| "loss": 2.882498550415039, | |
| "step": 9520 | |
| }, | |
| { | |
| "epoch": 19.978266561927207, | |
| "learning_rate": 0.00036963487462447303, | |
| "loss": 2.9787607192993164, | |
| "step": 9530 | |
| }, | |
| { | |
| "epoch": 19.99921445404556, | |
| "learning_rate": 0.0003694410948603691, | |
| "loss": 2.93823299407959, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 20.020947892118354, | |
| "learning_rate": 0.00036924761954211944, | |
| "loss": 3.033322334289551, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 20.04189578423671, | |
| "learning_rate": 0.0003690544478733707, | |
| "loss": 2.9187992095947264, | |
| "step": 9560 | |
| }, | |
| { | |
| "epoch": 20.062843676355065, | |
| "learning_rate": 0.0003688615790606828, | |
| "loss": 2.8914941787719726, | |
| "step": 9570 | |
| }, | |
| { | |
| "epoch": 20.083791568473423, | |
| "learning_rate": 0.000368669012313515, | |
| "loss": 2.8893537521362305, | |
| "step": 9580 | |
| }, | |
| { | |
| "epoch": 20.104739460591777, | |
| "learning_rate": 0.0003684767468442126, | |
| "loss": 2.906744384765625, | |
| "step": 9590 | |
| }, | |
| { | |
| "epoch": 20.125687352710134, | |
| "learning_rate": 0.0003682847818679935, | |
| "loss": 2.8719203948974608, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 20.146635244828488, | |
| "learning_rate": 0.0003680931166029342, | |
| "loss": 2.956478500366211, | |
| "step": 9610 | |
| }, | |
| { | |
| "epoch": 20.167583136946845, | |
| "learning_rate": 0.000367901750269957, | |
| "loss": 3.0235532760620116, | |
| "step": 9620 | |
| }, | |
| { | |
| "epoch": 20.1885310290652, | |
| "learning_rate": 0.00036771068209281657, | |
| "loss": 2.9067535400390625, | |
| "step": 9630 | |
| }, | |
| { | |
| "epoch": 20.209478921183557, | |
| "learning_rate": 0.0003675199112980863, | |
| "loss": 3.0424097061157225, | |
| "step": 9640 | |
| }, | |
| { | |
| "epoch": 20.23042681330191, | |
| "learning_rate": 0.0003673294371151458, | |
| "loss": 2.873898506164551, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 20.251374705420268, | |
| "learning_rate": 0.0003671392587761674, | |
| "loss": 2.9879985809326173, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 20.272322597538622, | |
| "learning_rate": 0.0003669493755161031, | |
| "loss": 2.9691110610961915, | |
| "step": 9670 | |
| }, | |
| { | |
| "epoch": 20.29327048965698, | |
| "learning_rate": 0.00036675978657267204, | |
| "loss": 2.8997966766357424, | |
| "step": 9680 | |
| }, | |
| { | |
| "epoch": 20.314218381775333, | |
| "learning_rate": 0.00036657049118634733, | |
| "loss": 3.0071743011474608, | |
| "step": 9690 | |
| }, | |
| { | |
| "epoch": 20.33516627389369, | |
| "learning_rate": 0.0003663814886003432, | |
| "loss": 2.9313344955444336, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 20.356114166012045, | |
| "learning_rate": 0.00036619277806060276, | |
| "loss": 2.9554468154907227, | |
| "step": 9710 | |
| }, | |
| { | |
| "epoch": 20.377062058130402, | |
| "learning_rate": 0.0003660043588157846, | |
| "loss": 2.9395862579345704, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 20.398009950248756, | |
| "learning_rate": 0.00036581623011725114, | |
| "loss": 2.8985124588012696, | |
| "step": 9730 | |
| }, | |
| { | |
| "epoch": 20.41895784236711, | |
| "learning_rate": 0.0003656283912190554, | |
| "loss": 2.92258243560791, | |
| "step": 9740 | |
| }, | |
| { | |
| "epoch": 20.439905734485468, | |
| "learning_rate": 0.00036544084137792883, | |
| "loss": 2.9244316101074217, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 20.46085362660382, | |
| "learning_rate": 0.00036525357985326903, | |
| "loss": 2.9558000564575195, | |
| "step": 9760 | |
| }, | |
| { | |
| "epoch": 20.48180151872218, | |
| "learning_rate": 0.0003650666059071275, | |
| "loss": 2.84803466796875, | |
| "step": 9770 | |
| }, | |
| { | |
| "epoch": 20.502749410840533, | |
| "learning_rate": 0.00036487991880419725, | |
| "loss": 2.9493398666381836, | |
| "step": 9780 | |
| }, | |
| { | |
| "epoch": 20.52369730295889, | |
| "learning_rate": 0.00036469351781180073, | |
| "loss": 2.925904083251953, | |
| "step": 9790 | |
| }, | |
| { | |
| "epoch": 20.544645195077244, | |
| "learning_rate": 0.00036450740219987765, | |
| "loss": 2.8513689041137695, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 20.5655930871956, | |
| "learning_rate": 0.0003643215712409734, | |
| "loss": 2.9981321334838866, | |
| "step": 9810 | |
| }, | |
| { | |
| "epoch": 20.586540979313956, | |
| "learning_rate": 0.00036413602421022653, | |
| "loss": 2.8937055587768556, | |
| "step": 9820 | |
| }, | |
| { | |
| "epoch": 20.607488871432313, | |
| "learning_rate": 0.0003639507603853572, | |
| "loss": 3.000600814819336, | |
| "step": 9830 | |
| }, | |
| { | |
| "epoch": 20.628436763550667, | |
| "learning_rate": 0.00036376577904665525, | |
| "loss": 2.8795480728149414, | |
| "step": 9840 | |
| }, | |
| { | |
| "epoch": 20.649384655669024, | |
| "learning_rate": 0.00036358107947696876, | |
| "loss": 2.8654504776000977, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 20.67033254778738, | |
| "learning_rate": 0.0003633966609616919, | |
| "loss": 2.9193105697631836, | |
| "step": 9860 | |
| }, | |
| { | |
| "epoch": 20.691280439905736, | |
| "learning_rate": 0.00036321252278875344, | |
| "loss": 3.000343894958496, | |
| "step": 9870 | |
| }, | |
| { | |
| "epoch": 20.71222833202409, | |
| "learning_rate": 0.00036302866424860566, | |
| "loss": 2.9843284606933596, | |
| "step": 9880 | |
| }, | |
| { | |
| "epoch": 20.733176224142447, | |
| "learning_rate": 0.00036284508463421217, | |
| "loss": 2.9358942031860353, | |
| "step": 9890 | |
| }, | |
| { | |
| "epoch": 20.7541241162608, | |
| "learning_rate": 0.0003626617832410371, | |
| "loss": 2.9408638000488283, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 20.77507200837916, | |
| "learning_rate": 0.00036247875936703335, | |
| "loss": 2.9035417556762697, | |
| "step": 9910 | |
| }, | |
| { | |
| "epoch": 20.796019900497512, | |
| "learning_rate": 0.00036229601231263145, | |
| "loss": 2.8692134857177733, | |
| "step": 9920 | |
| }, | |
| { | |
| "epoch": 20.81696779261587, | |
| "learning_rate": 0.0003621135413807282, | |
| "loss": 2.8154270172119142, | |
| "step": 9930 | |
| }, | |
| { | |
| "epoch": 20.837915684734224, | |
| "learning_rate": 0.0003619313458766758, | |
| "loss": 2.905722427368164, | |
| "step": 9940 | |
| }, | |
| { | |
| "epoch": 20.858863576852578, | |
| "learning_rate": 0.0003617494251082704, | |
| "loss": 2.9300207138061523, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 20.879811468970935, | |
| "learning_rate": 0.0003615677783857413, | |
| "loss": 2.9284923553466795, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 20.90075936108929, | |
| "learning_rate": 0.0003613864050217397, | |
| "loss": 2.8830732345581054, | |
| "step": 9970 | |
| }, | |
| { | |
| "epoch": 20.921707253207646, | |
| "learning_rate": 0.0003612053043313283, | |
| "loss": 2.922040557861328, | |
| "step": 9980 | |
| }, | |
| { | |
| "epoch": 20.942655145326, | |
| "learning_rate": 0.0003610244756319697, | |
| "loss": 2.919090461730957, | |
| "step": 9990 | |
| }, | |
| { | |
| "epoch": 20.963603037444358, | |
| "learning_rate": 0.00036084391824351607, | |
| "loss": 2.9119571685791015, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 20.98455092956271, | |
| "learning_rate": 0.00036066363148819854, | |
| "loss": 3.0065305709838865, | |
| "step": 10010 | |
| }, | |
| { | |
| "epoch": 21.006284367635505, | |
| "learning_rate": 0.000360483614690616, | |
| "loss": 3.088846206665039, | |
| "step": 10020 | |
| }, | |
| { | |
| "epoch": 21.027232259753863, | |
| "learning_rate": 0.00036030386717772494, | |
| "loss": 2.887124443054199, | |
| "step": 10030 | |
| }, | |
| { | |
| "epoch": 21.048180151872216, | |
| "learning_rate": 0.0003601243882788286, | |
| "loss": 2.9657873153686523, | |
| "step": 10040 | |
| }, | |
| { | |
| "epoch": 21.069128043990574, | |
| "learning_rate": 0.0003599451773255667, | |
| "loss": 2.8971757888793945, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 21.090075936108928, | |
| "learning_rate": 0.00035976623365190465, | |
| "loss": 2.8915260314941404, | |
| "step": 10060 | |
| }, | |
| { | |
| "epoch": 21.111023828227285, | |
| "learning_rate": 0.0003595875565941235, | |
| "loss": 2.9463220596313477, | |
| "step": 10070 | |
| }, | |
| { | |
| "epoch": 21.13197172034564, | |
| "learning_rate": 0.00035940914549080944, | |
| "loss": 2.9156425476074217, | |
| "step": 10080 | |
| }, | |
| { | |
| "epoch": 21.152919612463997, | |
| "learning_rate": 0.0003592309996828435, | |
| "loss": 2.996392250061035, | |
| "step": 10090 | |
| }, | |
| { | |
| "epoch": 21.17386750458235, | |
| "learning_rate": 0.0003590531185133913, | |
| "loss": 2.899692344665527, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 21.194815396700708, | |
| "learning_rate": 0.0003588755013278929, | |
| "loss": 2.9242908477783205, | |
| "step": 10110 | |
| }, | |
| { | |
| "epoch": 21.215763288819062, | |
| "learning_rate": 0.00035869814747405306, | |
| "loss": 2.846599578857422, | |
| "step": 10120 | |
| }, | |
| { | |
| "epoch": 21.23671118093742, | |
| "learning_rate": 0.00035852105630183027, | |
| "loss": 2.9692026138305665, | |
| "step": 10130 | |
| }, | |
| { | |
| "epoch": 21.257659073055773, | |
| "learning_rate": 0.0003583442271634278, | |
| "loss": 2.921228790283203, | |
| "step": 10140 | |
| }, | |
| { | |
| "epoch": 21.27860696517413, | |
| "learning_rate": 0.000358167659413283, | |
| "loss": 2.906037139892578, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 21.299554857292485, | |
| "learning_rate": 0.00035799135240805765, | |
| "loss": 2.8459890365600584, | |
| "step": 10160 | |
| }, | |
| { | |
| "epoch": 21.320502749410842, | |
| "learning_rate": 0.0003578153055066282, | |
| "loss": 2.839315986633301, | |
| "step": 10170 | |
| }, | |
| { | |
| "epoch": 21.341450641529196, | |
| "learning_rate": 0.00035763951807007597, | |
| "loss": 2.918286895751953, | |
| "step": 10180 | |
| }, | |
| { | |
| "epoch": 21.36239853364755, | |
| "learning_rate": 0.0003574639894616771, | |
| "loss": 2.9180910110473635, | |
| "step": 10190 | |
| }, | |
| { | |
| "epoch": 21.383346425765907, | |
| "learning_rate": 0.0003572887190468934, | |
| "loss": 2.96252498626709, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 21.40429431788426, | |
| "learning_rate": 0.00035711370619336214, | |
| "loss": 2.873885726928711, | |
| "step": 10210 | |
| }, | |
| { | |
| "epoch": 21.42524221000262, | |
| "learning_rate": 0.00035693895027088694, | |
| "loss": 2.8597929000854494, | |
| "step": 10220 | |
| }, | |
| { | |
| "epoch": 21.446190102120973, | |
| "learning_rate": 0.00035676445065142793, | |
| "loss": 2.9235706329345703, | |
| "step": 10230 | |
| }, | |
| { | |
| "epoch": 21.46713799423933, | |
| "learning_rate": 0.0003565902067090925, | |
| "loss": 2.843309783935547, | |
| "step": 10240 | |
| }, | |
| { | |
| "epoch": 21.488085886357684, | |
| "learning_rate": 0.0003564162178201257, | |
| "loss": 2.9598644256591795, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 21.50903377847604, | |
| "learning_rate": 0.0003562424833629007, | |
| "loss": 2.8366893768310546, | |
| "step": 10260 | |
| }, | |
| { | |
| "epoch": 21.529981670594395, | |
| "learning_rate": 0.0003560690027179101, | |
| "loss": 2.927451515197754, | |
| "step": 10270 | |
| }, | |
| { | |
| "epoch": 21.550929562712753, | |
| "learning_rate": 0.00035589577526775603, | |
| "loss": 2.8992708206176756, | |
| "step": 10280 | |
| }, | |
| { | |
| "epoch": 21.571877454831107, | |
| "learning_rate": 0.000355722800397141, | |
| "loss": 2.8801607131958007, | |
| "step": 10290 | |
| }, | |
| { | |
| "epoch": 21.592825346949464, | |
| "learning_rate": 0.00035555007749285897, | |
| "loss": 2.929705047607422, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 21.613773239067818, | |
| "learning_rate": 0.00035537760594378607, | |
| "loss": 2.9493310928344725, | |
| "step": 10310 | |
| }, | |
| { | |
| "epoch": 21.634721131186176, | |
| "learning_rate": 0.00035520538514087155, | |
| "loss": 2.9074274063110352, | |
| "step": 10320 | |
| }, | |
| { | |
| "epoch": 21.65566902330453, | |
| "learning_rate": 0.0003550334144771289, | |
| "loss": 2.8912160873413084, | |
| "step": 10330 | |
| }, | |
| { | |
| "epoch": 21.676616915422887, | |
| "learning_rate": 0.00035486169334762637, | |
| "loss": 2.8537384033203126, | |
| "step": 10340 | |
| }, | |
| { | |
| "epoch": 21.69756480754124, | |
| "learning_rate": 0.00035469022114947857, | |
| "loss": 2.899664878845215, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 21.7185126996596, | |
| "learning_rate": 0.00035451899728183736, | |
| "loss": 2.9585891723632813, | |
| "step": 10360 | |
| }, | |
| { | |
| "epoch": 21.739460591777952, | |
| "learning_rate": 0.00035434802114588305, | |
| "loss": 2.8878129959106444, | |
| "step": 10370 | |
| }, | |
| { | |
| "epoch": 21.760408483896306, | |
| "learning_rate": 0.00035417729214481556, | |
| "loss": 2.9733861923217773, | |
| "step": 10380 | |
| }, | |
| { | |
| "epoch": 21.781356376014664, | |
| "learning_rate": 0.0003540068096838456, | |
| "loss": 2.877838134765625, | |
| "step": 10390 | |
| }, | |
| { | |
| "epoch": 21.802304268133017, | |
| "learning_rate": 0.0003538365731701862, | |
| "loss": 2.874703598022461, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 21.823252160251375, | |
| "learning_rate": 0.0003536665820130437, | |
| "loss": 2.8807294845581053, | |
| "step": 10410 | |
| }, | |
| { | |
| "epoch": 21.84420005236973, | |
| "learning_rate": 0.00035349683562360966, | |
| "loss": 2.9425344467163086, | |
| "step": 10420 | |
| }, | |
| { | |
| "epoch": 21.865147944488086, | |
| "learning_rate": 0.0003533273334150517, | |
| "loss": 2.9000774383544923, | |
| "step": 10430 | |
| }, | |
| { | |
| "epoch": 21.88609583660644, | |
| "learning_rate": 0.0003531580748025054, | |
| "loss": 2.8818313598632814, | |
| "step": 10440 | |
| }, | |
| { | |
| "epoch": 21.907043728724798, | |
| "learning_rate": 0.00035298905920306563, | |
| "loss": 2.9183849334716796, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 21.92799162084315, | |
| "learning_rate": 0.00035282028603577823, | |
| "loss": 2.843509292602539, | |
| "step": 10460 | |
| }, | |
| { | |
| "epoch": 21.94893951296151, | |
| "learning_rate": 0.0003526517547216315, | |
| "loss": 2.9971471786499024, | |
| "step": 10470 | |
| }, | |
| { | |
| "epoch": 21.969887405079863, | |
| "learning_rate": 0.000352483464683548, | |
| "loss": 2.851757621765137, | |
| "step": 10480 | |
| }, | |
| { | |
| "epoch": 21.99083529719822, | |
| "learning_rate": 0.0003523154153463761, | |
| "loss": 2.889949417114258, | |
| "step": 10490 | |
| }, | |
| { | |
| "epoch": 22.012568735271014, | |
| "learning_rate": 0.00035214760613688187, | |
| "loss": 2.9378170013427733, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 22.033516627389368, | |
| "learning_rate": 0.0003519800364837407, | |
| "loss": 2.9217357635498047, | |
| "step": 10510 | |
| }, | |
| { | |
| "epoch": 22.054464519507725, | |
| "learning_rate": 0.0003518127058175293, | |
| "loss": 2.9642236709594725, | |
| "step": 10520 | |
| }, | |
| { | |
| "epoch": 22.07541241162608, | |
| "learning_rate": 0.00035164561357071755, | |
| "loss": 2.9043264389038086, | |
| "step": 10530 | |
| }, | |
| { | |
| "epoch": 22.096360303744436, | |
| "learning_rate": 0.0003514787591776602, | |
| "loss": 2.9355424880981444, | |
| "step": 10540 | |
| }, | |
| { | |
| "epoch": 22.11730819586279, | |
| "learning_rate": 0.0003513121420745892, | |
| "loss": 2.876350975036621, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 22.138256087981148, | |
| "learning_rate": 0.0003511457616996052, | |
| "loss": 2.9557785034179687, | |
| "step": 10560 | |
| }, | |
| { | |
| "epoch": 22.1592039800995, | |
| "learning_rate": 0.0003509796174926703, | |
| "loss": 2.9274904251098635, | |
| "step": 10570 | |
| }, | |
| { | |
| "epoch": 22.18015187221786, | |
| "learning_rate": 0.00035081370889559934, | |
| "loss": 2.873898506164551, | |
| "step": 10580 | |
| }, | |
| { | |
| "epoch": 22.201099764336213, | |
| "learning_rate": 0.0003506480353520526, | |
| "loss": 3.015602684020996, | |
| "step": 10590 | |
| }, | |
| { | |
| "epoch": 22.22204765645457, | |
| "learning_rate": 0.0003504825963075276, | |
| "loss": 2.836939239501953, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 22.242995548572924, | |
| "learning_rate": 0.00035031739120935175, | |
| "loss": 2.9317630767822265, | |
| "step": 10610 | |
| }, | |
| { | |
| "epoch": 22.263943440691282, | |
| "learning_rate": 0.0003501524195066741, | |
| "loss": 2.8972990036010744, | |
| "step": 10620 | |
| }, | |
| { | |
| "epoch": 22.284891332809636, | |
| "learning_rate": 0.0003499876806504578, | |
| "loss": 2.9171611785888674, | |
| "step": 10630 | |
| }, | |
| { | |
| "epoch": 22.305839224927993, | |
| "learning_rate": 0.00034982317409347263, | |
| "loss": 2.9115921020507813, | |
| "step": 10640 | |
| }, | |
| { | |
| "epoch": 22.326787117046347, | |
| "learning_rate": 0.00034965889929028707, | |
| "loss": 3.029188537597656, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 22.3477350091647, | |
| "learning_rate": 0.000349494855697261, | |
| "loss": 2.900659370422363, | |
| "step": 10660 | |
| }, | |
| { | |
| "epoch": 22.36868290128306, | |
| "learning_rate": 0.0003493310427725377, | |
| "loss": 2.8941118240356447, | |
| "step": 10670 | |
| }, | |
| { | |
| "epoch": 22.389630793401412, | |
| "learning_rate": 0.0003491674599760369, | |
| "loss": 2.913021278381348, | |
| "step": 10680 | |
| }, | |
| { | |
| "epoch": 22.41057868551977, | |
| "learning_rate": 0.0003490041067694469, | |
| "loss": 2.907943916320801, | |
| "step": 10690 | |
| }, | |
| { | |
| "epoch": 22.431526577638124, | |
| "learning_rate": 0.00034884098261621724, | |
| "loss": 2.8749153137207033, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 22.45247446975648, | |
| "learning_rate": 0.00034867808698155125, | |
| "loss": 2.836786460876465, | |
| "step": 10710 | |
| }, | |
| { | |
| "epoch": 22.473422361874835, | |
| "learning_rate": 0.0003485154193323988, | |
| "loss": 2.9680070877075195, | |
| "step": 10720 | |
| }, | |
| { | |
| "epoch": 22.494370253993193, | |
| "learning_rate": 0.00034835297913744903, | |
| "loss": 2.839517593383789, | |
| "step": 10730 | |
| }, | |
| { | |
| "epoch": 22.515318146111547, | |
| "learning_rate": 0.0003481907658671227, | |
| "loss": 2.9294412612915037, | |
| "step": 10740 | |
| }, | |
| { | |
| "epoch": 22.536266038229904, | |
| "learning_rate": 0.0003480287789935653, | |
| "loss": 2.931773376464844, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 22.557213930348258, | |
| "learning_rate": 0.00034786701799063976, | |
| "loss": 2.8776823043823243, | |
| "step": 10760 | |
| }, | |
| { | |
| "epoch": 22.578161822466615, | |
| "learning_rate": 0.00034770548233391924, | |
| "loss": 2.8429998397827148, | |
| "step": 10770 | |
| }, | |
| { | |
| "epoch": 22.59910971458497, | |
| "learning_rate": 0.0003475441715006799, | |
| "loss": 2.8952512741088867, | |
| "step": 10780 | |
| }, | |
| { | |
| "epoch": 22.620057606703327, | |
| "learning_rate": 0.0003473830849698938, | |
| "loss": 2.8576940536499023, | |
| "step": 10790 | |
| }, | |
| { | |
| "epoch": 22.64100549882168, | |
| "learning_rate": 0.0003472222222222222, | |
| "loss": 2.9290803909301757, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 22.661953390940038, | |
| "learning_rate": 0.00034706158274000796, | |
| "loss": 2.861796569824219, | |
| "step": 10810 | |
| }, | |
| { | |
| "epoch": 22.682901283058392, | |
| "learning_rate": 0.00034690116600726885, | |
| "loss": 2.9451555252075194, | |
| "step": 10820 | |
| }, | |
| { | |
| "epoch": 22.70384917517675, | |
| "learning_rate": 0.0003467409715096907, | |
| "loss": 2.8637598037719725, | |
| "step": 10830 | |
| }, | |
| { | |
| "epoch": 22.724797067295103, | |
| "learning_rate": 0.00034658099873462027, | |
| "loss": 2.919582176208496, | |
| "step": 10840 | |
| }, | |
| { | |
| "epoch": 22.745744959413457, | |
| "learning_rate": 0.0003464212471710583, | |
| "loss": 2.9625146865844725, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 22.766692851531815, | |
| "learning_rate": 0.0003462617163096529, | |
| "loss": 2.8601587295532225, | |
| "step": 10860 | |
| }, | |
| { | |
| "epoch": 22.78764074365017, | |
| "learning_rate": 0.00034610240564269265, | |
| "loss": 2.8949390411376954, | |
| "step": 10870 | |
| }, | |
| { | |
| "epoch": 22.808588635768526, | |
| "learning_rate": 0.0003459433146640997, | |
| "loss": 2.8205642700195312, | |
| "step": 10880 | |
| }, | |
| { | |
| "epoch": 22.82953652788688, | |
| "learning_rate": 0.00034578444286942307, | |
| "loss": 2.93542366027832, | |
| "step": 10890 | |
| }, | |
| { | |
| "epoch": 22.850484420005237, | |
| "learning_rate": 0.00034562578975583187, | |
| "loss": 2.8913852691650392, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 22.87143231212359, | |
| "learning_rate": 0.00034546735482210894, | |
| "loss": 2.9478212356567384, | |
| "step": 10910 | |
| }, | |
| { | |
| "epoch": 22.89238020424195, | |
| "learning_rate": 0.0003453091375686437, | |
| "loss": 2.854338455200195, | |
| "step": 10920 | |
| }, | |
| { | |
| "epoch": 22.913328096360303, | |
| "learning_rate": 0.00034515113749742586, | |
| "loss": 2.919601058959961, | |
| "step": 10930 | |
| }, | |
| { | |
| "epoch": 22.93427598847866, | |
| "learning_rate": 0.00034499335411203894, | |
| "loss": 2.88704776763916, | |
| "step": 10940 | |
| }, | |
| { | |
| "epoch": 22.955223880597014, | |
| "learning_rate": 0.00034483578691765326, | |
| "loss": 2.906253433227539, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 22.97617177271537, | |
| "learning_rate": 0.00034467843542102, | |
| "loss": 2.8981559753417967, | |
| "step": 10960 | |
| }, | |
| { | |
| "epoch": 22.997119664833725, | |
| "learning_rate": 0.0003445212991304641, | |
| "loss": 2.9073596954345704, | |
| "step": 10970 | |
| }, | |
| { | |
| "epoch": 23.01885310290652, | |
| "learning_rate": 0.00034436437755587827, | |
| "loss": 2.960616874694824, | |
| "step": 10980 | |
| }, | |
| { | |
| "epoch": 23.039800995024876, | |
| "learning_rate": 0.00034420767020871656, | |
| "loss": 2.9549840927124023, | |
| "step": 10990 | |
| }, | |
| { | |
| "epoch": 23.06074888714323, | |
| "learning_rate": 0.00034405117660198765, | |
| "loss": 2.9256917953491213, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 23.081696779261588, | |
| "learning_rate": 0.00034389489625024885, | |
| "loss": 2.89876651763916, | |
| "step": 11010 | |
| }, | |
| { | |
| "epoch": 23.10264467137994, | |
| "learning_rate": 0.00034373882866959936, | |
| "loss": 2.828813362121582, | |
| "step": 11020 | |
| }, | |
| { | |
| "epoch": 23.1235925634983, | |
| "learning_rate": 0.0003435829733776745, | |
| "loss": 2.9374326705932616, | |
| "step": 11030 | |
| }, | |
| { | |
| "epoch": 23.144540455616653, | |
| "learning_rate": 0.00034342732989363903, | |
| "loss": 2.928928184509277, | |
| "step": 11040 | |
| }, | |
| { | |
| "epoch": 23.16548834773501, | |
| "learning_rate": 0.0003432718977381811, | |
| "loss": 2.8702091217041015, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 23.186436239853364, | |
| "learning_rate": 0.0003431166764335058, | |
| "loss": 2.937228965759277, | |
| "step": 11060 | |
| }, | |
| { | |
| "epoch": 23.207384131971722, | |
| "learning_rate": 0.0003429616655033297, | |
| "loss": 2.9454578399658202, | |
| "step": 11070 | |
| }, | |
| { | |
| "epoch": 23.228332024090076, | |
| "learning_rate": 0.00034280686447287373, | |
| "loss": 2.8274587631225585, | |
| "step": 11080 | |
| }, | |
| { | |
| "epoch": 23.249279916208433, | |
| "learning_rate": 0.00034265227286885776, | |
| "loss": 2.876905632019043, | |
| "step": 11090 | |
| }, | |
| { | |
| "epoch": 23.270227808326787, | |
| "learning_rate": 0.00034249789021949435, | |
| "loss": 2.8264415740966795, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 23.29117570044514, | |
| "learning_rate": 0.0003423437160544826, | |
| "loss": 2.8484895706176756, | |
| "step": 11110 | |
| }, | |
| { | |
| "epoch": 23.3121235925635, | |
| "learning_rate": 0.0003421897499050022, | |
| "loss": 2.885685920715332, | |
| "step": 11120 | |
| }, | |
| { | |
| "epoch": 23.333071484681852, | |
| "learning_rate": 0.0003420359913037075, | |
| "loss": 2.9073020935058596, | |
| "step": 11130 | |
| }, | |
| { | |
| "epoch": 23.35401937680021, | |
| "learning_rate": 0.0003418824397847216, | |
| "loss": 2.919231986999512, | |
| "step": 11140 | |
| }, | |
| { | |
| "epoch": 23.374967268918564, | |
| "learning_rate": 0.00034172909488363007, | |
| "loss": 2.8528385162353516, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 23.39591516103692, | |
| "learning_rate": 0.00034157595613747545, | |
| "loss": 2.8856670379638674, | |
| "step": 11160 | |
| }, | |
| { | |
| "epoch": 23.416863053155275, | |
| "learning_rate": 0.00034142302308475133, | |
| "loss": 2.851297950744629, | |
| "step": 11170 | |
| }, | |
| { | |
| "epoch": 23.437810945273633, | |
| "learning_rate": 0.0003412702952653962, | |
| "loss": 2.8231760025024415, | |
| "step": 11180 | |
| }, | |
| { | |
| "epoch": 23.458758837391986, | |
| "learning_rate": 0.00034111777222078796, | |
| "loss": 2.9219854354858397, | |
| "step": 11190 | |
| }, | |
| { | |
| "epoch": 23.479706729510344, | |
| "learning_rate": 0.00034096545349373804, | |
| "loss": 2.977204132080078, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 23.500654621628698, | |
| "learning_rate": 0.0003408133386284857, | |
| "loss": 2.8231952667236326, | |
| "step": 11210 | |
| }, | |
| { | |
| "epoch": 23.521602513747055, | |
| "learning_rate": 0.0003406614271706919, | |
| "loss": 2.891893196105957, | |
| "step": 11220 | |
| }, | |
| { | |
| "epoch": 23.54255040586541, | |
| "learning_rate": 0.0003405097186674344, | |
| "loss": 2.8958648681640624, | |
| "step": 11230 | |
| }, | |
| { | |
| "epoch": 23.563498297983767, | |
| "learning_rate": 0.00034035821266720136, | |
| "loss": 2.8495506286621093, | |
| "step": 11240 | |
| }, | |
| { | |
| "epoch": 23.58444619010212, | |
| "learning_rate": 0.0003402069087198858, | |
| "loss": 2.935627746582031, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 23.605394082220478, | |
| "learning_rate": 0.00034005580637678053, | |
| "loss": 2.840359687805176, | |
| "step": 11260 | |
| }, | |
| { | |
| "epoch": 23.626341974338832, | |
| "learning_rate": 0.00033990490519057183, | |
| "loss": 2.9121625900268553, | |
| "step": 11270 | |
| }, | |
| { | |
| "epoch": 23.64728986645719, | |
| "learning_rate": 0.0003397542047153345, | |
| "loss": 2.897580146789551, | |
| "step": 11280 | |
| }, | |
| { | |
| "epoch": 23.668237758575543, | |
| "learning_rate": 0.0003396037045065257, | |
| "loss": 2.894269561767578, | |
| "step": 11290 | |
| }, | |
| { | |
| "epoch": 23.6891856506939, | |
| "learning_rate": 0.0003394534041209802, | |
| "loss": 2.923667335510254, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 23.710133542812255, | |
| "learning_rate": 0.0003393033031169043, | |
| "loss": 2.9479984283447265, | |
| "step": 11310 | |
| }, | |
| { | |
| "epoch": 23.73108143493061, | |
| "learning_rate": 0.0003391534010538705, | |
| "loss": 2.9156826019287108, | |
| "step": 11320 | |
| }, | |
| { | |
| "epoch": 23.752029327048966, | |
| "learning_rate": 0.00033900369749281225, | |
| "loss": 2.9133535385131837, | |
| "step": 11330 | |
| }, | |
| { | |
| "epoch": 23.77297721916732, | |
| "learning_rate": 0.00033885419199601845, | |
| "loss": 2.912689971923828, | |
| "step": 11340 | |
| }, | |
| { | |
| "epoch": 23.793925111285677, | |
| "learning_rate": 0.000338704884127128, | |
| "loss": 2.921385955810547, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 23.81487300340403, | |
| "learning_rate": 0.00033855577345112453, | |
| "loss": 2.9245376586914062, | |
| "step": 11360 | |
| }, | |
| { | |
| "epoch": 23.83582089552239, | |
| "learning_rate": 0.0003384068595343312, | |
| "loss": 2.921660232543945, | |
| "step": 11370 | |
| }, | |
| { | |
| "epoch": 23.856768787640743, | |
| "learning_rate": 0.00033825814194440504, | |
| "loss": 2.8597009658813475, | |
| "step": 11380 | |
| }, | |
| { | |
| "epoch": 23.8777166797591, | |
| "learning_rate": 0.0003381096202503321, | |
| "loss": 2.9699087142944336, | |
| "step": 11390 | |
| }, | |
| { | |
| "epoch": 23.898664571877454, | |
| "learning_rate": 0.00033796129402242193, | |
| "loss": 2.868007850646973, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 23.91961246399581, | |
| "learning_rate": 0.0003378131628323024, | |
| "loss": 2.8257036209106445, | |
| "step": 11410 | |
| }, | |
| { | |
| "epoch": 23.940560356114165, | |
| "learning_rate": 0.0003376652262529146, | |
| "loss": 2.8747650146484376, | |
| "step": 11420 | |
| }, | |
| { | |
| "epoch": 23.961508248232523, | |
| "learning_rate": 0.00033751748385850753, | |
| "loss": 2.8721830368041994, | |
| "step": 11430 | |
| }, | |
| { | |
| "epoch": 23.982456140350877, | |
| "learning_rate": 0.00033736993522463316, | |
| "loss": 2.8383148193359373, | |
| "step": 11440 | |
| }, | |
| { | |
| "epoch": 24.00418957842367, | |
| "learning_rate": 0.00033722257992814113, | |
| "loss": 2.9513004302978514, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 24.025137470542028, | |
| "learning_rate": 0.0003370754175471737, | |
| "loss": 2.9344108581542967, | |
| "step": 11460 | |
| }, | |
| { | |
| "epoch": 24.04608536266038, | |
| "learning_rate": 0.0003369284476611607, | |
| "loss": 2.9109573364257812, | |
| "step": 11470 | |
| }, | |
| { | |
| "epoch": 24.06703325477874, | |
| "learning_rate": 0.00033678166985081433, | |
| "loss": 2.8260976791381838, | |
| "step": 11480 | |
| }, | |
| { | |
| "epoch": 24.087981146897093, | |
| "learning_rate": 0.0003366350836981245, | |
| "loss": 2.790972137451172, | |
| "step": 11490 | |
| }, | |
| { | |
| "epoch": 24.10892903901545, | |
| "learning_rate": 0.0003364886887863534, | |
| "loss": 2.8716499328613283, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 24.129876931133804, | |
| "learning_rate": 0.0003363424847000309, | |
| "loss": 2.8708847045898436, | |
| "step": 11510 | |
| }, | |
| { | |
| "epoch": 24.15082482325216, | |
| "learning_rate": 0.0003361964710249494, | |
| "loss": 2.8762466430664064, | |
| "step": 11520 | |
| }, | |
| { | |
| "epoch": 24.171772715370516, | |
| "learning_rate": 0.00033605064734815865, | |
| "loss": 2.8436599731445313, | |
| "step": 11530 | |
| }, | |
| { | |
| "epoch": 24.192720607488873, | |
| "learning_rate": 0.0003359050132579615, | |
| "loss": 2.885796546936035, | |
| "step": 11540 | |
| }, | |
| { | |
| "epoch": 24.213668499607227, | |
| "learning_rate": 0.00033575956834390843, | |
| "loss": 2.8459619522094726, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 24.234616391725584, | |
| "learning_rate": 0.00033561431219679297, | |
| "loss": 2.8883172988891603, | |
| "step": 11560 | |
| }, | |
| { | |
| "epoch": 24.25556428384394, | |
| "learning_rate": 0.00033546924440864666, | |
| "loss": 2.8706939697265623, | |
| "step": 11570 | |
| }, | |
| { | |
| "epoch": 24.276512175962292, | |
| "learning_rate": 0.0003353243645727346, | |
| "loss": 2.8278776168823243, | |
| "step": 11580 | |
| }, | |
| { | |
| "epoch": 24.29746006808065, | |
| "learning_rate": 0.0003351796722835502, | |
| "loss": 2.8596363067626953, | |
| "step": 11590 | |
| }, | |
| { | |
| "epoch": 24.318407960199004, | |
| "learning_rate": 0.00033503516713681087, | |
| "loss": 2.8937658309936523, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 24.33935585231736, | |
| "learning_rate": 0.00033489084872945283, | |
| "loss": 2.9479068756103515, | |
| "step": 11610 | |
| }, | |
| { | |
| "epoch": 24.360303744435715, | |
| "learning_rate": 0.0003347467166596268, | |
| "loss": 2.913376235961914, | |
| "step": 11620 | |
| }, | |
| { | |
| "epoch": 24.381251636554072, | |
| "learning_rate": 0.0003346027705266929, | |
| "loss": 2.7945356369018555, | |
| "step": 11630 | |
| }, | |
| { | |
| "epoch": 24.402199528672426, | |
| "learning_rate": 0.0003344590099312164, | |
| "loss": 2.908190155029297, | |
| "step": 11640 | |
| }, | |
| { | |
| "epoch": 24.423147420790784, | |
| "learning_rate": 0.00033431543447496275, | |
| "loss": 2.8602962493896484, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 24.444095312909138, | |
| "learning_rate": 0.000334172043760893, | |
| "loss": 2.8535890579223633, | |
| "step": 11660 | |
| }, | |
| { | |
| "epoch": 24.465043205027495, | |
| "learning_rate": 0.0003340288373931593, | |
| "loss": 2.8150957107543944, | |
| "step": 11670 | |
| }, | |
| { | |
| "epoch": 24.48599109714585, | |
| "learning_rate": 0.0003338858149771002, | |
| "loss": 2.847452735900879, | |
| "step": 11680 | |
| }, | |
| { | |
| "epoch": 24.506938989264206, | |
| "learning_rate": 0.0003337429761192361, | |
| "loss": 2.7802717208862306, | |
| "step": 11690 | |
| }, | |
| { | |
| "epoch": 24.52788688138256, | |
| "learning_rate": 0.00033360032042726483, | |
| "loss": 2.9678937911987306, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 24.548834773500918, | |
| "learning_rate": 0.000333457847510057, | |
| "loss": 2.8469560623168944, | |
| "step": 11710 | |
| }, | |
| { | |
| "epoch": 24.56978266561927, | |
| "learning_rate": 0.0003333155569776514, | |
| "loss": 2.916895866394043, | |
| "step": 11720 | |
| }, | |
| { | |
| "epoch": 24.59073055773763, | |
| "learning_rate": 0.00033317344844125064, | |
| "loss": 2.8457548141479494, | |
| "step": 11730 | |
| }, | |
| { | |
| "epoch": 24.611678449855983, | |
| "learning_rate": 0.00033303152151321696, | |
| "loss": 2.872743606567383, | |
| "step": 11740 | |
| }, | |
| { | |
| "epoch": 24.63262634197434, | |
| "learning_rate": 0.00033288977580706714, | |
| "loss": 2.890146255493164, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 24.653574234092694, | |
| "learning_rate": 0.0003327482109374687, | |
| "loss": 2.847947883605957, | |
| "step": 11760 | |
| }, | |
| { | |
| "epoch": 24.67452212621105, | |
| "learning_rate": 0.00033260682652023517, | |
| "loss": 2.9236717224121094, | |
| "step": 11770 | |
| }, | |
| { | |
| "epoch": 24.695470018329406, | |
| "learning_rate": 0.0003324656221723217, | |
| "loss": 2.9201459884643555, | |
| "step": 11780 | |
| }, | |
| { | |
| "epoch": 24.71641791044776, | |
| "learning_rate": 0.000332324597511821, | |
| "loss": 2.8557527542114256, | |
| "step": 11790 | |
| }, | |
| { | |
| "epoch": 24.737365802566117, | |
| "learning_rate": 0.00033218375215795864, | |
| "loss": 2.875984001159668, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 24.75831369468447, | |
| "learning_rate": 0.00033204308573108897, | |
| "loss": 2.830782890319824, | |
| "step": 11810 | |
| }, | |
| { | |
| "epoch": 24.77926158680283, | |
| "learning_rate": 0.00033190259785269066, | |
| "loss": 2.844138526916504, | |
| "step": 11820 | |
| }, | |
| { | |
| "epoch": 24.800209478921182, | |
| "learning_rate": 0.0003317622881453626, | |
| "loss": 2.8771383285522463, | |
| "step": 11830 | |
| }, | |
| { | |
| "epoch": 24.82115737103954, | |
| "learning_rate": 0.0003316221562328194, | |
| "loss": 2.863381767272949, | |
| "step": 11840 | |
| }, | |
| { | |
| "epoch": 24.842105263157894, | |
| "learning_rate": 0.0003314822017398875, | |
| "loss": 2.8934911727905273, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 24.86305315527625, | |
| "learning_rate": 0.00033134242429250053, | |
| "loss": 2.7928911209106446, | |
| "step": 11860 | |
| }, | |
| { | |
| "epoch": 24.884001047394605, | |
| "learning_rate": 0.00033120282351769556, | |
| "loss": 2.8646501541137694, | |
| "step": 11870 | |
| }, | |
| { | |
| "epoch": 24.904948939512963, | |
| "learning_rate": 0.0003310633990436084, | |
| "loss": 2.8573431015014648, | |
| "step": 11880 | |
| }, | |
| { | |
| "epoch": 24.925896831631317, | |
| "learning_rate": 0.00033092415049947006, | |
| "loss": 2.9596303939819335, | |
| "step": 11890 | |
| }, | |
| { | |
| "epoch": 24.946844723749674, | |
| "learning_rate": 0.00033078507751560195, | |
| "loss": 2.809922790527344, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 24.967792615868028, | |
| "learning_rate": 0.00033064617972341235, | |
| "loss": 2.829710578918457, | |
| "step": 11910 | |
| }, | |
| { | |
| "epoch": 24.988740507986385, | |
| "learning_rate": 0.0003305074567553919, | |
| "loss": 2.837497520446777, | |
| "step": 11920 | |
| }, | |
| { | |
| "epoch": 25.01047394605918, | |
| "learning_rate": 0.0003303689082451096, | |
| "loss": 2.998362922668457, | |
| "step": 11930 | |
| }, | |
| { | |
| "epoch": 25.031421838177533, | |
| "learning_rate": 0.00033023053382720904, | |
| "loss": 2.8903406143188475, | |
| "step": 11940 | |
| }, | |
| { | |
| "epoch": 25.05236973029589, | |
| "learning_rate": 0.0003300923331374039, | |
| "loss": 2.88183536529541, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 25.073317622414244, | |
| "learning_rate": 0.00032995430581247417, | |
| "loss": 2.8853179931640627, | |
| "step": 11960 | |
| }, | |
| { | |
| "epoch": 25.0942655145326, | |
| "learning_rate": 0.0003298164514902622, | |
| "loss": 2.8495412826538087, | |
| "step": 11970 | |
| }, | |
| { | |
| "epoch": 25.115213406650955, | |
| "learning_rate": 0.0003296787698096686, | |
| "loss": 2.8417972564697265, | |
| "step": 11980 | |
| }, | |
| { | |
| "epoch": 25.136161298769313, | |
| "learning_rate": 0.0003295412604106482, | |
| "loss": 2.8254583358764647, | |
| "step": 11990 | |
| }, | |
| { | |
| "epoch": 25.157109190887667, | |
| "learning_rate": 0.00032940392293420614, | |
| "loss": 2.8476821899414064, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 25.178057083006024, | |
| "learning_rate": 0.00032926675702239425, | |
| "loss": 2.8646284103393556, | |
| "step": 12010 | |
| }, | |
| { | |
| "epoch": 25.199004975124378, | |
| "learning_rate": 0.00032912976231830646, | |
| "loss": 2.8645925521850586, | |
| "step": 12020 | |
| }, | |
| { | |
| "epoch": 25.219952867242732, | |
| "learning_rate": 0.0003289929384660757, | |
| "loss": 2.9142387390136717, | |
| "step": 12030 | |
| }, | |
| { | |
| "epoch": 25.24090075936109, | |
| "learning_rate": 0.0003288562851108693, | |
| "loss": 2.911361312866211, | |
| "step": 12040 | |
| }, | |
| { | |
| "epoch": 25.261848651479443, | |
| "learning_rate": 0.0003287198018988856, | |
| "loss": 2.810334014892578, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 25.2827965435978, | |
| "learning_rate": 0.00032858348847734985, | |
| "loss": 2.8691171646118163, | |
| "step": 12060 | |
| }, | |
| { | |
| "epoch": 25.303744435716155, | |
| "learning_rate": 0.00032844734449451055, | |
| "loss": 2.8615827560424805, | |
| "step": 12070 | |
| }, | |
| { | |
| "epoch": 25.324692327834512, | |
| "learning_rate": 0.00032831136959963553, | |
| "loss": 2.8075706481933596, | |
| "step": 12080 | |
| }, | |
| { | |
| "epoch": 25.345640219952866, | |
| "learning_rate": 0.00032817556344300823, | |
| "loss": 2.836076354980469, | |
| "step": 12090 | |
| }, | |
| { | |
| "epoch": 25.366588112071224, | |
| "learning_rate": 0.0003280399256759237, | |
| "loss": 2.873185729980469, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 25.387536004189577, | |
| "learning_rate": 0.0003279044559506852, | |
| "loss": 2.8835927963256838, | |
| "step": 12110 | |
| }, | |
| { | |
| "epoch": 25.408483896307935, | |
| "learning_rate": 0.0003277691539206003, | |
| "loss": 2.884838676452637, | |
| "step": 12120 | |
| }, | |
| { | |
| "epoch": 25.42943178842629, | |
| "learning_rate": 0.0003276340192399769, | |
| "loss": 2.9353681564331056, | |
| "step": 12130 | |
| }, | |
| { | |
| "epoch": 25.450379680544646, | |
| "learning_rate": 0.00032749905156412, | |
| "loss": 2.87127571105957, | |
| "step": 12140 | |
| }, | |
| { | |
| "epoch": 25.471327572663, | |
| "learning_rate": 0.0003273642505493275, | |
| "loss": 2.848041534423828, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 25.492275464781358, | |
| "learning_rate": 0.0003272296158528871, | |
| "loss": 2.8736820220947266, | |
| "step": 12160 | |
| }, | |
| { | |
| "epoch": 25.51322335689971, | |
| "learning_rate": 0.000327095147133072, | |
| "loss": 2.889766502380371, | |
| "step": 12170 | |
| }, | |
| { | |
| "epoch": 25.53417124901807, | |
| "learning_rate": 0.00032696084404913777, | |
| "loss": 2.8456445693969727, | |
| "step": 12180 | |
| }, | |
| { | |
| "epoch": 25.555119141136423, | |
| "learning_rate": 0.00032682670626131837, | |
| "loss": 2.8694175720214843, | |
| "step": 12190 | |
| }, | |
| { | |
| "epoch": 25.57606703325478, | |
| "learning_rate": 0.0003266927334308229, | |
| "loss": 2.863827705383301, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 25.597014925373134, | |
| "learning_rate": 0.0003265589252198317, | |
| "loss": 2.8949514389038087, | |
| "step": 12210 | |
| }, | |
| { | |
| "epoch": 25.617962817491488, | |
| "learning_rate": 0.0003264252812914928, | |
| "loss": 2.870989990234375, | |
| "step": 12220 | |
| }, | |
| { | |
| "epoch": 25.638910709609846, | |
| "learning_rate": 0.0003262918013099186, | |
| "loss": 2.8301280975341796, | |
| "step": 12230 | |
| }, | |
| { | |
| "epoch": 25.6598586017282, | |
| "learning_rate": 0.00032615848494018204, | |
| "loss": 2.7910818099975585, | |
| "step": 12240 | |
| }, | |
| { | |
| "epoch": 25.680806493846557, | |
| "learning_rate": 0.0003260253318483131, | |
| "loss": 2.8830698013305662, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 25.70175438596491, | |
| "learning_rate": 0.0003258923417012957, | |
| "loss": 2.886226844787598, | |
| "step": 12260 | |
| }, | |
| { | |
| "epoch": 25.72270227808327, | |
| "learning_rate": 0.00032575951416706354, | |
| "loss": 2.9646997451782227, | |
| "step": 12270 | |
| }, | |
| { | |
| "epoch": 25.743650170201622, | |
| "learning_rate": 0.0003256268489144972, | |
| "loss": 2.896713066101074, | |
| "step": 12280 | |
| }, | |
| { | |
| "epoch": 25.76459806231998, | |
| "learning_rate": 0.0003254943456134202, | |
| "loss": 2.8680368423461915, | |
| "step": 12290 | |
| }, | |
| { | |
| "epoch": 25.785545954438334, | |
| "learning_rate": 0.0003253620039345959, | |
| "loss": 2.866026496887207, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 25.80649384655669, | |
| "learning_rate": 0.0003252298235497241, | |
| "loss": 2.862067985534668, | |
| "step": 12310 | |
| }, | |
| { | |
| "epoch": 25.827441738675045, | |
| "learning_rate": 0.0003250978041314371, | |
| "loss": 2.8973188400268555, | |
| "step": 12320 | |
| }, | |
| { | |
| "epoch": 25.848389630793402, | |
| "learning_rate": 0.000324965945353297, | |
| "loss": 2.9389106750488283, | |
| "step": 12330 | |
| }, | |
| { | |
| "epoch": 25.869337522911756, | |
| "learning_rate": 0.0003248342468897917, | |
| "loss": 2.9147424697875977, | |
| "step": 12340 | |
| }, | |
| { | |
| "epoch": 25.890285415030114, | |
| "learning_rate": 0.00032470270841633195, | |
| "loss": 2.894465446472168, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 25.911233307148468, | |
| "learning_rate": 0.00032457132960924783, | |
| "loss": 2.9301485061645507, | |
| "step": 12360 | |
| }, | |
| { | |
| "epoch": 25.932181199266825, | |
| "learning_rate": 0.00032444011014578535, | |
| "loss": 2.8576644897460937, | |
| "step": 12370 | |
| }, | |
| { | |
| "epoch": 25.95312909138518, | |
| "learning_rate": 0.00032430904970410314, | |
| "loss": 2.836701202392578, | |
| "step": 12380 | |
| }, | |
| { | |
| "epoch": 25.974076983503537, | |
| "learning_rate": 0.0003241781479632693, | |
| "loss": 2.8457481384277346, | |
| "step": 12390 | |
| }, | |
| { | |
| "epoch": 25.99502487562189, | |
| "learning_rate": 0.0003240474046032579, | |
| "loss": 2.829239082336426, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 26.016758313694684, | |
| "learning_rate": 0.00032391681930494566, | |
| "loss": 3.0122323989868165, | |
| "step": 12410 | |
| }, | |
| { | |
| "epoch": 26.03770620581304, | |
| "learning_rate": 0.000323786391750109, | |
| "loss": 2.8899608612060548, | |
| "step": 12420 | |
| }, | |
| { | |
| "epoch": 26.058654097931395, | |
| "learning_rate": 0.0003236561216214202, | |
| "loss": 2.9380813598632813, | |
| "step": 12430 | |
| }, | |
| { | |
| "epoch": 26.079601990049753, | |
| "learning_rate": 0.000323526008602445, | |
| "loss": 2.9364286422729493, | |
| "step": 12440 | |
| }, | |
| { | |
| "epoch": 26.100549882168107, | |
| "learning_rate": 0.0003233960523776387, | |
| "loss": 2.8298776626586912, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 26.121497774286464, | |
| "learning_rate": 0.0003232662526323429, | |
| "loss": 2.868173027038574, | |
| "step": 12460 | |
| }, | |
| { | |
| "epoch": 26.142445666404818, | |
| "learning_rate": 0.0003231366090527828, | |
| "loss": 2.8364093780517576, | |
| "step": 12470 | |
| }, | |
| { | |
| "epoch": 26.163393558523175, | |
| "learning_rate": 0.00032300712132606366, | |
| "loss": 2.917738342285156, | |
| "step": 12480 | |
| }, | |
| { | |
| "epoch": 26.18434145064153, | |
| "learning_rate": 0.0003228777891401678, | |
| "loss": 2.8115827560424806, | |
| "step": 12490 | |
| }, | |
| { | |
| "epoch": 26.205289342759883, | |
| "learning_rate": 0.0003227486121839514, | |
| "loss": 2.8544151306152346, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 26.22623723487824, | |
| "learning_rate": 0.00032261959014714107, | |
| "loss": 2.890985870361328, | |
| "step": 12510 | |
| }, | |
| { | |
| "epoch": 26.247185126996595, | |
| "learning_rate": 0.0003224907227203312, | |
| "loss": 2.8269269943237303, | |
| "step": 12520 | |
| }, | |
| { | |
| "epoch": 26.268133019114952, | |
| "learning_rate": 0.0003223620095949806, | |
| "loss": 2.8392301559448243, | |
| "step": 12530 | |
| }, | |
| { | |
| "epoch": 26.289080911233306, | |
| "learning_rate": 0.00032223345046340936, | |
| "loss": 2.8283065795898437, | |
| "step": 12540 | |
| }, | |
| { | |
| "epoch": 26.310028803351663, | |
| "learning_rate": 0.00032210504501879576, | |
| "loss": 2.9033248901367186, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 26.330976695470017, | |
| "learning_rate": 0.0003219767929551733, | |
| "loss": 2.8192907333374024, | |
| "step": 12560 | |
| }, | |
| { | |
| "epoch": 26.351924587588375, | |
| "learning_rate": 0.00032184869396742754, | |
| "loss": 2.8758308410644533, | |
| "step": 12570 | |
| }, | |
| { | |
| "epoch": 26.37287247970673, | |
| "learning_rate": 0.00032172074775129323, | |
| "loss": 2.8491661071777346, | |
| "step": 12580 | |
| }, | |
| { | |
| "epoch": 26.393820371825086, | |
| "learning_rate": 0.00032159295400335114, | |
| "loss": 2.862008285522461, | |
| "step": 12590 | |
| }, | |
| { | |
| "epoch": 26.41476826394344, | |
| "learning_rate": 0.00032146531242102476, | |
| "loss": 2.854539489746094, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 26.435716156061797, | |
| "learning_rate": 0.0003213378227025779, | |
| "loss": 2.9059074401855467, | |
| "step": 12610 | |
| }, | |
| { | |
| "epoch": 26.45666404818015, | |
| "learning_rate": 0.00032121048454711114, | |
| "loss": 2.8347517013549806, | |
| "step": 12620 | |
| }, | |
| { | |
| "epoch": 26.47761194029851, | |
| "learning_rate": 0.00032108329765455926, | |
| "loss": 2.8621740341186523, | |
| "step": 12630 | |
| }, | |
| { | |
| "epoch": 26.498559832416863, | |
| "learning_rate": 0.00032095626172568784, | |
| "loss": 2.8287914276123045, | |
| "step": 12640 | |
| }, | |
| { | |
| "epoch": 26.51950772453522, | |
| "learning_rate": 0.00032082937646209084, | |
| "loss": 2.8201780319213867, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 26.540455616653574, | |
| "learning_rate": 0.0003207026415661871, | |
| "loss": 2.853387451171875, | |
| "step": 12660 | |
| }, | |
| { | |
| "epoch": 26.56140350877193, | |
| "learning_rate": 0.0003205760567412178, | |
| "loss": 2.8255029678344727, | |
| "step": 12670 | |
| }, | |
| { | |
| "epoch": 26.582351400890285, | |
| "learning_rate": 0.00032044962169124335, | |
| "loss": 2.8133966445922853, | |
| "step": 12680 | |
| }, | |
| { | |
| "epoch": 26.60329929300864, | |
| "learning_rate": 0.0003203233361211406, | |
| "loss": 2.8209064483642576, | |
| "step": 12690 | |
| }, | |
| { | |
| "epoch": 26.624247185126997, | |
| "learning_rate": 0.00032019719973659996, | |
| "loss": 2.839722442626953, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 26.64519507724535, | |
| "learning_rate": 0.00032007121224412224, | |
| "loss": 2.8414018630981444, | |
| "step": 12710 | |
| }, | |
| { | |
| "epoch": 26.666142969363708, | |
| "learning_rate": 0.0003199453733510162, | |
| "loss": 2.8677789688110353, | |
| "step": 12720 | |
| }, | |
| { | |
| "epoch": 26.687090861482062, | |
| "learning_rate": 0.00031981968276539543, | |
| "loss": 2.9177148818969725, | |
| "step": 12730 | |
| }, | |
| { | |
| "epoch": 26.70803875360042, | |
| "learning_rate": 0.0003196941401961754, | |
| "loss": 2.8555475234985352, | |
| "step": 12740 | |
| }, | |
| { | |
| "epoch": 26.728986645718773, | |
| "learning_rate": 0.000319568745353071, | |
| "loss": 2.8636154174804687, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 26.74993453783713, | |
| "learning_rate": 0.0003194434979465935, | |
| "loss": 2.810639190673828, | |
| "step": 12760 | |
| }, | |
| { | |
| "epoch": 26.770882429955485, | |
| "learning_rate": 0.0003193183976880476, | |
| "loss": 2.9356500625610353, | |
| "step": 12770 | |
| }, | |
| { | |
| "epoch": 26.791830322073842, | |
| "learning_rate": 0.00031919344428952895, | |
| "loss": 2.848637580871582, | |
| "step": 12780 | |
| }, | |
| { | |
| "epoch": 26.812778214192196, | |
| "learning_rate": 0.0003190686374639211, | |
| "loss": 2.8234004974365234, | |
| "step": 12790 | |
| }, | |
| { | |
| "epoch": 26.833726106310554, | |
| "learning_rate": 0.00031894397692489295, | |
| "loss": 2.8002485275268554, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 26.854673998428908, | |
| "learning_rate": 0.0003188194623868958, | |
| "loss": 2.841193199157715, | |
| "step": 12810 | |
| }, | |
| { | |
| "epoch": 26.875621890547265, | |
| "learning_rate": 0.00031869509356516063, | |
| "loss": 2.8377119064331056, | |
| "step": 12820 | |
| }, | |
| { | |
| "epoch": 26.89656978266562, | |
| "learning_rate": 0.00031857087017569556, | |
| "loss": 2.797208786010742, | |
| "step": 12830 | |
| }, | |
| { | |
| "epoch": 26.917517674783976, | |
| "learning_rate": 0.0003184467919352828, | |
| "loss": 2.778369140625, | |
| "step": 12840 | |
| }, | |
| { | |
| "epoch": 26.93846556690233, | |
| "learning_rate": 0.0003183228585614763, | |
| "loss": 2.8303714752197267, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 26.959413459020688, | |
| "learning_rate": 0.0003181990697725988, | |
| "loss": 2.805090141296387, | |
| "step": 12860 | |
| }, | |
| { | |
| "epoch": 26.98036135113904, | |
| "learning_rate": 0.0003180754252877392, | |
| "loss": 2.7620264053344727, | |
| "step": 12870 | |
| }, | |
| { | |
| "epoch": 27.002094789211835, | |
| "learning_rate": 0.0003179519248267498, | |
| "loss": 2.9018489837646486, | |
| "step": 12880 | |
| }, | |
| { | |
| "epoch": 27.023042681330192, | |
| "learning_rate": 0.000317828568110244, | |
| "loss": 2.8413219451904297, | |
| "step": 12890 | |
| }, | |
| { | |
| "epoch": 27.043990573448546, | |
| "learning_rate": 0.000317705354859593, | |
| "loss": 2.868427276611328, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 27.064938465566904, | |
| "learning_rate": 0.0003175822847969239, | |
| "loss": 2.84520263671875, | |
| "step": 12910 | |
| }, | |
| { | |
| "epoch": 27.085886357685258, | |
| "learning_rate": 0.00031745935764511645, | |
| "loss": 2.865756607055664, | |
| "step": 12920 | |
| }, | |
| { | |
| "epoch": 27.106834249803615, | |
| "learning_rate": 0.0003173365731278007, | |
| "loss": 2.8851186752319338, | |
| "step": 12930 | |
| }, | |
| { | |
| "epoch": 27.12778214192197, | |
| "learning_rate": 0.00031721393096935445, | |
| "loss": 2.8631362915039062, | |
| "step": 12940 | |
| }, | |
| { | |
| "epoch": 27.148730034040323, | |
| "learning_rate": 0.00031709143089490063, | |
| "loss": 2.8974273681640623, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 27.16967792615868, | |
| "learning_rate": 0.00031696907263030445, | |
| "loss": 2.8190950393676757, | |
| "step": 12960 | |
| }, | |
| { | |
| "epoch": 27.190625818277034, | |
| "learning_rate": 0.00031684685590217115, | |
| "loss": 2.861093521118164, | |
| "step": 12970 | |
| }, | |
| { | |
| "epoch": 27.211573710395392, | |
| "learning_rate": 0.00031672478043784336, | |
| "loss": 2.925172233581543, | |
| "step": 12980 | |
| }, | |
| { | |
| "epoch": 27.232521602513746, | |
| "learning_rate": 0.0003166028459653984, | |
| "loss": 2.7551206588745116, | |
| "step": 12990 | |
| }, | |
| { | |
| "epoch": 27.253469494632103, | |
| "learning_rate": 0.0003164810522136458, | |
| "loss": 2.8190824508666994, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 27.274417386750457, | |
| "learning_rate": 0.0003163593989121249, | |
| "loss": 2.884243965148926, | |
| "step": 13010 | |
| }, | |
| { | |
| "epoch": 27.295365278868815, | |
| "learning_rate": 0.0003162378857911022, | |
| "loss": 2.831955909729004, | |
| "step": 13020 | |
| }, | |
| { | |
| "epoch": 27.31631317098717, | |
| "learning_rate": 0.00031611651258156884, | |
| "loss": 2.891588020324707, | |
| "step": 13030 | |
| }, | |
| { | |
| "epoch": 27.337261063105526, | |
| "learning_rate": 0.0003159952790152381, | |
| "loss": 2.8689960479736327, | |
| "step": 13040 | |
| }, | |
| { | |
| "epoch": 27.35820895522388, | |
| "learning_rate": 0.0003158741848245431, | |
| "loss": 2.844234085083008, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 27.379156847342237, | |
| "learning_rate": 0.0003157532297426339, | |
| "loss": 2.783745765686035, | |
| "step": 13060 | |
| }, | |
| { | |
| "epoch": 27.40010473946059, | |
| "learning_rate": 0.00031563241350337546, | |
| "loss": 2.85959415435791, | |
| "step": 13070 | |
| }, | |
| { | |
| "epoch": 27.42105263157895, | |
| "learning_rate": 0.00031551173584134514, | |
| "loss": 2.828862762451172, | |
| "step": 13080 | |
| }, | |
| { | |
| "epoch": 27.442000523697303, | |
| "learning_rate": 0.0003153911964918298, | |
| "loss": 2.8126575469970705, | |
| "step": 13090 | |
| }, | |
| { | |
| "epoch": 27.46294841581566, | |
| "learning_rate": 0.0003152707951908239, | |
| "loss": 2.8336280822753905, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 27.483896307934014, | |
| "learning_rate": 0.0003151505316750269, | |
| "loss": 2.8542291641235353, | |
| "step": 13110 | |
| }, | |
| { | |
| "epoch": 27.50484420005237, | |
| "learning_rate": 0.0003150304056818405, | |
| "loss": 2.8555719375610353, | |
| "step": 13120 | |
| }, | |
| { | |
| "epoch": 27.525792092170725, | |
| "learning_rate": 0.00031491041694936697, | |
| "loss": 2.8440032958984376, | |
| "step": 13130 | |
| }, | |
| { | |
| "epoch": 27.54673998428908, | |
| "learning_rate": 0.000314790565216406, | |
| "loss": 2.899538040161133, | |
| "step": 13140 | |
| }, | |
| { | |
| "epoch": 27.567687876407437, | |
| "learning_rate": 0.0003146708502224526, | |
| "loss": 2.823881149291992, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 27.58863576852579, | |
| "learning_rate": 0.0003145512717076948, | |
| "loss": 2.8198898315429686, | |
| "step": 13160 | |
| }, | |
| { | |
| "epoch": 27.609583660644148, | |
| "learning_rate": 0.00031443182941301147, | |
| "loss": 2.8212156295776367, | |
| "step": 13170 | |
| }, | |
| { | |
| "epoch": 27.630531552762502, | |
| "learning_rate": 0.0003143125230799694, | |
| "loss": 2.7753381729125977, | |
| "step": 13180 | |
| }, | |
| { | |
| "epoch": 27.65147944488086, | |
| "learning_rate": 0.00031419335245082134, | |
| "loss": 2.812895393371582, | |
| "step": 13190 | |
| }, | |
| { | |
| "epoch": 27.672427336999213, | |
| "learning_rate": 0.00031407431726850375, | |
| "loss": 2.8747041702270506, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 27.69337522911757, | |
| "learning_rate": 0.00031395541727663413, | |
| "loss": 2.8663089752197264, | |
| "step": 13210 | |
| }, | |
| { | |
| "epoch": 27.714323121235925, | |
| "learning_rate": 0.0003138366522195088, | |
| "loss": 2.8993961334228517, | |
| "step": 13220 | |
| }, | |
| { | |
| "epoch": 27.735271013354282, | |
| "learning_rate": 0.0003137180218421011, | |
| "loss": 2.9394744873046874, | |
| "step": 13230 | |
| }, | |
| { | |
| "epoch": 27.756218905472636, | |
| "learning_rate": 0.0003135995258900582, | |
| "loss": 2.8471282958984374, | |
| "step": 13240 | |
| }, | |
| { | |
| "epoch": 27.777166797590993, | |
| "learning_rate": 0.0003134811641096994, | |
| "loss": 2.7851446151733397, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 27.798114689709347, | |
| "learning_rate": 0.00031336293624801393, | |
| "loss": 2.819938850402832, | |
| "step": 13260 | |
| }, | |
| { | |
| "epoch": 27.819062581827705, | |
| "learning_rate": 0.00031324484205265824, | |
| "loss": 2.8013900756835937, | |
| "step": 13270 | |
| }, | |
| { | |
| "epoch": 27.84001047394606, | |
| "learning_rate": 0.000313126881271954, | |
| "loss": 2.850057601928711, | |
| "step": 13280 | |
| }, | |
| { | |
| "epoch": 27.860958366064416, | |
| "learning_rate": 0.0003130090536548859, | |
| "loss": 2.7631250381469727, | |
| "step": 13290 | |
| }, | |
| { | |
| "epoch": 27.88190625818277, | |
| "learning_rate": 0.00031289135895109924, | |
| "loss": 2.8360868453979493, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 27.902854150301128, | |
| "learning_rate": 0.00031277379691089786, | |
| "loss": 2.804159927368164, | |
| "step": 13310 | |
| }, | |
| { | |
| "epoch": 27.92380204241948, | |
| "learning_rate": 0.00031265636728524174, | |
| "loss": 2.8401294708251954, | |
| "step": 13320 | |
| }, | |
| { | |
| "epoch": 27.94474993453784, | |
| "learning_rate": 0.000312539069825745, | |
| "loss": 2.850791168212891, | |
| "step": 13330 | |
| }, | |
| { | |
| "epoch": 27.965697826656193, | |
| "learning_rate": 0.00031242190428467325, | |
| "loss": 2.862323188781738, | |
| "step": 13340 | |
| }, | |
| { | |
| "epoch": 27.986645718774547, | |
| "learning_rate": 0.0003123048704149423, | |
| "loss": 2.8848134994506838, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 28.008379156847344, | |
| "learning_rate": 0.0003121879679701147, | |
| "loss": 2.9553651809692383, | |
| "step": 13360 | |
| }, | |
| { | |
| "epoch": 28.029327048965698, | |
| "learning_rate": 0.00031207119670439884, | |
| "loss": 2.893220138549805, | |
| "step": 13370 | |
| }, | |
| { | |
| "epoch": 28.050274941084055, | |
| "learning_rate": 0.00031195455637264574, | |
| "loss": 2.8204929351806642, | |
| "step": 13380 | |
| }, | |
| { | |
| "epoch": 28.07122283320241, | |
| "learning_rate": 0.00031183804673034756, | |
| "loss": 2.8356761932373047, | |
| "step": 13390 | |
| }, | |
| { | |
| "epoch": 28.092170725320763, | |
| "learning_rate": 0.0003117216675336353, | |
| "loss": 2.800448989868164, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 28.11311861743912, | |
| "learning_rate": 0.00031160541853927627, | |
| "loss": 2.8977182388305662, | |
| "step": 13410 | |
| }, | |
| { | |
| "epoch": 28.134066509557474, | |
| "learning_rate": 0.0003114892995046725, | |
| "loss": 2.8017560958862306, | |
| "step": 13420 | |
| }, | |
| { | |
| "epoch": 28.15501440167583, | |
| "learning_rate": 0.00031137331018785835, | |
| "loss": 2.7457189559936523, | |
| "step": 13430 | |
| }, | |
| { | |
| "epoch": 28.175962293794186, | |
| "learning_rate": 0.00031125745034749834, | |
| "loss": 2.8290485382080077, | |
| "step": 13440 | |
| }, | |
| { | |
| "epoch": 28.196910185912543, | |
| "learning_rate": 0.00031114171974288516, | |
| "loss": 2.8317813873291016, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 28.217858078030897, | |
| "learning_rate": 0.00031102611813393753, | |
| "loss": 2.7843399047851562, | |
| "step": 13460 | |
| }, | |
| { | |
| "epoch": 28.238805970149254, | |
| "learning_rate": 0.0003109106452811981, | |
| "loss": 2.8257192611694335, | |
| "step": 13470 | |
| }, | |
| { | |
| "epoch": 28.25975386226761, | |
| "learning_rate": 0.00031079530094583135, | |
| "loss": 2.8432809829711916, | |
| "step": 13480 | |
| }, | |
| { | |
| "epoch": 28.280701754385966, | |
| "learning_rate": 0.0003106800848896216, | |
| "loss": 2.882096862792969, | |
| "step": 13490 | |
| }, | |
| { | |
| "epoch": 28.30164964650432, | |
| "learning_rate": 0.0003105649968749708, | |
| "loss": 2.87137508392334, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 28.322597538622677, | |
| "learning_rate": 0.0003104500366648965, | |
| "loss": 2.8303447723388673, | |
| "step": 13510 | |
| }, | |
| { | |
| "epoch": 28.34354543074103, | |
| "learning_rate": 0.0003103352040230302, | |
| "loss": 2.955478477478027, | |
| "step": 13520 | |
| }, | |
| { | |
| "epoch": 28.36449332285939, | |
| "learning_rate": 0.00031022049871361445, | |
| "loss": 2.7974782943725587, | |
| "step": 13530 | |
| }, | |
| { | |
| "epoch": 28.385441214977742, | |
| "learning_rate": 0.0003101059205015017, | |
| "loss": 2.882868766784668, | |
| "step": 13540 | |
| }, | |
| { | |
| "epoch": 28.4063891070961, | |
| "learning_rate": 0.0003099914691521518, | |
| "loss": 2.9435708999633787, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 28.427336999214454, | |
| "learning_rate": 0.00030987714443163, | |
| "loss": 2.8506664276123046, | |
| "step": 13560 | |
| }, | |
| { | |
| "epoch": 28.44828489133281, | |
| "learning_rate": 0.00030976294610660516, | |
| "loss": 2.8492944717407225, | |
| "step": 13570 | |
| }, | |
| { | |
| "epoch": 28.469232783451165, | |
| "learning_rate": 0.00030964887394434754, | |
| "loss": 2.8658618927001953, | |
| "step": 13580 | |
| }, | |
| { | |
| "epoch": 28.49018067556952, | |
| "learning_rate": 0.000309534927712727, | |
| "loss": 2.8701282501220704, | |
| "step": 13590 | |
| }, | |
| { | |
| "epoch": 28.511128567687877, | |
| "learning_rate": 0.0003094211071802107, | |
| "loss": 2.8161798477172852, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 28.53207645980623, | |
| "learning_rate": 0.00030930741211586155, | |
| "loss": 2.768409538269043, | |
| "step": 13610 | |
| }, | |
| { | |
| "epoch": 28.553024351924588, | |
| "learning_rate": 0.0003091938422893358, | |
| "loss": 2.84487361907959, | |
| "step": 13620 | |
| }, | |
| { | |
| "epoch": 28.573972244042942, | |
| "learning_rate": 0.00030908039747088155, | |
| "loss": 2.8081539154052733, | |
| "step": 13630 | |
| }, | |
| { | |
| "epoch": 28.5949201361613, | |
| "learning_rate": 0.00030896707743133635, | |
| "loss": 2.8049062728881835, | |
| "step": 13640 | |
| }, | |
| { | |
| "epoch": 28.615868028279653, | |
| "learning_rate": 0.0003088538819421255, | |
| "loss": 2.8450254440307616, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 28.63681592039801, | |
| "learning_rate": 0.00030874081077526003, | |
| "loss": 2.8079158782958986, | |
| "step": 13660 | |
| }, | |
| { | |
| "epoch": 28.657763812516365, | |
| "learning_rate": 0.00030862786370333505, | |
| "loss": 2.8801244735717773, | |
| "step": 13670 | |
| }, | |
| { | |
| "epoch": 28.678711704634722, | |
| "learning_rate": 0.00030851504049952727, | |
| "loss": 2.8432153701782226, | |
| "step": 13680 | |
| }, | |
| { | |
| "epoch": 28.699659596753076, | |
| "learning_rate": 0.00030840234093759347, | |
| "loss": 2.913180923461914, | |
| "step": 13690 | |
| }, | |
| { | |
| "epoch": 28.720607488871433, | |
| "learning_rate": 0.0003082897647918688, | |
| "loss": 2.857924461364746, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 28.741555380989787, | |
| "learning_rate": 0.0003081773118372642, | |
| "loss": 2.7912296295166015, | |
| "step": 13710 | |
| }, | |
| { | |
| "epoch": 28.762503273108145, | |
| "learning_rate": 0.00030806498184926523, | |
| "loss": 2.8504261016845702, | |
| "step": 13720 | |
| }, | |
| { | |
| "epoch": 28.7834511652265, | |
| "learning_rate": 0.0003079527746039298, | |
| "loss": 2.8378028869628906, | |
| "step": 13730 | |
| }, | |
| { | |
| "epoch": 28.804399057344856, | |
| "learning_rate": 0.00030784068987788624, | |
| "loss": 2.803904914855957, | |
| "step": 13740 | |
| }, | |
| { | |
| "epoch": 28.82534694946321, | |
| "learning_rate": 0.00030772872744833183, | |
| "loss": 2.839299774169922, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 28.846294841581567, | |
| "learning_rate": 0.00030761688709303036, | |
| "loss": 2.7884681701660154, | |
| "step": 13760 | |
| }, | |
| { | |
| "epoch": 28.86724273369992, | |
| "learning_rate": 0.0003075051685903109, | |
| "loss": 2.88138427734375, | |
| "step": 13770 | |
| }, | |
| { | |
| "epoch": 28.888190625818275, | |
| "learning_rate": 0.00030739357171906536, | |
| "loss": 2.81328125, | |
| "step": 13780 | |
| }, | |
| { | |
| "epoch": 28.909138517936633, | |
| "learning_rate": 0.0003072820962587471, | |
| "loss": 2.8141046524047852, | |
| "step": 13790 | |
| }, | |
| { | |
| "epoch": 28.930086410054987, | |
| "learning_rate": 0.00030717074198936904, | |
| "loss": 2.7672204971313477, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 28.951034302173344, | |
| "learning_rate": 0.0003070595086915015, | |
| "loss": 2.8781991958618165, | |
| "step": 13810 | |
| }, | |
| { | |
| "epoch": 28.971982194291698, | |
| "learning_rate": 0.00030694839614627076, | |
| "loss": 2.7781099319458007, | |
| "step": 13820 | |
| }, | |
| { | |
| "epoch": 28.992930086410055, | |
| "learning_rate": 0.0003068374041353571, | |
| "loss": 2.879766082763672, | |
| "step": 13830 | |
| }, | |
| { | |
| "epoch": 29.01466352448285, | |
| "learning_rate": 0.000306726532440993, | |
| "loss": 2.8991397857666015, | |
| "step": 13840 | |
| }, | |
| { | |
| "epoch": 29.035611416601206, | |
| "learning_rate": 0.0003066157808459613, | |
| "loss": 2.8512521743774415, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 29.05655930871956, | |
| "learning_rate": 0.0003065051491335936, | |
| "loss": 2.833390235900879, | |
| "step": 13860 | |
| }, | |
| { | |
| "epoch": 29.077507200837914, | |
| "learning_rate": 0.0003063946370877681, | |
| "loss": 2.8554765701293947, | |
| "step": 13870 | |
| }, | |
| { | |
| "epoch": 29.09845509295627, | |
| "learning_rate": 0.0003062842444929085, | |
| "loss": 2.7805418014526366, | |
| "step": 13880 | |
| }, | |
| { | |
| "epoch": 29.119402985074625, | |
| "learning_rate": 0.00030617397113398125, | |
| "loss": 3.09820671081543, | |
| "step": 13890 | |
| }, | |
| { | |
| "epoch": 29.140350877192983, | |
| "learning_rate": 0.00030606381679649483, | |
| "loss": 2.900446128845215, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 29.161298769311337, | |
| "learning_rate": 0.00030595378126649727, | |
| "loss": 2.852696418762207, | |
| "step": 13910 | |
| }, | |
| { | |
| "epoch": 29.182246661429694, | |
| "learning_rate": 0.0003058438643305747, | |
| "loss": 3.0798343658447265, | |
| "step": 13920 | |
| }, | |
| { | |
| "epoch": 29.203194553548048, | |
| "learning_rate": 0.00030573406577584955, | |
| "loss": 2.8329389572143553, | |
| "step": 13930 | |
| }, | |
| { | |
| "epoch": 29.224142445666406, | |
| "learning_rate": 0.000305624385389979, | |
| "loss": 2.7638198852539064, | |
| "step": 13940 | |
| }, | |
| { | |
| "epoch": 29.24509033778476, | |
| "learning_rate": 0.0003055148229611527, | |
| "loss": 2.7774702072143556, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 29.266038229903117, | |
| "learning_rate": 0.00030540537827809176, | |
| "loss": 2.884586524963379, | |
| "step": 13960 | |
| }, | |
| { | |
| "epoch": 29.28698612202147, | |
| "learning_rate": 0.0003052960511300467, | |
| "loss": 2.858045196533203, | |
| "step": 13970 | |
| }, | |
| { | |
| "epoch": 29.30793401413983, | |
| "learning_rate": 0.0003051868413067956, | |
| "loss": 2.8505125045776367, | |
| "step": 13980 | |
| }, | |
| { | |
| "epoch": 29.328881906258182, | |
| "learning_rate": 0.00030507774859864277, | |
| "loss": 2.840318298339844, | |
| "step": 13990 | |
| }, | |
| { | |
| "epoch": 29.34982979837654, | |
| "learning_rate": 0.0003049687727964166, | |
| "loss": 2.871793746948242, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 29.370777690494894, | |
| "learning_rate": 0.00030485991369146834, | |
| "loss": 2.814739990234375, | |
| "step": 14010 | |
| }, | |
| { | |
| "epoch": 29.39172558261325, | |
| "learning_rate": 0.00030475117107567015, | |
| "loss": 2.8241125106811524, | |
| "step": 14020 | |
| }, | |
| { | |
| "epoch": 29.412673474731605, | |
| "learning_rate": 0.0003046425447414135, | |
| "loss": 2.802973747253418, | |
| "step": 14030 | |
| }, | |
| { | |
| "epoch": 29.433621366849962, | |
| "learning_rate": 0.0003045340344816073, | |
| "loss": 2.829861068725586, | |
| "step": 14040 | |
| }, | |
| { | |
| "epoch": 29.454569258968316, | |
| "learning_rate": 0.0003044256400896769, | |
| "loss": 2.823344612121582, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 29.47551715108667, | |
| "learning_rate": 0.0003043173613595614, | |
| "loss": 2.811284065246582, | |
| "step": 14060 | |
| }, | |
| { | |
| "epoch": 29.496465043205028, | |
| "learning_rate": 0.0003042091980857131, | |
| "loss": 2.8590465545654298, | |
| "step": 14070 | |
| }, | |
| { | |
| "epoch": 29.51741293532338, | |
| "learning_rate": 0.0003041011500630949, | |
| "loss": 2.8229595184326173, | |
| "step": 14080 | |
| }, | |
| { | |
| "epoch": 29.53836082744174, | |
| "learning_rate": 0.00030399321708717947, | |
| "loss": 2.8343103408813475, | |
| "step": 14090 | |
| }, | |
| { | |
| "epoch": 29.559308719560093, | |
| "learning_rate": 0.00030388539895394697, | |
| "loss": 2.804738235473633, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 29.58025661167845, | |
| "learning_rate": 0.00030377769545988394, | |
| "loss": 2.8719600677490233, | |
| "step": 14110 | |
| }, | |
| { | |
| "epoch": 29.601204503796804, | |
| "learning_rate": 0.00030367010640198143, | |
| "loss": 2.777914810180664, | |
| "step": 14120 | |
| }, | |
| { | |
| "epoch": 29.622152395915162, | |
| "learning_rate": 0.0003035626315777333, | |
| "loss": 2.837109375, | |
| "step": 14130 | |
| }, | |
| { | |
| "epoch": 29.643100288033516, | |
| "learning_rate": 0.00030345527078513493, | |
| "loss": 2.8141595840454103, | |
| "step": 14140 | |
| }, | |
| { | |
| "epoch": 29.664048180151873, | |
| "learning_rate": 0.0003033480238226813, | |
| "loss": 2.8648092269897463, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 29.684996072270227, | |
| "learning_rate": 0.0003032408904893656, | |
| "loss": 2.7934087753295898, | |
| "step": 14160 | |
| }, | |
| { | |
| "epoch": 29.705943964388585, | |
| "learning_rate": 0.00030313387058467756, | |
| "loss": 2.834004783630371, | |
| "step": 14170 | |
| }, | |
| { | |
| "epoch": 29.72689185650694, | |
| "learning_rate": 0.0003030269639086021, | |
| "loss": 2.8099668502807615, | |
| "step": 14180 | |
| }, | |
| { | |
| "epoch": 29.747839748625296, | |
| "learning_rate": 0.0003029201702616173, | |
| "loss": 2.830114555358887, | |
| "step": 14190 | |
| }, | |
| { | |
| "epoch": 29.76878764074365, | |
| "learning_rate": 0.0003028134894446933, | |
| "loss": 2.780957794189453, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 29.789735532862007, | |
| "learning_rate": 0.00030270692125929034, | |
| "loss": 2.829334831237793, | |
| "step": 14210 | |
| }, | |
| { | |
| "epoch": 29.81068342498036, | |
| "learning_rate": 0.00030260046550735763, | |
| "loss": 2.840847969055176, | |
| "step": 14220 | |
| }, | |
| { | |
| "epoch": 29.83163131709872, | |
| "learning_rate": 0.0003024941219913316, | |
| "loss": 2.855925369262695, | |
| "step": 14230 | |
| }, | |
| { | |
| "epoch": 29.852579209217073, | |
| "learning_rate": 0.00030238789051413416, | |
| "loss": 2.8478092193603515, | |
| "step": 14240 | |
| }, | |
| { | |
| "epoch": 29.873527101335426, | |
| "learning_rate": 0.00030228177087917153, | |
| "loss": 2.8140996932983398, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 29.894474993453784, | |
| "learning_rate": 0.00030217576289033235, | |
| "loss": 2.803069496154785, | |
| "step": 14260 | |
| }, | |
| { | |
| "epoch": 29.915422885572138, | |
| "learning_rate": 0.00030206986635198654, | |
| "loss": 2.7434965133666993, | |
| "step": 14270 | |
| }, | |
| { | |
| "epoch": 29.936370777690495, | |
| "learning_rate": 0.00030196408106898356, | |
| "loss": 2.859099006652832, | |
| "step": 14280 | |
| }, | |
| { | |
| "epoch": 29.95731866980885, | |
| "learning_rate": 0.0003018584068466507, | |
| "loss": 2.9088722229003907, | |
| "step": 14290 | |
| }, | |
| { | |
| "epoch": 29.978266561927207, | |
| "learning_rate": 0.0003017528434907922, | |
| "loss": 2.7880224227905273, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 29.99921445404556, | |
| "learning_rate": 0.00030164739080768704, | |
| "loss": 2.8113405227661135, | |
| "step": 14310 | |
| }, | |
| { | |
| "epoch": 30.020947892118354, | |
| "learning_rate": 0.0003015420486040879, | |
| "loss": 2.893621826171875, | |
| "step": 14320 | |
| }, | |
| { | |
| "epoch": 30.04189578423671, | |
| "learning_rate": 0.00030143681668721935, | |
| "loss": 2.8216567993164063, | |
| "step": 14330 | |
| }, | |
| { | |
| "epoch": 30.062843676355065, | |
| "learning_rate": 0.00030133169486477694, | |
| "loss": 2.8155281066894533, | |
| "step": 14340 | |
| }, | |
| { | |
| "epoch": 30.083791568473423, | |
| "learning_rate": 0.0003012266829449249, | |
| "loss": 2.8872468948364256, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 30.104739460591777, | |
| "learning_rate": 0.00030112178073629544, | |
| "loss": 2.820456886291504, | |
| "step": 14360 | |
| }, | |
| { | |
| "epoch": 30.125687352710134, | |
| "learning_rate": 0.0003010169880479867, | |
| "loss": 2.869482231140137, | |
| "step": 14370 | |
| }, | |
| { | |
| "epoch": 30.146635244828488, | |
| "learning_rate": 0.0003009123046895618, | |
| "loss": 2.8011972427368166, | |
| "step": 14380 | |
| }, | |
| { | |
| "epoch": 30.167583136946845, | |
| "learning_rate": 0.00030080773047104687, | |
| "loss": 2.8537342071533205, | |
| "step": 14390 | |
| }, | |
| { | |
| "epoch": 30.1885310290652, | |
| "learning_rate": 0.0003007032652029301, | |
| "loss": 2.808944892883301, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 30.209478921183557, | |
| "learning_rate": 0.00030059890869615983, | |
| "loss": 2.833651542663574, | |
| "step": 14410 | |
| }, | |
| { | |
| "epoch": 30.23042681330191, | |
| "learning_rate": 0.0003004946607621435, | |
| "loss": 2.860894203186035, | |
| "step": 14420 | |
| }, | |
| { | |
| "epoch": 30.251374705420268, | |
| "learning_rate": 0.0003003905212127461, | |
| "loss": 2.835972213745117, | |
| "step": 14430 | |
| }, | |
| { | |
| "epoch": 30.272322597538622, | |
| "learning_rate": 0.00030028648986028843, | |
| "loss": 2.857589912414551, | |
| "step": 14440 | |
| }, | |
| { | |
| "epoch": 30.29327048965698, | |
| "learning_rate": 0.00030018256651754633, | |
| "loss": 2.845281219482422, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 30.314218381775333, | |
| "learning_rate": 0.00030007875099774864, | |
| "loss": 2.7922155380249025, | |
| "step": 14460 | |
| }, | |
| { | |
| "epoch": 30.33516627389369, | |
| "learning_rate": 0.0002999750431145761, | |
| "loss": 2.846644973754883, | |
| "step": 14470 | |
| }, | |
| { | |
| "epoch": 30.356114166012045, | |
| "learning_rate": 0.0002998714426821599, | |
| "loss": 2.83693904876709, | |
| "step": 14480 | |
| }, | |
| { | |
| "epoch": 30.377062058130402, | |
| "learning_rate": 0.00029976794951508027, | |
| "loss": 2.8328250885009765, | |
| "step": 14490 | |
| }, | |
| { | |
| "epoch": 30.398009950248756, | |
| "learning_rate": 0.00029966456342836505, | |
| "loss": 2.8287097930908205, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 30.41895784236711, | |
| "learning_rate": 0.0002995612842374884, | |
| "loss": 2.818513298034668, | |
| "step": 14510 | |
| }, | |
| { | |
| "epoch": 30.439905734485468, | |
| "learning_rate": 0.0002994581117583693, | |
| "loss": 2.804762077331543, | |
| "step": 14520 | |
| }, | |
| { | |
| "epoch": 30.46085362660382, | |
| "learning_rate": 0.00029935504580737006, | |
| "loss": 2.8560808181762694, | |
| "step": 14530 | |
| }, | |
| { | |
| "epoch": 30.48180151872218, | |
| "learning_rate": 0.00029925208620129546, | |
| "loss": 2.7961631774902345, | |
| "step": 14540 | |
| }, | |
| { | |
| "epoch": 30.502749410840533, | |
| "learning_rate": 0.0002991492327573909, | |
| "loss": 2.8281347274780275, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 30.52369730295889, | |
| "learning_rate": 0.0002990464852933409, | |
| "loss": 2.813071060180664, | |
| "step": 14560 | |
| }, | |
| { | |
| "epoch": 30.544645195077244, | |
| "learning_rate": 0.0002989438436272684, | |
| "loss": 2.765872001647949, | |
| "step": 14570 | |
| }, | |
| { | |
| "epoch": 30.5655930871956, | |
| "learning_rate": 0.00029884130757773275, | |
| "loss": 2.7835336685180665, | |
| "step": 14580 | |
| }, | |
| { | |
| "epoch": 30.586540979313956, | |
| "learning_rate": 0.0002987388769637288, | |
| "loss": 2.855548286437988, | |
| "step": 14590 | |
| }, | |
| { | |
| "epoch": 30.607488871432313, | |
| "learning_rate": 0.00029863655160468534, | |
| "loss": 2.804723358154297, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 30.628436763550667, | |
| "learning_rate": 0.0002985343313204637, | |
| "loss": 2.8737287521362305, | |
| "step": 14610 | |
| }, | |
| { | |
| "epoch": 30.649384655669024, | |
| "learning_rate": 0.0002984322159313568, | |
| "loss": 2.871350860595703, | |
| "step": 14620 | |
| }, | |
| { | |
| "epoch": 30.67033254778738, | |
| "learning_rate": 0.00029833020525808714, | |
| "loss": 2.780613327026367, | |
| "step": 14630 | |
| }, | |
| { | |
| "epoch": 30.691280439905736, | |
| "learning_rate": 0.00029822829912180636, | |
| "loss": 2.8216100692749024, | |
| "step": 14640 | |
| }, | |
| { | |
| "epoch": 30.71222833202409, | |
| "learning_rate": 0.0002981264973440931, | |
| "loss": 2.789328956604004, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 30.733176224142447, | |
| "learning_rate": 0.00029802479974695223, | |
| "loss": 2.7879051208496093, | |
| "step": 14660 | |
| }, | |
| { | |
| "epoch": 30.7541241162608, | |
| "learning_rate": 0.00029792320615281337, | |
| "loss": 2.762567710876465, | |
| "step": 14670 | |
| }, | |
| { | |
| "epoch": 30.77507200837916, | |
| "learning_rate": 0.00029782171638452937, | |
| "loss": 2.8410247802734374, | |
| "step": 14680 | |
| }, | |
| { | |
| "epoch": 30.796019900497512, | |
| "learning_rate": 0.0002977203302653755, | |
| "loss": 2.7910200119018556, | |
| "step": 14690 | |
| }, | |
| { | |
| "epoch": 30.81696779261587, | |
| "learning_rate": 0.0002976190476190476, | |
| "loss": 2.8463191986083984, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 30.837915684734224, | |
| "learning_rate": 0.0002975178682696613, | |
| "loss": 2.816401481628418, | |
| "step": 14710 | |
| }, | |
| { | |
| "epoch": 30.858863576852578, | |
| "learning_rate": 0.0002974167920417504, | |
| "loss": 2.850655746459961, | |
| "step": 14720 | |
| }, | |
| { | |
| "epoch": 30.879811468970935, | |
| "learning_rate": 0.00029731581876026557, | |
| "loss": 2.844277191162109, | |
| "step": 14730 | |
| }, | |
| { | |
| "epoch": 30.90075936108929, | |
| "learning_rate": 0.00029721494825057357, | |
| "loss": 2.8203685760498045, | |
| "step": 14740 | |
| }, | |
| { | |
| "epoch": 30.921707253207646, | |
| "learning_rate": 0.00029711418033845523, | |
| "loss": 2.848883628845215, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 30.942655145326, | |
| "learning_rate": 0.0002970135148501047, | |
| "loss": 2.7703632354736327, | |
| "step": 14760 | |
| }, | |
| { | |
| "epoch": 30.963603037444358, | |
| "learning_rate": 0.00029691295161212816, | |
| "loss": 2.8733938217163084, | |
| "step": 14770 | |
| }, | |
| { | |
| "epoch": 30.98455092956271, | |
| "learning_rate": 0.0002968124904515423, | |
| "loss": 2.8676376342773438, | |
| "step": 14780 | |
| }, | |
| { | |
| "epoch": 31.006284367635505, | |
| "learning_rate": 0.00029671213119577346, | |
| "loss": 2.8960426330566404, | |
| "step": 14790 | |
| }, | |
| { | |
| "epoch": 31.027232259753863, | |
| "learning_rate": 0.00029661187367265593, | |
| "loss": 2.8205034255981447, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 31.048180151872216, | |
| "learning_rate": 0.0002965117177104311, | |
| "loss": 2.8493398666381835, | |
| "step": 14810 | |
| }, | |
| { | |
| "epoch": 31.069128043990574, | |
| "learning_rate": 0.0002964116631377459, | |
| "loss": 2.808573913574219, | |
| "step": 14820 | |
| }, | |
| { | |
| "epoch": 31.090075936108928, | |
| "learning_rate": 0.000296311709783652, | |
| "loss": 2.770844078063965, | |
| "step": 14830 | |
| }, | |
| { | |
| "epoch": 31.111023828227285, | |
| "learning_rate": 0.00029621185747760406, | |
| "loss": 2.7819324493408204, | |
| "step": 14840 | |
| }, | |
| { | |
| "epoch": 31.13197172034564, | |
| "learning_rate": 0.0002961121060494589, | |
| "loss": 2.7976245880126953, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 31.152919612463997, | |
| "learning_rate": 0.00029601245532947417, | |
| "loss": 2.8540115356445312, | |
| "step": 14860 | |
| }, | |
| { | |
| "epoch": 31.17386750458235, | |
| "learning_rate": 0.0002959129051483069, | |
| "loss": 2.7655929565429687, | |
| "step": 14870 | |
| }, | |
| { | |
| "epoch": 31.194815396700708, | |
| "learning_rate": 0.00029581345533701285, | |
| "loss": 2.847081184387207, | |
| "step": 14880 | |
| }, | |
| { | |
| "epoch": 31.215763288819062, | |
| "learning_rate": 0.0002957141057270448, | |
| "loss": 2.82701416015625, | |
| "step": 14890 | |
| }, | |
| { | |
| "epoch": 31.23671118093742, | |
| "learning_rate": 0.0002956148561502513, | |
| "loss": 2.8076833724975585, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 31.257659073055773, | |
| "learning_rate": 0.00029551570643887603, | |
| "loss": 2.7729957580566404, | |
| "step": 14910 | |
| }, | |
| { | |
| "epoch": 31.27860696517413, | |
| "learning_rate": 0.00029541665642555606, | |
| "loss": 2.8175632476806642, | |
| "step": 14920 | |
| }, | |
| { | |
| "epoch": 31.299554857292485, | |
| "learning_rate": 0.00029531770594332096, | |
| "loss": 2.781933403015137, | |
| "step": 14930 | |
| }, | |
| { | |
| "epoch": 31.320502749410842, | |
| "learning_rate": 0.0002952188548255915, | |
| "loss": 2.826693534851074, | |
| "step": 14940 | |
| }, | |
| { | |
| "epoch": 31.341450641529196, | |
| "learning_rate": 0.00029512010290617854, | |
| "loss": 2.7952367782592775, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 31.36239853364755, | |
| "learning_rate": 0.0002950214500192816, | |
| "loss": 2.7863574981689454, | |
| "step": 14960 | |
| }, | |
| { | |
| "epoch": 31.383346425765907, | |
| "learning_rate": 0.00029492289599948834, | |
| "loss": 2.8061588287353514, | |
| "step": 14970 | |
| }, | |
| { | |
| "epoch": 31.40429431788426, | |
| "learning_rate": 0.0002948244406817725, | |
| "loss": 2.8176244735717773, | |
| "step": 14980 | |
| }, | |
| { | |
| "epoch": 31.42524221000262, | |
| "learning_rate": 0.00029472608390149343, | |
| "loss": 2.8314136505126952, | |
| "step": 14990 | |
| }, | |
| { | |
| "epoch": 31.446190102120973, | |
| "learning_rate": 0.00029462782549439473, | |
| "loss": 2.802597999572754, | |
| "step": 15000 | |
| } | |
| ], | |
| "max_steps": 15000, | |
| "num_train_epochs": 32, | |
| "total_flos": 4124974225514526720, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |