{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 4.933333333333334, "eval_steps": 150, "global_step": 335, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.014814814814814815, "grad_norm": 24.245065689086914, "learning_rate": 0.0, "loss": 1.5986, "step": 1 }, { "epoch": 0.02962962962962963, "grad_norm": 24.59937286376953, "learning_rate": 1.965616322328226e-06, "loss": 1.5387, "step": 2 }, { "epoch": 0.044444444444444446, "grad_norm": 22.017257690429688, "learning_rate": 3.1154281616956676e-06, "loss": 1.4677, "step": 3 }, { "epoch": 0.05925925925925926, "grad_norm": 19.979198455810547, "learning_rate": 3.931232644656452e-06, "loss": 1.3971, "step": 4 }, { "epoch": 0.07407407407407407, "grad_norm": 10.355873107910156, "learning_rate": 4.5640197625830816e-06, "loss": 1.1965, "step": 5 }, { "epoch": 0.08888888888888889, "grad_norm": 7.20529842376709, "learning_rate": 5.081044484023894e-06, "loss": 1.0327, "step": 6 }, { "epoch": 0.1037037037037037, "grad_norm": 6.543303489685059, "learning_rate": 5.518182657364912e-06, "loss": 1.0091, "step": 7 }, { "epoch": 0.11851851851851852, "grad_norm": 5.638514041900635, "learning_rate": 5.896848966984678e-06, "loss": 0.8685, "step": 8 }, { "epoch": 0.13333333333333333, "grad_norm": 4.796297073364258, "learning_rate": 6.230856323391335e-06, "loss": 0.816, "step": 9 }, { "epoch": 0.14814814814814814, "grad_norm": 6.3762431144714355, "learning_rate": 6.5296360849113085e-06, "loss": 0.8679, "step": 10 }, { "epoch": 0.16296296296296298, "grad_norm": 4.701241493225098, "learning_rate": 6.7999152555718276e-06, "loss": 0.7973, "step": 11 }, { "epoch": 0.17777777777777778, "grad_norm": 4.334161281585693, "learning_rate": 7.04666080635212e-06, "loss": 0.6289, "step": 12 }, { "epoch": 0.1925925925925926, "grad_norm": 2.9507179260253906, "learning_rate": 7.273644709769792e-06, "loss": 0.7202, "step": 13 }, { "epoch": 0.2074074074074074, "grad_norm": 3.3083744049072266, "learning_rate": 7.483798979693137e-06, "loss": 0.6864, "step": 14 }, { "epoch": 0.2222222222222222, "grad_norm": 3.1791484355926514, "learning_rate": 7.679447924278749e-06, "loss": 0.7458, "step": 15 }, { "epoch": 0.23703703703703705, "grad_norm": 2.6846699714660645, "learning_rate": 7.862465289312904e-06, "loss": 0.6543, "step": 16 }, { "epoch": 0.2518518518518518, "grad_norm": 3.1359522342681885, "learning_rate": 8.034383677671775e-06, "loss": 0.531, "step": 17 }, { "epoch": 0.26666666666666666, "grad_norm": 3.0577006340026855, "learning_rate": 8.19647264571956e-06, "loss": 0.6345, "step": 18 }, { "epoch": 0.2814814814814815, "grad_norm": 3.066653251647949, "learning_rate": 8.349795656491867e-06, "loss": 0.5942, "step": 19 }, { "epoch": 0.2962962962962963, "grad_norm": 3.3706552982330322, "learning_rate": 8.495252407239533e-06, "loss": 0.638, "step": 20 }, { "epoch": 0.3111111111111111, "grad_norm": 2.55362606048584, "learning_rate": 8.633610819060579e-06, "loss": 0.602, "step": 21 }, { "epoch": 0.32592592592592595, "grad_norm": 2.864042282104492, "learning_rate": 8.765531577900054e-06, "loss": 0.6849, "step": 22 }, { "epoch": 0.34074074074074073, "grad_norm": 2.357649326324463, "learning_rate": 8.891587215888663e-06, "loss": 0.5194, "step": 23 }, { "epoch": 0.35555555555555557, "grad_norm": 2.659815549850464, "learning_rate": 9.012277128680346e-06, "loss": 0.4647, "step": 24 }, { "epoch": 0.37037037037037035, "grad_norm": 3.5252552032470703, "learning_rate": 9.128039525166163e-06, "loss": 0.6621, "step": 25 }, { "epoch": 0.3851851851851852, "grad_norm": 2.833387613296509, "learning_rate": 9.239261032098019e-06, "loss": 0.5758, "step": 26 }, { "epoch": 0.4, "grad_norm": 2.5673093795776367, "learning_rate": 9.346284485087002e-06, "loss": 0.4742, "step": 27 }, { "epoch": 0.4148148148148148, "grad_norm": 3.8754830360412598, "learning_rate": 9.449415302021363e-06, "loss": 0.5632, "step": 28 }, { "epoch": 0.42962962962962964, "grad_norm": 2.8832461833953857, "learning_rate": 9.548926737583076e-06, "loss": 0.5039, "step": 29 }, { "epoch": 0.4444444444444444, "grad_norm": 2.340471029281616, "learning_rate": 9.645064246606975e-06, "loss": 0.4832, "step": 30 }, { "epoch": 0.45925925925925926, "grad_norm": 3.4840638637542725, "learning_rate": 9.738049131714717e-06, "loss": 0.6216, "step": 31 }, { "epoch": 0.4740740740740741, "grad_norm": 1.927802562713623, "learning_rate": 9.828081611641132e-06, "loss": 0.4316, "step": 32 }, { "epoch": 0.4888888888888889, "grad_norm": 1.9146533012390137, "learning_rate": 9.915343417267494e-06, "loss": 0.4261, "step": 33 }, { "epoch": 0.5037037037037037, "grad_norm": 2.636239767074585, "learning_rate": 1e-05, "loss": 0.4445, "step": 34 }, { "epoch": 0.5185185185185185, "grad_norm": 2.7775421142578125, "learning_rate": 1e-05, "loss": 0.5895, "step": 35 }, { "epoch": 0.5333333333333333, "grad_norm": 3.3831987380981445, "learning_rate": 9.966777408637874e-06, "loss": 0.5247, "step": 36 }, { "epoch": 0.5481481481481482, "grad_norm": 2.883329391479492, "learning_rate": 9.933554817275748e-06, "loss": 0.4712, "step": 37 }, { "epoch": 0.562962962962963, "grad_norm": 2.908726215362549, "learning_rate": 9.900332225913623e-06, "loss": 0.4991, "step": 38 }, { "epoch": 0.5777777777777777, "grad_norm": 2.307657480239868, "learning_rate": 9.867109634551495e-06, "loss": 0.4174, "step": 39 }, { "epoch": 0.5925925925925926, "grad_norm": 2.3641955852508545, "learning_rate": 9.83388704318937e-06, "loss": 0.5014, "step": 40 }, { "epoch": 0.6074074074074074, "grad_norm": 2.850147247314453, "learning_rate": 9.800664451827243e-06, "loss": 0.5025, "step": 41 }, { "epoch": 0.6222222222222222, "grad_norm": 2.9469687938690186, "learning_rate": 9.767441860465117e-06, "loss": 0.4127, "step": 42 }, { "epoch": 0.6370370370370371, "grad_norm": 2.0221896171569824, "learning_rate": 9.734219269102992e-06, "loss": 0.4793, "step": 43 }, { "epoch": 0.6518518518518519, "grad_norm": 2.2807424068450928, "learning_rate": 9.700996677740865e-06, "loss": 0.4363, "step": 44 }, { "epoch": 0.6666666666666666, "grad_norm": 1.9812514781951904, "learning_rate": 9.66777408637874e-06, "loss": 0.472, "step": 45 }, { "epoch": 0.6814814814814815, "grad_norm": 2.3653407096862793, "learning_rate": 9.634551495016612e-06, "loss": 0.4747, "step": 46 }, { "epoch": 0.6962962962962963, "grad_norm": 1.875173807144165, "learning_rate": 9.601328903654485e-06, "loss": 0.4466, "step": 47 }, { "epoch": 0.7111111111111111, "grad_norm": 2.5644772052764893, "learning_rate": 9.56810631229236e-06, "loss": 0.4908, "step": 48 }, { "epoch": 0.725925925925926, "grad_norm": 2.5701663494110107, "learning_rate": 9.534883720930234e-06, "loss": 0.5709, "step": 49 }, { "epoch": 0.7407407407407407, "grad_norm": 3.4693243503570557, "learning_rate": 9.501661129568107e-06, "loss": 0.4436, "step": 50 }, { "epoch": 0.7555555555555555, "grad_norm": 1.7804033756256104, "learning_rate": 9.468438538205981e-06, "loss": 0.5096, "step": 51 }, { "epoch": 0.7703703703703704, "grad_norm": 2.171896457672119, "learning_rate": 9.435215946843854e-06, "loss": 0.4063, "step": 52 }, { "epoch": 0.7851851851851852, "grad_norm": 2.830214500427246, "learning_rate": 9.401993355481728e-06, "loss": 0.6981, "step": 53 }, { "epoch": 0.8, "grad_norm": 3.3130273818969727, "learning_rate": 9.368770764119603e-06, "loss": 0.5553, "step": 54 }, { "epoch": 0.8148148148148148, "grad_norm": 2.5522916316986084, "learning_rate": 9.335548172757476e-06, "loss": 0.5471, "step": 55 }, { "epoch": 0.8296296296296296, "grad_norm": 1.8053007125854492, "learning_rate": 9.30232558139535e-06, "loss": 0.4449, "step": 56 }, { "epoch": 0.8444444444444444, "grad_norm": 2.2297627925872803, "learning_rate": 9.269102990033223e-06, "loss": 0.4782, "step": 57 }, { "epoch": 0.8592592592592593, "grad_norm": 2.388592004776001, "learning_rate": 9.235880398671098e-06, "loss": 0.4333, "step": 58 }, { "epoch": 0.8740740740740741, "grad_norm": 2.2160580158233643, "learning_rate": 9.20265780730897e-06, "loss": 0.397, "step": 59 }, { "epoch": 0.8888888888888888, "grad_norm": 2.1153295040130615, "learning_rate": 9.169435215946845e-06, "loss": 0.4976, "step": 60 }, { "epoch": 0.9037037037037037, "grad_norm": 2.1571428775787354, "learning_rate": 9.136212624584718e-06, "loss": 0.5009, "step": 61 }, { "epoch": 0.9185185185185185, "grad_norm": 1.9368093013763428, "learning_rate": 9.102990033222592e-06, "loss": 0.4225, "step": 62 }, { "epoch": 0.9333333333333333, "grad_norm": 1.793280839920044, "learning_rate": 9.069767441860465e-06, "loss": 0.4407, "step": 63 }, { "epoch": 0.9481481481481482, "grad_norm": 1.7864512205123901, "learning_rate": 9.03654485049834e-06, "loss": 0.481, "step": 64 }, { "epoch": 0.9629629629629629, "grad_norm": 2.982163429260254, "learning_rate": 9.003322259136214e-06, "loss": 0.5735, "step": 65 }, { "epoch": 0.9777777777777777, "grad_norm": 2.1836278438568115, "learning_rate": 8.970099667774087e-06, "loss": 0.4532, "step": 66 }, { "epoch": 0.9925925925925926, "grad_norm": 2.726743221282959, "learning_rate": 8.93687707641196e-06, "loss": 0.5093, "step": 67 }, { "epoch": 1.0, "grad_norm": 2.726743221282959, "learning_rate": 8.93687707641196e-06, "loss": 0.4299, "step": 68 }, { "epoch": 1.0148148148148148, "grad_norm": 2.7928483486175537, "learning_rate": 8.903654485049834e-06, "loss": 0.3207, "step": 69 }, { "epoch": 1.0296296296296297, "grad_norm": 1.4182758331298828, "learning_rate": 8.870431893687709e-06, "loss": 0.4154, "step": 70 }, { "epoch": 1.0444444444444445, "grad_norm": 2.5209672451019287, "learning_rate": 8.837209302325582e-06, "loss": 0.4017, "step": 71 }, { "epoch": 1.0592592592592593, "grad_norm": 1.528294324874878, "learning_rate": 8.803986710963456e-06, "loss": 0.3515, "step": 72 }, { "epoch": 1.074074074074074, "grad_norm": 2.7746169567108154, "learning_rate": 8.770764119601329e-06, "loss": 0.4338, "step": 73 }, { "epoch": 1.0888888888888888, "grad_norm": 1.9682352542877197, "learning_rate": 8.737541528239203e-06, "loss": 0.4154, "step": 74 }, { "epoch": 1.1037037037037036, "grad_norm": 2.2559337615966797, "learning_rate": 8.704318936877078e-06, "loss": 0.368, "step": 75 }, { "epoch": 1.1185185185185185, "grad_norm": 1.3220306634902954, "learning_rate": 8.67109634551495e-06, "loss": 0.3142, "step": 76 }, { "epoch": 1.1333333333333333, "grad_norm": 2.455972671508789, "learning_rate": 8.637873754152825e-06, "loss": 0.4777, "step": 77 }, { "epoch": 1.1481481481481481, "grad_norm": 2.32432222366333, "learning_rate": 8.604651162790698e-06, "loss": 0.4585, "step": 78 }, { "epoch": 1.162962962962963, "grad_norm": 2.2095751762390137, "learning_rate": 8.571428571428571e-06, "loss": 0.4513, "step": 79 }, { "epoch": 1.1777777777777778, "grad_norm": 2.29775071144104, "learning_rate": 8.538205980066447e-06, "loss": 0.3315, "step": 80 }, { "epoch": 1.1925925925925926, "grad_norm": 2.390573740005493, "learning_rate": 8.50498338870432e-06, "loss": 0.4328, "step": 81 }, { "epoch": 1.2074074074074075, "grad_norm": 2.1682214736938477, "learning_rate": 8.471760797342193e-06, "loss": 0.3708, "step": 82 }, { "epoch": 1.2222222222222223, "grad_norm": 2.305809736251831, "learning_rate": 8.438538205980067e-06, "loss": 0.333, "step": 83 }, { "epoch": 1.237037037037037, "grad_norm": 1.9154317378997803, "learning_rate": 8.40531561461794e-06, "loss": 0.3514, "step": 84 }, { "epoch": 1.2518518518518518, "grad_norm": 2.1949751377105713, "learning_rate": 8.372093023255815e-06, "loss": 0.3423, "step": 85 }, { "epoch": 1.2666666666666666, "grad_norm": 1.4846103191375732, "learning_rate": 8.338870431893689e-06, "loss": 0.3935, "step": 86 }, { "epoch": 1.2814814814814814, "grad_norm": 2.3148951530456543, "learning_rate": 8.305647840531562e-06, "loss": 0.3683, "step": 87 }, { "epoch": 1.2962962962962963, "grad_norm": 1.7715332508087158, "learning_rate": 8.272425249169436e-06, "loss": 0.3415, "step": 88 }, { "epoch": 1.3111111111111111, "grad_norm": 1.6804319620132446, "learning_rate": 8.23920265780731e-06, "loss": 0.4075, "step": 89 }, { "epoch": 1.325925925925926, "grad_norm": 1.976668119430542, "learning_rate": 8.205980066445184e-06, "loss": 0.295, "step": 90 }, { "epoch": 1.3407407407407408, "grad_norm": 1.8023229837417603, "learning_rate": 8.172757475083057e-06, "loss": 0.389, "step": 91 }, { "epoch": 1.3555555555555556, "grad_norm": 1.7759190797805786, "learning_rate": 8.139534883720931e-06, "loss": 0.3517, "step": 92 }, { "epoch": 1.3703703703703702, "grad_norm": 1.3482650518417358, "learning_rate": 8.106312292358804e-06, "loss": 0.3606, "step": 93 }, { "epoch": 1.3851851851851853, "grad_norm": 2.5421998500823975, "learning_rate": 8.073089700996678e-06, "loss": 0.5241, "step": 94 }, { "epoch": 1.4, "grad_norm": 3.181865692138672, "learning_rate": 8.039867109634553e-06, "loss": 0.375, "step": 95 }, { "epoch": 1.4148148148148147, "grad_norm": 1.7613658905029297, "learning_rate": 8.006644518272426e-06, "loss": 0.4339, "step": 96 }, { "epoch": 1.4296296296296296, "grad_norm": 1.4631813764572144, "learning_rate": 7.9734219269103e-06, "loss": 0.3762, "step": 97 }, { "epoch": 1.4444444444444444, "grad_norm": 1.71363365650177, "learning_rate": 7.940199335548173e-06, "loss": 0.3224, "step": 98 }, { "epoch": 1.4592592592592593, "grad_norm": 2.1795740127563477, "learning_rate": 7.906976744186048e-06, "loss": 0.3816, "step": 99 }, { "epoch": 1.474074074074074, "grad_norm": 1.5822690725326538, "learning_rate": 7.873754152823922e-06, "loss": 0.4523, "step": 100 }, { "epoch": 1.488888888888889, "grad_norm": 1.7851877212524414, "learning_rate": 7.840531561461795e-06, "loss": 0.3665, "step": 101 }, { "epoch": 1.5037037037037035, "grad_norm": 1.671633243560791, "learning_rate": 7.807308970099668e-06, "loss": 0.3217, "step": 102 }, { "epoch": 1.5185185185185186, "grad_norm": 2.666489601135254, "learning_rate": 7.774086378737542e-06, "loss": 0.4801, "step": 103 }, { "epoch": 1.5333333333333332, "grad_norm": 1.3500992059707642, "learning_rate": 7.740863787375415e-06, "loss": 0.3407, "step": 104 }, { "epoch": 1.5481481481481483, "grad_norm": 1.4269295930862427, "learning_rate": 7.70764119601329e-06, "loss": 0.3082, "step": 105 }, { "epoch": 1.5629629629629629, "grad_norm": 1.4882391691207886, "learning_rate": 7.674418604651164e-06, "loss": 0.3497, "step": 106 }, { "epoch": 1.5777777777777777, "grad_norm": 1.4180535078048706, "learning_rate": 7.641196013289037e-06, "loss": 0.3963, "step": 107 }, { "epoch": 1.5925925925925926, "grad_norm": 1.4838252067565918, "learning_rate": 7.6079734219269106e-06, "loss": 0.3956, "step": 108 }, { "epoch": 1.6074074074074074, "grad_norm": 2.0815930366516113, "learning_rate": 7.574750830564784e-06, "loss": 0.4667, "step": 109 }, { "epoch": 1.6222222222222222, "grad_norm": 1.80450439453125, "learning_rate": 7.541528239202659e-06, "loss": 0.5329, "step": 110 }, { "epoch": 1.637037037037037, "grad_norm": 2.5666868686676025, "learning_rate": 7.508305647840532e-06, "loss": 0.5433, "step": 111 }, { "epoch": 1.651851851851852, "grad_norm": 1.733236312866211, "learning_rate": 7.475083056478406e-06, "loss": 0.4304, "step": 112 }, { "epoch": 1.6666666666666665, "grad_norm": 2.059544086456299, "learning_rate": 7.44186046511628e-06, "loss": 0.3601, "step": 113 }, { "epoch": 1.6814814814814816, "grad_norm": 1.8904452323913574, "learning_rate": 7.408637873754153e-06, "loss": 0.385, "step": 114 }, { "epoch": 1.6962962962962962, "grad_norm": 1.3712379932403564, "learning_rate": 7.375415282392027e-06, "loss": 0.2972, "step": 115 }, { "epoch": 1.7111111111111112, "grad_norm": 1.5603693723678589, "learning_rate": 7.342192691029902e-06, "loss": 0.3953, "step": 116 }, { "epoch": 1.7259259259259259, "grad_norm": 1.4126758575439453, "learning_rate": 7.308970099667775e-06, "loss": 0.3582, "step": 117 }, { "epoch": 1.7407407407407407, "grad_norm": 2.0345335006713867, "learning_rate": 7.275747508305648e-06, "loss": 0.5015, "step": 118 }, { "epoch": 1.7555555555555555, "grad_norm": 2.6806693077087402, "learning_rate": 7.242524916943522e-06, "loss": 0.4006, "step": 119 }, { "epoch": 1.7703703703703704, "grad_norm": 1.6245332956314087, "learning_rate": 7.209302325581395e-06, "loss": 0.4188, "step": 120 }, { "epoch": 1.7851851851851852, "grad_norm": 1.6593953371047974, "learning_rate": 7.17607973421927e-06, "loss": 0.3694, "step": 121 }, { "epoch": 1.8, "grad_norm": 1.595201849937439, "learning_rate": 7.1428571428571436e-06, "loss": 0.2838, "step": 122 }, { "epoch": 1.8148148148148149, "grad_norm": 1.921108603477478, "learning_rate": 7.109634551495017e-06, "loss": 0.3044, "step": 123 }, { "epoch": 1.8296296296296295, "grad_norm": 1.5152816772460938, "learning_rate": 7.076411960132891e-06, "loss": 0.3705, "step": 124 }, { "epoch": 1.8444444444444446, "grad_norm": 1.7763299942016602, "learning_rate": 7.0431893687707646e-06, "loss": 0.3935, "step": 125 }, { "epoch": 1.8592592592592592, "grad_norm": 1.3404872417449951, "learning_rate": 7.009966777408639e-06, "loss": 0.3951, "step": 126 }, { "epoch": 1.8740740740740742, "grad_norm": 1.7480947971343994, "learning_rate": 6.976744186046513e-06, "loss": 0.3694, "step": 127 }, { "epoch": 1.8888888888888888, "grad_norm": 1.3903143405914307, "learning_rate": 6.9435215946843855e-06, "loss": 0.3963, "step": 128 }, { "epoch": 1.9037037037037037, "grad_norm": 1.579624056816101, "learning_rate": 6.910299003322259e-06, "loss": 0.4147, "step": 129 }, { "epoch": 1.9185185185185185, "grad_norm": 1.6633076667785645, "learning_rate": 6.877076411960133e-06, "loss": 0.415, "step": 130 }, { "epoch": 1.9333333333333333, "grad_norm": 1.4897836446762085, "learning_rate": 6.843853820598007e-06, "loss": 0.3438, "step": 131 }, { "epoch": 1.9481481481481482, "grad_norm": 1.7405959367752075, "learning_rate": 6.810631229235881e-06, "loss": 0.3973, "step": 132 }, { "epoch": 1.9629629629629628, "grad_norm": 1.6084774732589722, "learning_rate": 6.777408637873755e-06, "loss": 0.3904, "step": 133 }, { "epoch": 1.9777777777777779, "grad_norm": 1.4988147020339966, "learning_rate": 6.744186046511628e-06, "loss": 0.4101, "step": 134 }, { "epoch": 1.9925925925925925, "grad_norm": 2.449335813522339, "learning_rate": 6.710963455149502e-06, "loss": 0.3886, "step": 135 }, { "epoch": 2.0, "grad_norm": 2.2566375732421875, "learning_rate": 6.6777408637873766e-06, "loss": 0.3651, "step": 136 }, { "epoch": 2.0148148148148146, "grad_norm": 1.5977928638458252, "learning_rate": 6.64451827242525e-06, "loss": 0.3709, "step": 137 }, { "epoch": 2.0296296296296297, "grad_norm": 2.1500344276428223, "learning_rate": 6.611295681063124e-06, "loss": 0.4124, "step": 138 }, { "epoch": 2.0444444444444443, "grad_norm": 1.3736720085144043, "learning_rate": 6.578073089700997e-06, "loss": 0.3334, "step": 139 }, { "epoch": 2.0592592592592593, "grad_norm": 1.6602332592010498, "learning_rate": 6.54485049833887e-06, "loss": 0.3443, "step": 140 }, { "epoch": 2.074074074074074, "grad_norm": 1.5972793102264404, "learning_rate": 6.511627906976745e-06, "loss": 0.3969, "step": 141 }, { "epoch": 2.088888888888889, "grad_norm": 1.5588147640228271, "learning_rate": 6.4784053156146185e-06, "loss": 0.2632, "step": 142 }, { "epoch": 2.1037037037037036, "grad_norm": 1.4944829940795898, "learning_rate": 6.445182724252492e-06, "loss": 0.3403, "step": 143 }, { "epoch": 2.1185185185185187, "grad_norm": 1.799228310585022, "learning_rate": 6.411960132890366e-06, "loss": 0.3669, "step": 144 }, { "epoch": 2.1333333333333333, "grad_norm": 1.5002415180206299, "learning_rate": 6.3787375415282395e-06, "loss": 0.3037, "step": 145 }, { "epoch": 2.148148148148148, "grad_norm": 1.5463218688964844, "learning_rate": 6.345514950166114e-06, "loss": 0.2999, "step": 146 }, { "epoch": 2.162962962962963, "grad_norm": 1.8339827060699463, "learning_rate": 6.312292358803988e-06, "loss": 0.3781, "step": 147 }, { "epoch": 2.1777777777777776, "grad_norm": 1.9693403244018555, "learning_rate": 6.279069767441861e-06, "loss": 0.3527, "step": 148 }, { "epoch": 2.1925925925925926, "grad_norm": 1.3716813325881958, "learning_rate": 6.245847176079734e-06, "loss": 0.3193, "step": 149 }, { "epoch": 2.2074074074074073, "grad_norm": 1.2879470586776733, "learning_rate": 6.212624584717608e-06, "loss": 0.3079, "step": 150 }, { "epoch": 2.2222222222222223, "grad_norm": 1.3284484148025513, "learning_rate": 6.179401993355482e-06, "loss": 0.3552, "step": 151 }, { "epoch": 2.237037037037037, "grad_norm": 1.1561126708984375, "learning_rate": 6.146179401993356e-06, "loss": 0.2879, "step": 152 }, { "epoch": 2.251851851851852, "grad_norm": 1.2936991453170776, "learning_rate": 6.11295681063123e-06, "loss": 0.2926, "step": 153 }, { "epoch": 2.2666666666666666, "grad_norm": 1.5603609085083008, "learning_rate": 6.079734219269103e-06, "loss": 0.3591, "step": 154 }, { "epoch": 2.2814814814814817, "grad_norm": 1.408933162689209, "learning_rate": 6.046511627906977e-06, "loss": 0.3313, "step": 155 }, { "epoch": 2.2962962962962963, "grad_norm": 1.9709101915359497, "learning_rate": 6.0132890365448515e-06, "loss": 0.3948, "step": 156 }, { "epoch": 2.311111111111111, "grad_norm": 1.5349321365356445, "learning_rate": 5.980066445182725e-06, "loss": 0.3227, "step": 157 }, { "epoch": 2.325925925925926, "grad_norm": 1.5960361957550049, "learning_rate": 5.946843853820599e-06, "loss": 0.3532, "step": 158 }, { "epoch": 2.3407407407407406, "grad_norm": 1.8431493043899536, "learning_rate": 5.9136212624584725e-06, "loss": 0.3621, "step": 159 }, { "epoch": 2.3555555555555556, "grad_norm": 1.875801920890808, "learning_rate": 5.880398671096345e-06, "loss": 0.4026, "step": 160 }, { "epoch": 2.3703703703703702, "grad_norm": 1.7672104835510254, "learning_rate": 5.847176079734221e-06, "loss": 0.3889, "step": 161 }, { "epoch": 2.3851851851851853, "grad_norm": 1.5263267755508423, "learning_rate": 5.8139534883720935e-06, "loss": 0.332, "step": 162 }, { "epoch": 2.4, "grad_norm": 1.624315857887268, "learning_rate": 5.780730897009967e-06, "loss": 0.3167, "step": 163 }, { "epoch": 2.414814814814815, "grad_norm": 1.9534611701965332, "learning_rate": 5.747508305647841e-06, "loss": 0.3733, "step": 164 }, { "epoch": 2.4296296296296296, "grad_norm": 1.6925545930862427, "learning_rate": 5.7142857142857145e-06, "loss": 0.3631, "step": 165 }, { "epoch": 2.4444444444444446, "grad_norm": 1.6101276874542236, "learning_rate": 5.681063122923588e-06, "loss": 0.3561, "step": 166 }, { "epoch": 2.4592592592592593, "grad_norm": 1.5481266975402832, "learning_rate": 5.647840531561463e-06, "loss": 0.349, "step": 167 }, { "epoch": 2.474074074074074, "grad_norm": 1.6006097793579102, "learning_rate": 5.614617940199336e-06, "loss": 0.3862, "step": 168 }, { "epoch": 2.488888888888889, "grad_norm": 1.4279389381408691, "learning_rate": 5.58139534883721e-06, "loss": 0.3236, "step": 169 }, { "epoch": 2.5037037037037035, "grad_norm": 1.2405160665512085, "learning_rate": 5.548172757475083e-06, "loss": 0.3013, "step": 170 }, { "epoch": 2.5185185185185186, "grad_norm": 1.3768985271453857, "learning_rate": 5.5149501661129565e-06, "loss": 0.3626, "step": 171 }, { "epoch": 2.533333333333333, "grad_norm": 1.6408584117889404, "learning_rate": 5.481727574750831e-06, "loss": 0.3196, "step": 172 }, { "epoch": 2.5481481481481483, "grad_norm": 1.6769694089889526, "learning_rate": 5.448504983388705e-06, "loss": 0.3694, "step": 173 }, { "epoch": 2.562962962962963, "grad_norm": 1.3804996013641357, "learning_rate": 5.415282392026578e-06, "loss": 0.3502, "step": 174 }, { "epoch": 2.5777777777777775, "grad_norm": 1.4021036624908447, "learning_rate": 5.382059800664452e-06, "loss": 0.3306, "step": 175 }, { "epoch": 2.5925925925925926, "grad_norm": 1.2684727907180786, "learning_rate": 5.348837209302326e-06, "loss": 0.3106, "step": 176 }, { "epoch": 2.6074074074074076, "grad_norm": 1.5925308465957642, "learning_rate": 5.3156146179402e-06, "loss": 0.4211, "step": 177 }, { "epoch": 2.6222222222222222, "grad_norm": 1.5546250343322754, "learning_rate": 5.282392026578074e-06, "loss": 0.3301, "step": 178 }, { "epoch": 2.637037037037037, "grad_norm": 1.4001022577285767, "learning_rate": 5.2491694352159475e-06, "loss": 0.3746, "step": 179 }, { "epoch": 2.651851851851852, "grad_norm": 1.4525564908981323, "learning_rate": 5.215946843853821e-06, "loss": 0.2681, "step": 180 }, { "epoch": 2.6666666666666665, "grad_norm": 1.27578604221344, "learning_rate": 5.182724252491694e-06, "loss": 0.3106, "step": 181 }, { "epoch": 2.6814814814814816, "grad_norm": 1.4694350957870483, "learning_rate": 5.149501661129569e-06, "loss": 0.3065, "step": 182 }, { "epoch": 2.696296296296296, "grad_norm": 1.3258838653564453, "learning_rate": 5.116279069767442e-06, "loss": 0.3233, "step": 183 }, { "epoch": 2.7111111111111112, "grad_norm": 1.1991218328475952, "learning_rate": 5.083056478405316e-06, "loss": 0.2786, "step": 184 }, { "epoch": 2.725925925925926, "grad_norm": 1.134046196937561, "learning_rate": 5.0498338870431895e-06, "loss": 0.2788, "step": 185 }, { "epoch": 2.7407407407407405, "grad_norm": 1.1837220191955566, "learning_rate": 5.016611295681063e-06, "loss": 0.2802, "step": 186 }, { "epoch": 2.7555555555555555, "grad_norm": 1.5560253858566284, "learning_rate": 4.983388704318937e-06, "loss": 0.3686, "step": 187 }, { "epoch": 2.7703703703703706, "grad_norm": 1.442020297050476, "learning_rate": 4.950166112956811e-06, "loss": 0.3648, "step": 188 }, { "epoch": 2.785185185185185, "grad_norm": 1.2866986989974976, "learning_rate": 4.916943521594685e-06, "loss": 0.335, "step": 189 }, { "epoch": 2.8, "grad_norm": 0.9921827912330627, "learning_rate": 4.883720930232559e-06, "loss": 0.3052, "step": 190 }, { "epoch": 2.814814814814815, "grad_norm": 1.3171212673187256, "learning_rate": 4.850498338870432e-06, "loss": 0.3365, "step": 191 }, { "epoch": 2.8296296296296295, "grad_norm": 1.202288031578064, "learning_rate": 4.817275747508306e-06, "loss": 0.2726, "step": 192 }, { "epoch": 2.8444444444444446, "grad_norm": 1.5031251907348633, "learning_rate": 4.78405315614618e-06, "loss": 0.366, "step": 193 }, { "epoch": 2.859259259259259, "grad_norm": 1.2141317129135132, "learning_rate": 4.750830564784053e-06, "loss": 0.2861, "step": 194 }, { "epoch": 2.8740740740740742, "grad_norm": 1.64356529712677, "learning_rate": 4.717607973421927e-06, "loss": 0.3415, "step": 195 }, { "epoch": 2.888888888888889, "grad_norm": 1.2897413969039917, "learning_rate": 4.6843853820598015e-06, "loss": 0.3334, "step": 196 }, { "epoch": 2.9037037037037035, "grad_norm": 1.192264199256897, "learning_rate": 4.651162790697675e-06, "loss": 0.2927, "step": 197 }, { "epoch": 2.9185185185185185, "grad_norm": 1.36125648021698, "learning_rate": 4.617940199335549e-06, "loss": 0.3129, "step": 198 }, { "epoch": 2.9333333333333336, "grad_norm": 1.3408805131912231, "learning_rate": 4.5847176079734225e-06, "loss": 0.3277, "step": 199 }, { "epoch": 2.948148148148148, "grad_norm": 1.3519293069839478, "learning_rate": 4.551495016611296e-06, "loss": 0.2881, "step": 200 }, { "epoch": 2.962962962962963, "grad_norm": 1.2503290176391602, "learning_rate": 4.51827242524917e-06, "loss": 0.3282, "step": 201 }, { "epoch": 2.977777777777778, "grad_norm": 1.357706069946289, "learning_rate": 4.4850498338870435e-06, "loss": 0.3319, "step": 202 }, { "epoch": 2.9925925925925925, "grad_norm": 1.173913598060608, "learning_rate": 4.451827242524917e-06, "loss": 0.3099, "step": 203 }, { "epoch": 3.0, "grad_norm": 1.173913598060608, "learning_rate": 4.451827242524917e-06, "loss": 0.2291, "step": 204 }, { "epoch": 3.0148148148148146, "grad_norm": 1.3551340103149414, "learning_rate": 4.418604651162791e-06, "loss": 0.3068, "step": 205 }, { "epoch": 3.0296296296296297, "grad_norm": 1.3206164836883545, "learning_rate": 4.3853820598006645e-06, "loss": 0.3416, "step": 206 }, { "epoch": 3.0444444444444443, "grad_norm": 1.417013168334961, "learning_rate": 4.352159468438539e-06, "loss": 0.2771, "step": 207 }, { "epoch": 3.0592592592592593, "grad_norm": 1.115646243095398, "learning_rate": 4.318936877076413e-06, "loss": 0.2828, "step": 208 }, { "epoch": 3.074074074074074, "grad_norm": 1.4733177423477173, "learning_rate": 4.2857142857142855e-06, "loss": 0.3038, "step": 209 }, { "epoch": 3.088888888888889, "grad_norm": 1.3604916334152222, "learning_rate": 4.25249169435216e-06, "loss": 0.3128, "step": 210 }, { "epoch": 3.1037037037037036, "grad_norm": 1.6025428771972656, "learning_rate": 4.219269102990034e-06, "loss": 0.2808, "step": 211 }, { "epoch": 3.1185185185185187, "grad_norm": 1.5683448314666748, "learning_rate": 4.186046511627907e-06, "loss": 0.3538, "step": 212 }, { "epoch": 3.1333333333333333, "grad_norm": 1.5745620727539062, "learning_rate": 4.152823920265781e-06, "loss": 0.3071, "step": 213 }, { "epoch": 3.148148148148148, "grad_norm": 1.26665461063385, "learning_rate": 4.119601328903655e-06, "loss": 0.2568, "step": 214 }, { "epoch": 3.162962962962963, "grad_norm": 1.1559772491455078, "learning_rate": 4.086378737541528e-06, "loss": 0.2655, "step": 215 }, { "epoch": 3.1777777777777776, "grad_norm": 1.3667874336242676, "learning_rate": 4.053156146179402e-06, "loss": 0.2943, "step": 216 }, { "epoch": 3.1925925925925926, "grad_norm": 1.4076658487319946, "learning_rate": 4.0199335548172765e-06, "loss": 0.242, "step": 217 }, { "epoch": 3.2074074074074073, "grad_norm": 1.461408257484436, "learning_rate": 3.98671096345515e-06, "loss": 0.2792, "step": 218 }, { "epoch": 3.2222222222222223, "grad_norm": 1.3601031303405762, "learning_rate": 3.953488372093024e-06, "loss": 0.2788, "step": 219 }, { "epoch": 3.237037037037037, "grad_norm": 1.9735394716262817, "learning_rate": 3.9202657807308975e-06, "loss": 0.3367, "step": 220 }, { "epoch": 3.251851851851852, "grad_norm": 1.5413546562194824, "learning_rate": 3.887043189368771e-06, "loss": 0.2887, "step": 221 }, { "epoch": 3.2666666666666666, "grad_norm": 1.5698870420455933, "learning_rate": 3.853820598006645e-06, "loss": 0.3121, "step": 222 }, { "epoch": 3.2814814814814817, "grad_norm": 1.4351308345794678, "learning_rate": 3.8205980066445185e-06, "loss": 0.2859, "step": 223 }, { "epoch": 3.2962962962962963, "grad_norm": 1.5523098707199097, "learning_rate": 3.787375415282392e-06, "loss": 0.3496, "step": 224 }, { "epoch": 3.311111111111111, "grad_norm": 1.5026061534881592, "learning_rate": 3.754152823920266e-06, "loss": 0.2744, "step": 225 }, { "epoch": 3.325925925925926, "grad_norm": 1.565753698348999, "learning_rate": 3.72093023255814e-06, "loss": 0.2761, "step": 226 }, { "epoch": 3.3407407407407406, "grad_norm": 1.514609456062317, "learning_rate": 3.6877076411960135e-06, "loss": 0.2385, "step": 227 }, { "epoch": 3.3555555555555556, "grad_norm": 1.233588695526123, "learning_rate": 3.6544850498338876e-06, "loss": 0.2849, "step": 228 }, { "epoch": 3.3703703703703702, "grad_norm": 1.3447370529174805, "learning_rate": 3.621262458471761e-06, "loss": 0.3197, "step": 229 }, { "epoch": 3.3851851851851853, "grad_norm": 1.3895092010498047, "learning_rate": 3.588039867109635e-06, "loss": 0.2869, "step": 230 }, { "epoch": 3.4, "grad_norm": 1.326200246810913, "learning_rate": 3.5548172757475086e-06, "loss": 0.2937, "step": 231 }, { "epoch": 3.414814814814815, "grad_norm": 1.2751388549804688, "learning_rate": 3.5215946843853823e-06, "loss": 0.2611, "step": 232 }, { "epoch": 3.4296296296296296, "grad_norm": 1.1317442655563354, "learning_rate": 3.4883720930232564e-06, "loss": 0.2733, "step": 233 }, { "epoch": 3.4444444444444446, "grad_norm": 1.3669071197509766, "learning_rate": 3.4551495016611296e-06, "loss": 0.2967, "step": 234 }, { "epoch": 3.4592592592592593, "grad_norm": 1.451213002204895, "learning_rate": 3.4219269102990037e-06, "loss": 0.3398, "step": 235 }, { "epoch": 3.474074074074074, "grad_norm": 2.680572748184204, "learning_rate": 3.3887043189368774e-06, "loss": 0.2973, "step": 236 }, { "epoch": 3.488888888888889, "grad_norm": 1.5603687763214111, "learning_rate": 3.355481727574751e-06, "loss": 0.2669, "step": 237 }, { "epoch": 3.5037037037037035, "grad_norm": 1.8472412824630737, "learning_rate": 3.322259136212625e-06, "loss": 0.2956, "step": 238 }, { "epoch": 3.5185185185185186, "grad_norm": 1.6677846908569336, "learning_rate": 3.2890365448504984e-06, "loss": 0.329, "step": 239 }, { "epoch": 3.533333333333333, "grad_norm": 1.8696357011795044, "learning_rate": 3.2558139534883724e-06, "loss": 0.2735, "step": 240 }, { "epoch": 3.5481481481481483, "grad_norm": 1.2237964868545532, "learning_rate": 3.222591362126246e-06, "loss": 0.3011, "step": 241 }, { "epoch": 3.562962962962963, "grad_norm": 1.2561728954315186, "learning_rate": 3.1893687707641198e-06, "loss": 0.2568, "step": 242 }, { "epoch": 3.5777777777777775, "grad_norm": 1.1599830389022827, "learning_rate": 3.156146179401994e-06, "loss": 0.3025, "step": 243 }, { "epoch": 3.5925925925925926, "grad_norm": 1.7549623250961304, "learning_rate": 3.122923588039867e-06, "loss": 0.3252, "step": 244 }, { "epoch": 3.6074074074074076, "grad_norm": 1.5652899742126465, "learning_rate": 3.089700996677741e-06, "loss": 0.341, "step": 245 }, { "epoch": 3.6222222222222222, "grad_norm": 1.7946059703826904, "learning_rate": 3.056478405315615e-06, "loss": 0.3244, "step": 246 }, { "epoch": 3.637037037037037, "grad_norm": 1.5564749240875244, "learning_rate": 3.0232558139534885e-06, "loss": 0.2686, "step": 247 }, { "epoch": 3.651851851851852, "grad_norm": 1.439971923828125, "learning_rate": 2.9900332225913626e-06, "loss": 0.3198, "step": 248 }, { "epoch": 3.6666666666666665, "grad_norm": 1.590828537940979, "learning_rate": 2.9568106312292363e-06, "loss": 0.2874, "step": 249 }, { "epoch": 3.6814814814814816, "grad_norm": 1.2883110046386719, "learning_rate": 2.9235880398671104e-06, "loss": 0.2475, "step": 250 }, { "epoch": 3.696296296296296, "grad_norm": 1.2013444900512695, "learning_rate": 2.8903654485049836e-06, "loss": 0.2911, "step": 251 }, { "epoch": 3.7111111111111112, "grad_norm": 1.5711296796798706, "learning_rate": 2.8571428571428573e-06, "loss": 0.3113, "step": 252 }, { "epoch": 3.725925925925926, "grad_norm": 1.5199639797210693, "learning_rate": 2.8239202657807313e-06, "loss": 0.2462, "step": 253 }, { "epoch": 3.7407407407407405, "grad_norm": 1.2884533405303955, "learning_rate": 2.790697674418605e-06, "loss": 0.245, "step": 254 }, { "epoch": 3.7555555555555555, "grad_norm": 1.4477325677871704, "learning_rate": 2.7574750830564782e-06, "loss": 0.2409, "step": 255 }, { "epoch": 3.7703703703703706, "grad_norm": 1.2022032737731934, "learning_rate": 2.7242524916943523e-06, "loss": 0.2806, "step": 256 }, { "epoch": 3.785185185185185, "grad_norm": 1.467168927192688, "learning_rate": 2.691029900332226e-06, "loss": 0.2866, "step": 257 }, { "epoch": 3.8, "grad_norm": 2.0435290336608887, "learning_rate": 2.6578073089701e-06, "loss": 0.3211, "step": 258 }, { "epoch": 3.814814814814815, "grad_norm": 1.2129874229431152, "learning_rate": 2.6245847176079738e-06, "loss": 0.2765, "step": 259 }, { "epoch": 3.8296296296296295, "grad_norm": 1.4887040853500366, "learning_rate": 2.591362126245847e-06, "loss": 0.282, "step": 260 }, { "epoch": 3.8444444444444446, "grad_norm": 1.3326512575149536, "learning_rate": 2.558139534883721e-06, "loss": 0.307, "step": 261 }, { "epoch": 3.859259259259259, "grad_norm": 1.2139685153961182, "learning_rate": 2.5249169435215947e-06, "loss": 0.3053, "step": 262 }, { "epoch": 3.8740740740740742, "grad_norm": 1.6648317575454712, "learning_rate": 2.4916943521594684e-06, "loss": 0.2859, "step": 263 }, { "epoch": 3.888888888888889, "grad_norm": 1.550366997718811, "learning_rate": 2.4584717607973425e-06, "loss": 0.3527, "step": 264 }, { "epoch": 3.9037037037037035, "grad_norm": 1.641121506690979, "learning_rate": 2.425249169435216e-06, "loss": 0.2735, "step": 265 }, { "epoch": 3.9185185185185185, "grad_norm": 1.244675874710083, "learning_rate": 2.39202657807309e-06, "loss": 0.3208, "step": 266 }, { "epoch": 3.9333333333333336, "grad_norm": 1.4783977270126343, "learning_rate": 2.3588039867109635e-06, "loss": 0.2869, "step": 267 }, { "epoch": 3.948148148148148, "grad_norm": 1.5731481313705444, "learning_rate": 2.3255813953488376e-06, "loss": 0.4189, "step": 268 }, { "epoch": 3.962962962962963, "grad_norm": 1.654645562171936, "learning_rate": 2.2923588039867112e-06, "loss": 0.2766, "step": 269 }, { "epoch": 3.977777777777778, "grad_norm": 1.6379601955413818, "learning_rate": 2.259136212624585e-06, "loss": 0.3227, "step": 270 }, { "epoch": 3.9925925925925925, "grad_norm": 1.3923341035842896, "learning_rate": 2.2259136212624586e-06, "loss": 0.3051, "step": 271 }, { "epoch": 4.0, "grad_norm": 1.9158486127853394, "learning_rate": 2.1926910299003322e-06, "loss": 0.2157, "step": 272 }, { "epoch": 4.014814814814815, "grad_norm": 1.2792537212371826, "learning_rate": 2.1594684385382063e-06, "loss": 0.2568, "step": 273 }, { "epoch": 4.029629629629629, "grad_norm": 1.1235581636428833, "learning_rate": 2.12624584717608e-06, "loss": 0.2664, "step": 274 }, { "epoch": 4.044444444444444, "grad_norm": 1.2936333417892456, "learning_rate": 2.0930232558139536e-06, "loss": 0.2966, "step": 275 }, { "epoch": 4.059259259259259, "grad_norm": 1.2000664472579956, "learning_rate": 2.0598006644518273e-06, "loss": 0.2487, "step": 276 }, { "epoch": 4.074074074074074, "grad_norm": 1.4936013221740723, "learning_rate": 2.026578073089701e-06, "loss": 0.283, "step": 277 }, { "epoch": 4.088888888888889, "grad_norm": 1.551798939704895, "learning_rate": 1.993355481727575e-06, "loss": 0.275, "step": 278 }, { "epoch": 4.103703703703704, "grad_norm": 1.8705427646636963, "learning_rate": 1.9601328903654487e-06, "loss": 0.3045, "step": 279 }, { "epoch": 4.118518518518519, "grad_norm": 1.5146915912628174, "learning_rate": 1.9269102990033224e-06, "loss": 0.2575, "step": 280 }, { "epoch": 4.133333333333334, "grad_norm": 2.1336557865142822, "learning_rate": 1.893687707641196e-06, "loss": 0.293, "step": 281 }, { "epoch": 4.148148148148148, "grad_norm": 1.673853874206543, "learning_rate": 1.86046511627907e-06, "loss": 0.2228, "step": 282 }, { "epoch": 4.162962962962963, "grad_norm": 1.4035648107528687, "learning_rate": 1.8272425249169438e-06, "loss": 0.236, "step": 283 }, { "epoch": 4.177777777777778, "grad_norm": 1.5520694255828857, "learning_rate": 1.7940199335548175e-06, "loss": 0.2381, "step": 284 }, { "epoch": 4.192592592592592, "grad_norm": 1.8831232786178589, "learning_rate": 1.7607973421926911e-06, "loss": 0.3056, "step": 285 }, { "epoch": 4.207407407407407, "grad_norm": 1.524223804473877, "learning_rate": 1.7275747508305648e-06, "loss": 0.2288, "step": 286 }, { "epoch": 4.222222222222222, "grad_norm": 1.7227911949157715, "learning_rate": 1.6943521594684387e-06, "loss": 0.3001, "step": 287 }, { "epoch": 4.237037037037037, "grad_norm": 1.6464650630950928, "learning_rate": 1.6611295681063126e-06, "loss": 0.2767, "step": 288 }, { "epoch": 4.2518518518518515, "grad_norm": 1.462727665901184, "learning_rate": 1.6279069767441862e-06, "loss": 0.2583, "step": 289 }, { "epoch": 4.266666666666667, "grad_norm": 1.3373020887374878, "learning_rate": 1.5946843853820599e-06, "loss": 0.1861, "step": 290 }, { "epoch": 4.281481481481482, "grad_norm": 1.3069071769714355, "learning_rate": 1.5614617940199335e-06, "loss": 0.2279, "step": 291 }, { "epoch": 4.296296296296296, "grad_norm": 2.0804059505462646, "learning_rate": 1.5282392026578074e-06, "loss": 0.3373, "step": 292 }, { "epoch": 4.311111111111111, "grad_norm": 1.5276566743850708, "learning_rate": 1.4950166112956813e-06, "loss": 0.2211, "step": 293 }, { "epoch": 4.325925925925926, "grad_norm": 1.7238777875900269, "learning_rate": 1.4617940199335552e-06, "loss": 0.2492, "step": 294 }, { "epoch": 4.340740740740741, "grad_norm": 1.7892625331878662, "learning_rate": 1.4285714285714286e-06, "loss": 0.2521, "step": 295 }, { "epoch": 4.355555555555555, "grad_norm": 1.83062744140625, "learning_rate": 1.3953488372093025e-06, "loss": 0.2445, "step": 296 }, { "epoch": 4.37037037037037, "grad_norm": 1.665940284729004, "learning_rate": 1.3621262458471762e-06, "loss": 0.2188, "step": 297 }, { "epoch": 4.385185185185185, "grad_norm": 1.7071560621261597, "learning_rate": 1.32890365448505e-06, "loss": 0.2294, "step": 298 }, { "epoch": 4.4, "grad_norm": 1.5902904272079468, "learning_rate": 1.2956810631229235e-06, "loss": 0.2443, "step": 299 }, { "epoch": 4.4148148148148145, "grad_norm": 1.6071012020111084, "learning_rate": 1.2624584717607974e-06, "loss": 0.2268, "step": 300 }, { "epoch": 4.42962962962963, "grad_norm": 1.481217384338379, "learning_rate": 1.2292358803986712e-06, "loss": 0.2245, "step": 301 }, { "epoch": 4.444444444444445, "grad_norm": 1.7338011264801025, "learning_rate": 1.196013289036545e-06, "loss": 0.2481, "step": 302 }, { "epoch": 4.459259259259259, "grad_norm": 1.83079195022583, "learning_rate": 1.1627906976744188e-06, "loss": 0.2771, "step": 303 }, { "epoch": 4.474074074074074, "grad_norm": 1.6256608963012695, "learning_rate": 1.1295681063122925e-06, "loss": 0.2418, "step": 304 }, { "epoch": 4.488888888888889, "grad_norm": 1.7368595600128174, "learning_rate": 1.0963455149501661e-06, "loss": 0.2325, "step": 305 }, { "epoch": 4.503703703703704, "grad_norm": 1.8986256122589111, "learning_rate": 1.06312292358804e-06, "loss": 0.2685, "step": 306 }, { "epoch": 4.518518518518518, "grad_norm": 1.9530152082443237, "learning_rate": 1.0299003322259137e-06, "loss": 0.2734, "step": 307 }, { "epoch": 4.533333333333333, "grad_norm": 2.2950448989868164, "learning_rate": 9.966777408637875e-07, "loss": 0.2775, "step": 308 }, { "epoch": 4.548148148148148, "grad_norm": 1.7885551452636719, "learning_rate": 9.634551495016612e-07, "loss": 0.2299, "step": 309 }, { "epoch": 4.562962962962963, "grad_norm": 1.8106021881103516, "learning_rate": 9.30232558139535e-07, "loss": 0.2665, "step": 310 }, { "epoch": 4.5777777777777775, "grad_norm": 1.7523531913757324, "learning_rate": 8.970099667774087e-07, "loss": 0.2514, "step": 311 }, { "epoch": 4.592592592592593, "grad_norm": 1.5606251955032349, "learning_rate": 8.637873754152824e-07, "loss": 0.216, "step": 312 }, { "epoch": 4.607407407407408, "grad_norm": 1.3957760334014893, "learning_rate": 8.305647840531563e-07, "loss": 0.2133, "step": 313 }, { "epoch": 4.622222222222222, "grad_norm": 1.4215636253356934, "learning_rate": 7.973421926910299e-07, "loss": 0.2159, "step": 314 }, { "epoch": 4.637037037037037, "grad_norm": 1.9139257669448853, "learning_rate": 7.641196013289037e-07, "loss": 0.2799, "step": 315 }, { "epoch": 4.651851851851852, "grad_norm": 1.5956664085388184, "learning_rate": 7.308970099667776e-07, "loss": 0.2386, "step": 316 }, { "epoch": 4.666666666666667, "grad_norm": 1.7055957317352295, "learning_rate": 6.976744186046513e-07, "loss": 0.2256, "step": 317 }, { "epoch": 4.681481481481481, "grad_norm": 2.2495598793029785, "learning_rate": 6.64451827242525e-07, "loss": 0.2795, "step": 318 }, { "epoch": 4.696296296296296, "grad_norm": 1.4281206130981445, "learning_rate": 6.312292358803987e-07, "loss": 0.2295, "step": 319 }, { "epoch": 4.711111111111111, "grad_norm": 1.7233028411865234, "learning_rate": 5.980066445182725e-07, "loss": 0.2586, "step": 320 }, { "epoch": 4.725925925925926, "grad_norm": 2.274329662322998, "learning_rate": 5.647840531561462e-07, "loss": 0.2477, "step": 321 }, { "epoch": 4.7407407407407405, "grad_norm": 1.6101205348968506, "learning_rate": 5.3156146179402e-07, "loss": 0.2304, "step": 322 }, { "epoch": 4.7555555555555555, "grad_norm": 1.9844639301300049, "learning_rate": 4.983388704318938e-07, "loss": 0.2888, "step": 323 }, { "epoch": 4.770370370370371, "grad_norm": 1.7177938222885132, "learning_rate": 4.651162790697675e-07, "loss": 0.2543, "step": 324 }, { "epoch": 4.785185185185185, "grad_norm": 1.7451181411743164, "learning_rate": 4.318936877076412e-07, "loss": 0.2362, "step": 325 }, { "epoch": 4.8, "grad_norm": 1.9255115985870361, "learning_rate": 3.9867109634551497e-07, "loss": 0.2689, "step": 326 }, { "epoch": 4.814814814814815, "grad_norm": 1.6680419445037842, "learning_rate": 3.654485049833888e-07, "loss": 0.2505, "step": 327 }, { "epoch": 4.82962962962963, "grad_norm": 1.8660964965820312, "learning_rate": 3.322259136212625e-07, "loss": 0.2671, "step": 328 }, { "epoch": 4.844444444444444, "grad_norm": 1.9916861057281494, "learning_rate": 2.9900332225913623e-07, "loss": 0.2519, "step": 329 }, { "epoch": 4.859259259259259, "grad_norm": 1.5888378620147705, "learning_rate": 2.6578073089701e-07, "loss": 0.2505, "step": 330 }, { "epoch": 4.874074074074074, "grad_norm": 1.654725193977356, "learning_rate": 2.3255813953488374e-07, "loss": 0.2466, "step": 331 }, { "epoch": 4.888888888888889, "grad_norm": 1.4593290090560913, "learning_rate": 1.9933554817275749e-07, "loss": 0.2039, "step": 332 }, { "epoch": 4.9037037037037035, "grad_norm": 1.5956478118896484, "learning_rate": 1.6611295681063126e-07, "loss": 0.2221, "step": 333 }, { "epoch": 4.9185185185185185, "grad_norm": 1.33909273147583, "learning_rate": 1.32890365448505e-07, "loss": 0.192, "step": 334 }, { "epoch": 4.933333333333334, "grad_norm": 1.7868711948394775, "learning_rate": 9.966777408637874e-08, "loss": 0.2477, "step": 335 } ], "logging_steps": 1, "max_steps": 335, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 150, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4351170969600.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }