VLAC / trainer_state.json
futurefantasy's picture
Upload folder using huggingface_hub
a6897ca verified
raw
history blame
180 kB
{
"best_metric": 0.65735477,
"best_model_checkpoint": "/cpfs04/shared/rlproject/zhangqi/model_garden/0709_intern2b_v7-1-part15-19-resize-decay/v0-20250710-072707/checkpoint-3000",
"epoch": 0.9599616015359386,
"eval_steps": 250,
"global_step": 3000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0003199872005119795,
"grad_norm": 0.06346331978668461,
"learning_rate": 2.0000000000000002e-07,
"loss": 0.4898327589035034,
"memory(GiB)": 37.39,
"step": 1,
"token_acc": 0.8589147286821706,
"train_speed(iter/s)": 0.017141
},
{
"epoch": 0.0015999360025598975,
"grad_norm": 0.06701772816278752,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.479672372341156,
"memory(GiB)": 57.22,
"step": 5,
"token_acc": 0.8759901666211418,
"train_speed(iter/s)": 0.029212
},
{
"epoch": 0.003199872005119795,
"grad_norm": 0.0577242460197625,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.4767899990081787,
"memory(GiB)": 57.22,
"step": 10,
"token_acc": 0.8920853213584058,
"train_speed(iter/s)": 0.030985
},
{
"epoch": 0.004799808007679693,
"grad_norm": 0.05378464715409511,
"learning_rate": 3e-06,
"loss": 0.46750531196594236,
"memory(GiB)": 57.22,
"step": 15,
"token_acc": 0.846137292877125,
"train_speed(iter/s)": 0.030693
},
{
"epoch": 0.00639974401023959,
"grad_norm": 0.05482863789544427,
"learning_rate": 4.000000000000001e-06,
"loss": 0.47017059326171873,
"memory(GiB)": 57.22,
"step": 20,
"token_acc": 0.8701638201463925,
"train_speed(iter/s)": 0.031839
},
{
"epoch": 0.007999680012799487,
"grad_norm": 0.05575025715512655,
"learning_rate": 5e-06,
"loss": 0.46974716186523435,
"memory(GiB)": 57.22,
"step": 25,
"token_acc": 0.8939419941650936,
"train_speed(iter/s)": 0.031203
},
{
"epoch": 0.009599616015359386,
"grad_norm": 0.05007200892621057,
"learning_rate": 6e-06,
"loss": 0.4700496196746826,
"memory(GiB)": 57.22,
"step": 30,
"token_acc": 0.872761844398669,
"train_speed(iter/s)": 0.030738
},
{
"epoch": 0.011199552017919284,
"grad_norm": 0.06284973201247342,
"learning_rate": 7e-06,
"loss": 0.477018404006958,
"memory(GiB)": 57.22,
"step": 35,
"token_acc": 0.8738273921200751,
"train_speed(iter/s)": 0.031491
},
{
"epoch": 0.01279948802047918,
"grad_norm": 0.051858307829802386,
"learning_rate": 8.000000000000001e-06,
"loss": 0.46584124565124513,
"memory(GiB)": 57.22,
"step": 40,
"token_acc": 0.8791312559017942,
"train_speed(iter/s)": 0.031035
},
{
"epoch": 0.014399424023039079,
"grad_norm": 0.05024484287076301,
"learning_rate": 9e-06,
"loss": 0.4685808658599854,
"memory(GiB)": 57.22,
"step": 45,
"token_acc": 0.8846325167037862,
"train_speed(iter/s)": 0.031503
},
{
"epoch": 0.015999360025598975,
"grad_norm": 0.06025612278216295,
"learning_rate": 1e-05,
"loss": 0.4697974681854248,
"memory(GiB)": 57.22,
"step": 50,
"token_acc": 0.854153041203401,
"train_speed(iter/s)": 0.031549
},
{
"epoch": 0.017599296028158875,
"grad_norm": 0.05171252494611451,
"learning_rate": 1.1000000000000001e-05,
"loss": 0.4622661113739014,
"memory(GiB)": 73.06,
"step": 55,
"token_acc": 0.8281821878812525,
"train_speed(iter/s)": 0.031181
},
{
"epoch": 0.01919923203071877,
"grad_norm": 0.06238294030934267,
"learning_rate": 1.2e-05,
"loss": 0.47000856399536134,
"memory(GiB)": 73.06,
"step": 60,
"token_acc": 0.8703465982028241,
"train_speed(iter/s)": 0.031617
},
{
"epoch": 0.020799168033278668,
"grad_norm": 0.05576663204628902,
"learning_rate": 1.3000000000000001e-05,
"loss": 0.4765446186065674,
"memory(GiB)": 73.06,
"step": 65,
"token_acc": 0.867019517036057,
"train_speed(iter/s)": 0.031138
},
{
"epoch": 0.022399104035838568,
"grad_norm": 0.054406694286476175,
"learning_rate": 1.4e-05,
"loss": 0.47959036827087403,
"memory(GiB)": 73.06,
"step": 70,
"token_acc": 0.8927940657011657,
"train_speed(iter/s)": 0.030807
},
{
"epoch": 0.023999040038398464,
"grad_norm": 0.05458413282309297,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.45742173194885255,
"memory(GiB)": 73.06,
"step": 75,
"token_acc": 0.8916037316748112,
"train_speed(iter/s)": 0.031204
},
{
"epoch": 0.02559897604095836,
"grad_norm": 0.05986601718141533,
"learning_rate": 1.6000000000000003e-05,
"loss": 0.46221466064453126,
"memory(GiB)": 73.06,
"step": 80,
"token_acc": 0.8513287970214405,
"train_speed(iter/s)": 0.030907
},
{
"epoch": 0.02719891204351826,
"grad_norm": 0.0610597448330617,
"learning_rate": 1.7e-05,
"loss": 0.4636178493499756,
"memory(GiB)": 73.06,
"step": 85,
"token_acc": 0.8755391250770179,
"train_speed(iter/s)": 0.030852
},
{
"epoch": 0.028798848046078157,
"grad_norm": 0.0537272421077024,
"learning_rate": 1.8e-05,
"loss": 0.4674212455749512,
"memory(GiB)": 73.06,
"step": 90,
"token_acc": 0.8330635370295427,
"train_speed(iter/s)": 0.03096
},
{
"epoch": 0.030398784048638054,
"grad_norm": 0.05874793146275866,
"learning_rate": 1.9e-05,
"loss": 0.459043550491333,
"memory(GiB)": 73.06,
"step": 95,
"token_acc": 0.8686557483379277,
"train_speed(iter/s)": 0.030744
},
{
"epoch": 0.03199872005119795,
"grad_norm": 0.0635697375742028,
"learning_rate": 2e-05,
"loss": 0.4696988582611084,
"memory(GiB)": 73.06,
"step": 100,
"token_acc": 0.8658624414836155,
"train_speed(iter/s)": 0.030967
},
{
"epoch": 0.03359865605375785,
"grad_norm": 0.05212754941929234,
"learning_rate": 1.9999865178850847e-05,
"loss": 0.47245235443115235,
"memory(GiB)": 73.06,
"step": 105,
"token_acc": 0.8773064185459707,
"train_speed(iter/s)": 0.030857
},
{
"epoch": 0.03519859205631775,
"grad_norm": 0.0584305183084298,
"learning_rate": 1.999946071903873e-05,
"loss": 0.4617309093475342,
"memory(GiB)": 73.06,
"step": 110,
"token_acc": 0.8634434872058634,
"train_speed(iter/s)": 0.030617
},
{
"epoch": 0.03679852805887764,
"grad_norm": 0.0594956422290189,
"learning_rate": 1.9998786631469602e-05,
"loss": 0.4737922191619873,
"memory(GiB)": 73.06,
"step": 115,
"token_acc": 0.8586429725363489,
"train_speed(iter/s)": 0.030917
},
{
"epoch": 0.03839846406143754,
"grad_norm": 0.05542732612540538,
"learning_rate": 1.999784293431971e-05,
"loss": 0.46903514862060547,
"memory(GiB)": 73.06,
"step": 120,
"token_acc": 0.8490970309152127,
"train_speed(iter/s)": 0.030707
},
{
"epoch": 0.03999840006399744,
"grad_norm": 0.06279644130355853,
"learning_rate": 1.9996629653035128e-05,
"loss": 0.47660508155822756,
"memory(GiB)": 73.06,
"step": 125,
"token_acc": 0.8797399783315276,
"train_speed(iter/s)": 0.03059
},
{
"epoch": 0.041598336066557336,
"grad_norm": 0.0589281224878424,
"learning_rate": 1.999514682033104e-05,
"loss": 0.47182955741882326,
"memory(GiB)": 73.06,
"step": 130,
"token_acc": 0.8641215106732348,
"train_speed(iter/s)": 0.030792
},
{
"epoch": 0.043198272069117236,
"grad_norm": 0.06201767062726238,
"learning_rate": 1.99933944761909e-05,
"loss": 0.4760914325714111,
"memory(GiB)": 73.06,
"step": 135,
"token_acc": 0.856929955290611,
"train_speed(iter/s)": 0.030572
},
{
"epoch": 0.044798208071677136,
"grad_norm": 0.06842512509591067,
"learning_rate": 1.999137266786531e-05,
"loss": 0.4673017501831055,
"memory(GiB)": 73.06,
"step": 140,
"token_acc": 0.8725868725868726,
"train_speed(iter/s)": 0.030583
},
{
"epoch": 0.04639814407423703,
"grad_norm": 0.06206860242026748,
"learning_rate": 1.998908144987078e-05,
"loss": 0.4650784969329834,
"memory(GiB)": 73.06,
"step": 145,
"token_acc": 0.8592991206195039,
"train_speed(iter/s)": 0.030589
},
{
"epoch": 0.04799808007679693,
"grad_norm": 0.07443005443822684,
"learning_rate": 1.9986520883988233e-05,
"loss": 0.46671695709228517,
"memory(GiB)": 73.06,
"step": 150,
"token_acc": 0.8422288642186165,
"train_speed(iter/s)": 0.030331
},
{
"epoch": 0.04959801607935683,
"grad_norm": 0.06105984389773331,
"learning_rate": 1.9983691039261358e-05,
"loss": 0.47356271743774414,
"memory(GiB)": 73.06,
"step": 155,
"token_acc": 0.8814565604591332,
"train_speed(iter/s)": 0.030476
},
{
"epoch": 0.05119795208191672,
"grad_norm": 0.08001789963695773,
"learning_rate": 1.998059199199474e-05,
"loss": 0.4710524559020996,
"memory(GiB)": 73.06,
"step": 160,
"token_acc": 0.8710629921259843,
"train_speed(iter/s)": 0.03042
},
{
"epoch": 0.05279788808447662,
"grad_norm": 0.05874066660842649,
"learning_rate": 1.9977223825751802e-05,
"loss": 0.46933708190917967,
"memory(GiB)": 73.06,
"step": 165,
"token_acc": 0.894688221709007,
"train_speed(iter/s)": 0.030265
},
{
"epoch": 0.05439782408703652,
"grad_norm": 0.06544476253513323,
"learning_rate": 1.997358663135255e-05,
"loss": 0.46097607612609864,
"memory(GiB)": 73.06,
"step": 170,
"token_acc": 0.8832285384319261,
"train_speed(iter/s)": 0.030391
},
{
"epoch": 0.055997760089596414,
"grad_norm": 0.052612734205809274,
"learning_rate": 1.9969680506871138e-05,
"loss": 0.4674376010894775,
"memory(GiB)": 73.06,
"step": 175,
"token_acc": 0.8822400558269365,
"train_speed(iter/s)": 0.030298
},
{
"epoch": 0.057597696092156314,
"grad_norm": 0.05831524002340728,
"learning_rate": 1.9965505557633188e-05,
"loss": 0.47021942138671874,
"memory(GiB)": 73.06,
"step": 180,
"token_acc": 0.8729036501150937,
"train_speed(iter/s)": 0.030222
},
{
"epoch": 0.059197632094716214,
"grad_norm": 0.05558463298745032,
"learning_rate": 1.9961061896213006e-05,
"loss": 0.4707474708557129,
"memory(GiB)": 73.06,
"step": 185,
"token_acc": 0.8507278220269157,
"train_speed(iter/s)": 0.030322
},
{
"epoch": 0.06079756809727611,
"grad_norm": 0.05835065786169716,
"learning_rate": 1.9956349642430494e-05,
"loss": 0.4792951583862305,
"memory(GiB)": 73.06,
"step": 190,
"token_acc": 0.8596896665566194,
"train_speed(iter/s)": 0.030221
},
{
"epoch": 0.06239750409983601,
"grad_norm": 0.055138234334700054,
"learning_rate": 1.9951368923347945e-05,
"loss": 0.4755210876464844,
"memory(GiB)": 73.06,
"step": 195,
"token_acc": 0.8501170960187353,
"train_speed(iter/s)": 0.030259
},
{
"epoch": 0.0639974401023959,
"grad_norm": 0.06763946666899583,
"learning_rate": 1.9946119873266615e-05,
"loss": 0.4560092926025391,
"memory(GiB)": 73.06,
"step": 200,
"token_acc": 0.8891928864569083,
"train_speed(iter/s)": 0.030234
},
{
"epoch": 0.0655973761049558,
"grad_norm": 0.056758126149685124,
"learning_rate": 1.9940602633723097e-05,
"loss": 0.470977258682251,
"memory(GiB)": 73.06,
"step": 205,
"token_acc": 0.8635224424698109,
"train_speed(iter/s)": 0.030086
},
{
"epoch": 0.0671973121075157,
"grad_norm": 0.06089496841996932,
"learning_rate": 1.99348173534855e-05,
"loss": 0.4699739456176758,
"memory(GiB)": 73.06,
"step": 210,
"token_acc": 0.8621679827709978,
"train_speed(iter/s)": 0.03019
},
{
"epoch": 0.06879724811007559,
"grad_norm": 0.05802383504947049,
"learning_rate": 1.9928764188549462e-05,
"loss": 0.46386079788208007,
"memory(GiB)": 73.06,
"step": 215,
"token_acc": 0.8754250939681403,
"train_speed(iter/s)": 0.030114
},
{
"epoch": 0.0703971841126355,
"grad_norm": 0.05515703750577632,
"learning_rate": 1.9922443302133906e-05,
"loss": 0.4679898262023926,
"memory(GiB)": 73.06,
"step": 220,
"token_acc": 0.88738807102747,
"train_speed(iter/s)": 0.030007
},
{
"epoch": 0.07199712011519539,
"grad_norm": 0.05967474480415036,
"learning_rate": 1.9915854864676665e-05,
"loss": 0.47310919761657716,
"memory(GiB)": 73.06,
"step": 225,
"token_acc": 0.8383795309168444,
"train_speed(iter/s)": 0.030071
},
{
"epoch": 0.07359705611775529,
"grad_norm": 0.06076387420670948,
"learning_rate": 1.990899905382988e-05,
"loss": 0.4678232192993164,
"memory(GiB)": 73.06,
"step": 230,
"token_acc": 0.8767123287671232,
"train_speed(iter/s)": 0.029948
},
{
"epoch": 0.07519699212031519,
"grad_norm": 0.05635803811030448,
"learning_rate": 1.9901876054455217e-05,
"loss": 0.4821170330047607,
"memory(GiB)": 73.06,
"step": 235,
"token_acc": 0.8841222879684418,
"train_speed(iter/s)": 0.029913
},
{
"epoch": 0.07679692812287509,
"grad_norm": 0.053886114557468945,
"learning_rate": 1.9894486058618863e-05,
"loss": 0.46213107109069823,
"memory(GiB)": 73.06,
"step": 240,
"token_acc": 0.8886558627264061,
"train_speed(iter/s)": 0.02992
},
{
"epoch": 0.07839686412543498,
"grad_norm": 0.06048992108753748,
"learning_rate": 1.9886829265586368e-05,
"loss": 0.4749046802520752,
"memory(GiB)": 73.06,
"step": 245,
"token_acc": 0.8758281279575999,
"train_speed(iter/s)": 0.0298
},
{
"epoch": 0.07999680012799489,
"grad_norm": 0.06662896613700448,
"learning_rate": 1.9878905881817254e-05,
"loss": 0.47487664222717285,
"memory(GiB)": 73.06,
"step": 250,
"token_acc": 0.8239743295897318,
"train_speed(iter/s)": 0.029855
},
{
"epoch": 0.07999680012799489,
"eval_loss": 0.6802101731300354,
"eval_runtime": 108.8605,
"eval_samples_per_second": 184.53,
"eval_steps_per_second": 0.928,
"eval_token_acc": 0.8656411339267154,
"step": 250
},
{
"epoch": 0.08159673613055478,
"grad_norm": 0.058967589577093804,
"learning_rate": 1.9870716120959462e-05,
"loss": 0.4691306591033936,
"memory(GiB)": 73.24,
"step": 255,
"token_acc": 0.8612697569398327,
"train_speed(iter/s)": 0.029606
},
{
"epoch": 0.08319667213311467,
"grad_norm": 0.057353651690814994,
"learning_rate": 1.986226020384359e-05,
"loss": 0.46143622398376466,
"memory(GiB)": 73.24,
"step": 260,
"token_acc": 0.8685547371094742,
"train_speed(iter/s)": 0.029681
},
{
"epoch": 0.08479660813567458,
"grad_norm": 0.05409688809510523,
"learning_rate": 1.9853538358476933e-05,
"loss": 0.4704445838928223,
"memory(GiB)": 73.24,
"step": 265,
"token_acc": 0.8804637020144431,
"train_speed(iter/s)": 0.02979
},
{
"epoch": 0.08639654413823447,
"grad_norm": 0.06968473514476099,
"learning_rate": 1.9844550820037326e-05,
"loss": 0.4717890739440918,
"memory(GiB)": 73.24,
"step": 270,
"token_acc": 0.8638003174145145,
"train_speed(iter/s)": 0.029834
},
{
"epoch": 0.08799648014079436,
"grad_norm": 0.06009720175343309,
"learning_rate": 1.9835297830866827e-05,
"loss": 0.4709662437438965,
"memory(GiB)": 73.24,
"step": 275,
"token_acc": 0.8634590377113134,
"train_speed(iter/s)": 0.029835
},
{
"epoch": 0.08959641614335427,
"grad_norm": 0.058778539356308675,
"learning_rate": 1.9825779640465157e-05,
"loss": 0.47084336280822753,
"memory(GiB)": 73.24,
"step": 280,
"token_acc": 0.9203691779351793,
"train_speed(iter/s)": 0.029942
},
{
"epoch": 0.09119635214591416,
"grad_norm": 0.054325246749067864,
"learning_rate": 1.9815996505483e-05,
"loss": 0.4666774749755859,
"memory(GiB)": 73.24,
"step": 285,
"token_acc": 0.8521723454119344,
"train_speed(iter/s)": 0.029906
},
{
"epoch": 0.09279628814847406,
"grad_norm": 0.058698263071843435,
"learning_rate": 1.9805948689715043e-05,
"loss": 0.45826416015625,
"memory(GiB)": 73.24,
"step": 290,
"token_acc": 0.8421138211382114,
"train_speed(iter/s)": 0.029895
},
{
"epoch": 0.09439622415103396,
"grad_norm": 0.05517972536097747,
"learning_rate": 1.979563646409291e-05,
"loss": 0.47627692222595214,
"memory(GiB)": 73.24,
"step": 295,
"token_acc": 0.8784122999686226,
"train_speed(iter/s)": 0.029966
},
{
"epoch": 0.09599616015359386,
"grad_norm": 0.06223926082468345,
"learning_rate": 1.9785060106677818e-05,
"loss": 0.4711057186126709,
"memory(GiB)": 73.24,
"step": 300,
"token_acc": 0.876372039283651,
"train_speed(iter/s)": 0.029911
},
{
"epoch": 0.09759609615615375,
"grad_norm": 0.060678733702642235,
"learning_rate": 1.97742199026531e-05,
"loss": 0.46833024024963377,
"memory(GiB)": 73.24,
"step": 305,
"token_acc": 0.8586995355484102,
"train_speed(iter/s)": 0.029998
},
{
"epoch": 0.09919603215871366,
"grad_norm": 0.06117494885421727,
"learning_rate": 1.9763116144316506e-05,
"loss": 0.4692807197570801,
"memory(GiB)": 73.24,
"step": 310,
"token_acc": 0.8383072793304911,
"train_speed(iter/s)": 0.029983
},
{
"epoch": 0.10079596816127355,
"grad_norm": 0.059512004342169564,
"learning_rate": 1.9751749131072335e-05,
"loss": 0.462421178817749,
"memory(GiB)": 73.24,
"step": 315,
"token_acc": 0.865073787772312,
"train_speed(iter/s)": 0.029959
},
{
"epoch": 0.10239590416383344,
"grad_norm": 0.05759903892800583,
"learning_rate": 1.9740119169423337e-05,
"loss": 0.4749638080596924,
"memory(GiB)": 73.24,
"step": 320,
"token_acc": 0.8657438292194797,
"train_speed(iter/s)": 0.030064
},
{
"epoch": 0.10399584016639335,
"grad_norm": 0.05512670495542287,
"learning_rate": 1.9728226572962474e-05,
"loss": 0.48053979873657227,
"memory(GiB)": 73.24,
"step": 325,
"token_acc": 0.9068181818181819,
"train_speed(iter/s)": 0.03004
},
{
"epoch": 0.10559577616895324,
"grad_norm": 0.05723038100011267,
"learning_rate": 1.9716071662364454e-05,
"loss": 0.47551665306091306,
"memory(GiB)": 73.24,
"step": 330,
"token_acc": 0.8362432269717038,
"train_speed(iter/s)": 0.030003
},
{
"epoch": 0.10719571217151314,
"grad_norm": 0.057638605082885846,
"learning_rate": 1.970365476537707e-05,
"loss": 0.4652701854705811,
"memory(GiB)": 73.24,
"step": 335,
"token_acc": 0.8735049401976079,
"train_speed(iter/s)": 0.030082
},
{
"epoch": 0.10879564817407304,
"grad_norm": 0.05903871731521889,
"learning_rate": 1.9690976216812397e-05,
"loss": 0.4698742389678955,
"memory(GiB)": 73.24,
"step": 340,
"token_acc": 0.8620361560418649,
"train_speed(iter/s)": 0.030027
},
{
"epoch": 0.11039558417663294,
"grad_norm": 0.053856521964694516,
"learning_rate": 1.9678036358537726e-05,
"loss": 0.4701416015625,
"memory(GiB)": 73.24,
"step": 345,
"token_acc": 0.8708435421771089,
"train_speed(iter/s)": 0.03002
},
{
"epoch": 0.11199552017919283,
"grad_norm": 0.05586893539038131,
"learning_rate": 1.966483553946637e-05,
"loss": 0.47447028160095217,
"memory(GiB)": 73.24,
"step": 350,
"token_acc": 0.8617533718689788,
"train_speed(iter/s)": 0.030041
},
{
"epoch": 0.11359545618175274,
"grad_norm": 0.052599438001953325,
"learning_rate": 1.9651374115548255e-05,
"loss": 0.4637298583984375,
"memory(GiB)": 73.24,
"step": 355,
"token_acc": 0.8874341610233258,
"train_speed(iter/s)": 0.029967
},
{
"epoch": 0.11519539218431263,
"grad_norm": 0.05804143123407663,
"learning_rate": 1.9637652449760297e-05,
"loss": 0.4660144329071045,
"memory(GiB)": 73.24,
"step": 360,
"token_acc": 0.8349885408708938,
"train_speed(iter/s)": 0.030034
},
{
"epoch": 0.11679532818687252,
"grad_norm": 0.06055547849970778,
"learning_rate": 1.9623670912096656e-05,
"loss": 0.4716383934020996,
"memory(GiB)": 73.24,
"step": 365,
"token_acc": 0.8751012473675684,
"train_speed(iter/s)": 0.02998
},
{
"epoch": 0.11839526418943243,
"grad_norm": 0.058520598293842735,
"learning_rate": 1.9609429879558726e-05,
"loss": 0.46298699378967284,
"memory(GiB)": 73.24,
"step": 370,
"token_acc": 0.8553921568627451,
"train_speed(iter/s)": 0.029931
},
{
"epoch": 0.11999520019199232,
"grad_norm": 0.058584318589478955,
"learning_rate": 1.9594929736144978e-05,
"loss": 0.4756875514984131,
"memory(GiB)": 73.24,
"step": 375,
"token_acc": 0.8618346545866364,
"train_speed(iter/s)": 0.030006
},
{
"epoch": 0.12159513619455221,
"grad_norm": 0.05966533070217228,
"learning_rate": 1.958017087284061e-05,
"loss": 0.4596414089202881,
"memory(GiB)": 73.24,
"step": 380,
"token_acc": 0.8836156297165856,
"train_speed(iter/s)": 0.029968
},
{
"epoch": 0.12319507219711212,
"grad_norm": 0.06510894340277039,
"learning_rate": 1.9565153687607006e-05,
"loss": 0.4687026023864746,
"memory(GiB)": 73.24,
"step": 385,
"token_acc": 0.8694005270092227,
"train_speed(iter/s)": 0.029956
},
{
"epoch": 0.12479500819967201,
"grad_norm": 0.05180588304383506,
"learning_rate": 1.9549878585371006e-05,
"loss": 0.4649878978729248,
"memory(GiB)": 73.24,
"step": 390,
"token_acc": 0.8677233429394813,
"train_speed(iter/s)": 0.029999
},
{
"epoch": 0.1263949442022319,
"grad_norm": 0.060875424512666344,
"learning_rate": 1.9534345978013972e-05,
"loss": 0.47073874473571775,
"memory(GiB)": 73.24,
"step": 395,
"token_acc": 0.8484663512894858,
"train_speed(iter/s)": 0.029947
},
{
"epoch": 0.1279948802047918,
"grad_norm": 0.0571374353277554,
"learning_rate": 1.9518556284360696e-05,
"loss": 0.4666412353515625,
"memory(GiB)": 73.24,
"step": 400,
"token_acc": 0.8975701436434421,
"train_speed(iter/s)": 0.029998
},
{
"epoch": 0.1295948162073517,
"grad_norm": 0.05215050598306155,
"learning_rate": 1.9502509930168113e-05,
"loss": 0.4628121376037598,
"memory(GiB)": 73.24,
"step": 405,
"token_acc": 0.8816677696889477,
"train_speed(iter/s)": 0.029966
},
{
"epoch": 0.1311947522099116,
"grad_norm": 0.07947570193916972,
"learning_rate": 1.9486207348113803e-05,
"loss": 0.4593012809753418,
"memory(GiB)": 73.24,
"step": 410,
"token_acc": 0.8692473832862602,
"train_speed(iter/s)": 0.029911
},
{
"epoch": 0.1327946882124715,
"grad_norm": 0.07262611466641217,
"learning_rate": 1.946964897778433e-05,
"loss": 0.47004990577697753,
"memory(GiB)": 73.24,
"step": 415,
"token_acc": 0.8736337958983176,
"train_speed(iter/s)": 0.029969
},
{
"epoch": 0.1343946242150314,
"grad_norm": 0.053754461298334506,
"learning_rate": 1.9452835265663404e-05,
"loss": 0.4695271015167236,
"memory(GiB)": 73.24,
"step": 420,
"token_acc": 0.8747993579454254,
"train_speed(iter/s)": 0.029901
},
{
"epoch": 0.1359945602175913,
"grad_norm": 0.0742051800083311,
"learning_rate": 1.9435766665119823e-05,
"loss": 0.47011446952819824,
"memory(GiB)": 73.24,
"step": 425,
"token_acc": 0.8356736242884251,
"train_speed(iter/s)": 0.029856
},
{
"epoch": 0.13759449622015119,
"grad_norm": 0.06429200177825628,
"learning_rate": 1.941844363639525e-05,
"loss": 0.476796817779541,
"memory(GiB)": 73.24,
"step": 430,
"token_acc": 0.869019972131909,
"train_speed(iter/s)": 0.0299
},
{
"epoch": 0.13919443222271108,
"grad_norm": 0.06544854557851852,
"learning_rate": 1.9400866646591816e-05,
"loss": 0.4666853904724121,
"memory(GiB)": 73.24,
"step": 435,
"token_acc": 0.8204667863554758,
"train_speed(iter/s)": 0.029847
},
{
"epoch": 0.140794368225271,
"grad_norm": 0.0546565929911768,
"learning_rate": 1.9383036169659513e-05,
"loss": 0.4738778591156006,
"memory(GiB)": 73.24,
"step": 440,
"token_acc": 0.8605809128630706,
"train_speed(iter/s)": 0.029855
},
{
"epoch": 0.1423943042278309,
"grad_norm": 0.06789336848906298,
"learning_rate": 1.936495268638342e-05,
"loss": 0.47726120948791506,
"memory(GiB)": 73.24,
"step": 445,
"token_acc": 0.8404369243949454,
"train_speed(iter/s)": 0.029875
},
{
"epoch": 0.14399424023039079,
"grad_norm": 0.049909982274150465,
"learning_rate": 1.934661668437073e-05,
"loss": 0.47165632247924805,
"memory(GiB)": 73.24,
"step": 450,
"token_acc": 0.848471615720524,
"train_speed(iter/s)": 0.029826
},
{
"epoch": 0.14559417623295068,
"grad_norm": 0.057441474731933166,
"learning_rate": 1.932802865803763e-05,
"loss": 0.4703391075134277,
"memory(GiB)": 73.24,
"step": 455,
"token_acc": 0.8466442358774571,
"train_speed(iter/s)": 0.029877
},
{
"epoch": 0.14719411223551057,
"grad_norm": 0.07263904251491092,
"learning_rate": 1.930918910859592e-05,
"loss": 0.467697811126709,
"memory(GiB)": 73.24,
"step": 460,
"token_acc": 0.8491142333536957,
"train_speed(iter/s)": 0.02984
},
{
"epoch": 0.14879404823807046,
"grad_norm": 0.06769237623086669,
"learning_rate": 1.9290098544039546e-05,
"loss": 0.46541628837585447,
"memory(GiB)": 73.24,
"step": 465,
"token_acc": 0.8555353301340394,
"train_speed(iter/s)": 0.02978
},
{
"epoch": 0.15039398424063039,
"grad_norm": 0.06751583633556477,
"learning_rate": 1.927075747913088e-05,
"loss": 0.47134056091308596,
"memory(GiB)": 73.24,
"step": 470,
"token_acc": 0.8708000507163687,
"train_speed(iter/s)": 0.029832
},
{
"epoch": 0.15199392024319028,
"grad_norm": 0.0539492567012165,
"learning_rate": 1.9251166435386837e-05,
"loss": 0.4645866394042969,
"memory(GiB)": 73.24,
"step": 475,
"token_acc": 0.849832526981764,
"train_speed(iter/s)": 0.029779
},
{
"epoch": 0.15359385624575017,
"grad_norm": 0.06038706866556876,
"learning_rate": 1.923132594106483e-05,
"loss": 0.46890692710876464,
"memory(GiB)": 73.24,
"step": 480,
"token_acc": 0.8665925514174542,
"train_speed(iter/s)": 0.02976
},
{
"epoch": 0.15519379224831006,
"grad_norm": 0.05215840717634863,
"learning_rate": 1.92112365311485e-05,
"loss": 0.46829919815063475,
"memory(GiB)": 73.24,
"step": 485,
"token_acc": 0.861963565228023,
"train_speed(iter/s)": 0.029794
},
{
"epoch": 0.15679372825086996,
"grad_norm": 0.06554142579397569,
"learning_rate": 1.919089874733332e-05,
"loss": 0.4702622413635254,
"memory(GiB)": 73.24,
"step": 490,
"token_acc": 0.8809186723297153,
"train_speed(iter/s)": 0.029747
},
{
"epoch": 0.15839366425342985,
"grad_norm": 0.0601172563145885,
"learning_rate": 1.9170313138011964e-05,
"loss": 0.46490135192871096,
"memory(GiB)": 73.24,
"step": 495,
"token_acc": 0.8890911637025627,
"train_speed(iter/s)": 0.02977
},
{
"epoch": 0.15999360025598977,
"grad_norm": 0.05924399402367875,
"learning_rate": 1.9149480258259535e-05,
"loss": 0.46698894500732424,
"memory(GiB)": 73.24,
"step": 500,
"token_acc": 0.8781434114096853,
"train_speed(iter/s)": 0.029766
},
{
"epoch": 0.15999360025598977,
"eval_loss": 0.677643895149231,
"eval_runtime": 109.3458,
"eval_samples_per_second": 183.711,
"eval_steps_per_second": 0.924,
"eval_token_acc": 0.8661408286670019,
"step": 500
},
{
"epoch": 0.16159353625854966,
"grad_norm": 0.05118070522939682,
"learning_rate": 1.9128400669818586e-05,
"loss": 0.4606743812561035,
"memory(GiB)": 73.24,
"step": 505,
"token_acc": 0.8727327237295758,
"train_speed(iter/s)": 0.029628
},
{
"epoch": 0.16319347226110956,
"grad_norm": 0.05904937387674259,
"learning_rate": 1.9107074941083987e-05,
"loss": 0.47115492820739746,
"memory(GiB)": 73.24,
"step": 510,
"token_acc": 0.8801781737193763,
"train_speed(iter/s)": 0.029663
},
{
"epoch": 0.16479340826366945,
"grad_norm": 0.061211680590962145,
"learning_rate": 1.9085503647087588e-05,
"loss": 0.46154184341430665,
"memory(GiB)": 73.24,
"step": 515,
"token_acc": 0.8573438874230431,
"train_speed(iter/s)": 0.029714
},
{
"epoch": 0.16639334426622934,
"grad_norm": 0.05461804298242196,
"learning_rate": 1.906368736948272e-05,
"loss": 0.46891465187072756,
"memory(GiB)": 73.24,
"step": 520,
"token_acc": 0.8665508756694167,
"train_speed(iter/s)": 0.029721
},
{
"epoch": 0.16799328026878924,
"grad_norm": 0.059072521440841075,
"learning_rate": 1.9041626696528503e-05,
"loss": 0.4666083812713623,
"memory(GiB)": 73.24,
"step": 525,
"token_acc": 0.8742783835792175,
"train_speed(iter/s)": 0.029735
},
{
"epoch": 0.16959321627134916,
"grad_norm": 0.06762878495647719,
"learning_rate": 1.9019322223073997e-05,
"loss": 0.4684437274932861,
"memory(GiB)": 73.24,
"step": 530,
"token_acc": 0.8906074591493077,
"train_speed(iter/s)": 0.029782
},
{
"epoch": 0.17119315227390905,
"grad_norm": 0.05741557316745661,
"learning_rate": 1.899677455054215e-05,
"loss": 0.4690097332000732,
"memory(GiB)": 73.24,
"step": 535,
"token_acc": 0.8231878958479943,
"train_speed(iter/s)": 0.029785
},
{
"epoch": 0.17279308827646894,
"grad_norm": 0.049026865135578496,
"learning_rate": 1.8973984286913584e-05,
"loss": 0.469140625,
"memory(GiB)": 73.24,
"step": 540,
"token_acc": 0.8849415539766216,
"train_speed(iter/s)": 0.029789
},
{
"epoch": 0.17439302427902884,
"grad_norm": 0.059746465018255104,
"learning_rate": 1.895095204671021e-05,
"loss": 0.4646149158477783,
"memory(GiB)": 73.24,
"step": 545,
"token_acc": 0.8944385405596883,
"train_speed(iter/s)": 0.029813
},
{
"epoch": 0.17599296028158873,
"grad_norm": 0.049833714934798115,
"learning_rate": 1.892767845097864e-05,
"loss": 0.47077240943908694,
"memory(GiB)": 73.24,
"step": 550,
"token_acc": 0.8640860961638605,
"train_speed(iter/s)": 0.029794
},
{
"epoch": 0.17759289628414862,
"grad_norm": 0.06593845007149325,
"learning_rate": 1.890416412727346e-05,
"loss": 0.46265759468078616,
"memory(GiB)": 73.24,
"step": 555,
"token_acc": 0.8249895412076419,
"train_speed(iter/s)": 0.02984
},
{
"epoch": 0.17919283228670854,
"grad_norm": 0.058254003445636866,
"learning_rate": 1.88804097096403e-05,
"loss": 0.459829044342041,
"memory(GiB)": 73.24,
"step": 560,
"token_acc": 0.8835873095178616,
"train_speed(iter/s)": 0.029842
},
{
"epoch": 0.18079276828926844,
"grad_norm": 0.07335953644753283,
"learning_rate": 1.8856415838598738e-05,
"loss": 0.45765042304992676,
"memory(GiB)": 73.24,
"step": 565,
"token_acc": 0.8755007210382951,
"train_speed(iter/s)": 0.029818
},
{
"epoch": 0.18239270429182833,
"grad_norm": 0.06659181547700674,
"learning_rate": 1.8832183161125026e-05,
"loss": 0.4609128475189209,
"memory(GiB)": 73.24,
"step": 570,
"token_acc": 0.8344311377245509,
"train_speed(iter/s)": 0.029871
},
{
"epoch": 0.18399264029438822,
"grad_norm": 0.05836437871791382,
"learning_rate": 1.8807712330634645e-05,
"loss": 0.4691438674926758,
"memory(GiB)": 73.24,
"step": 575,
"token_acc": 0.8848027659908848,
"train_speed(iter/s)": 0.029828
},
{
"epoch": 0.18559257629694811,
"grad_norm": 0.05735059462858394,
"learning_rate": 1.87830040069647e-05,
"loss": 0.4602513790130615,
"memory(GiB)": 73.24,
"step": 580,
"token_acc": 0.8959147903465012,
"train_speed(iter/s)": 0.029816
},
{
"epoch": 0.187192512299508,
"grad_norm": 0.05337219773586585,
"learning_rate": 1.87580588563561e-05,
"loss": 0.46318631172180175,
"memory(GiB)": 73.24,
"step": 585,
"token_acc": 0.8725881039706586,
"train_speed(iter/s)": 0.029851
},
{
"epoch": 0.18879244830206793,
"grad_norm": 0.05886716832883729,
"learning_rate": 1.873287755143563e-05,
"loss": 0.4604507923126221,
"memory(GiB)": 73.24,
"step": 590,
"token_acc": 0.9041755130927105,
"train_speed(iter/s)": 0.029822
},
{
"epoch": 0.19039238430462782,
"grad_norm": 0.053483810048332456,
"learning_rate": 1.8707460771197773e-05,
"loss": 0.46618080139160156,
"memory(GiB)": 73.24,
"step": 595,
"token_acc": 0.8785046728971962,
"train_speed(iter/s)": 0.029819
},
{
"epoch": 0.1919923203071877,
"grad_norm": 0.0518592001281956,
"learning_rate": 1.868180920098644e-05,
"loss": 0.4680916786193848,
"memory(GiB)": 73.24,
"step": 600,
"token_acc": 0.8467063770147162,
"train_speed(iter/s)": 0.029843
},
{
"epoch": 0.1935922563097476,
"grad_norm": 0.07018232236413237,
"learning_rate": 1.8655923532476463e-05,
"loss": 0.46170759201049805,
"memory(GiB)": 73.24,
"step": 605,
"token_acc": 0.889030612244898,
"train_speed(iter/s)": 0.02981
},
{
"epoch": 0.1951921923123075,
"grad_norm": 0.06030421269833889,
"learning_rate": 1.8629804463654956e-05,
"loss": 0.46511187553405764,
"memory(GiB)": 73.24,
"step": 610,
"token_acc": 0.8554680664916885,
"train_speed(iter/s)": 0.029852
},
{
"epoch": 0.1967921283148674,
"grad_norm": 0.056137765321266526,
"learning_rate": 1.8603452698802498e-05,
"loss": 0.47327299118041993,
"memory(GiB)": 76.61,
"step": 615,
"token_acc": 0.8645191852202747,
"train_speed(iter/s)": 0.029831
},
{
"epoch": 0.1983920643174273,
"grad_norm": 0.05458475201274465,
"learning_rate": 1.857686894847413e-05,
"loss": 0.45963249206542967,
"memory(GiB)": 76.61,
"step": 620,
"token_acc": 0.8517509197438343,
"train_speed(iter/s)": 0.029791
},
{
"epoch": 0.1999920003199872,
"grad_norm": 0.059902578480064236,
"learning_rate": 1.8550053929480202e-05,
"loss": 0.4687147617340088,
"memory(GiB)": 76.61,
"step": 625,
"token_acc": 0.8958185683912119,
"train_speed(iter/s)": 0.029833
},
{
"epoch": 0.2015919363225471,
"grad_norm": 0.0539478773118384,
"learning_rate": 1.8523008364867056e-05,
"loss": 0.4696544647216797,
"memory(GiB)": 76.61,
"step": 630,
"token_acc": 0.8439355385920272,
"train_speed(iter/s)": 0.029796
},
{
"epoch": 0.203191872325107,
"grad_norm": 0.05688926646164217,
"learning_rate": 1.8495732983897504e-05,
"loss": 0.4628334045410156,
"memory(GiB)": 76.61,
"step": 635,
"token_acc": 0.8406656465187452,
"train_speed(iter/s)": 0.029775
},
{
"epoch": 0.20479180832766689,
"grad_norm": 0.055104479428209605,
"learning_rate": 1.8468228522031197e-05,
"loss": 0.4559271812438965,
"memory(GiB)": 76.61,
"step": 640,
"token_acc": 0.8823529411764706,
"train_speed(iter/s)": 0.029794
},
{
"epoch": 0.20639174433022678,
"grad_norm": 0.058080447436547736,
"learning_rate": 1.8440495720904758e-05,
"loss": 0.4649765968322754,
"memory(GiB)": 76.61,
"step": 645,
"token_acc": 0.8708735027753433,
"train_speed(iter/s)": 0.029752
},
{
"epoch": 0.2079916803327867,
"grad_norm": 0.06300003986546152,
"learning_rate": 1.8412535328311813e-05,
"loss": 0.47095327377319335,
"memory(GiB)": 76.61,
"step": 650,
"token_acc": 0.8504976200778883,
"train_speed(iter/s)": 0.029755
},
{
"epoch": 0.2095916163353466,
"grad_norm": 0.06584526718748161,
"learning_rate": 1.8384348098182815e-05,
"loss": 0.46697392463684084,
"memory(GiB)": 76.61,
"step": 655,
"token_acc": 0.8224407171775593,
"train_speed(iter/s)": 0.029762
},
{
"epoch": 0.21119155233790649,
"grad_norm": 0.07147957728971413,
"learning_rate": 1.8355934790564718e-05,
"loss": 0.4684570789337158,
"memory(GiB)": 76.61,
"step": 660,
"token_acc": 0.8842165898617511,
"train_speed(iter/s)": 0.029723
},
{
"epoch": 0.21279148834046638,
"grad_norm": 0.06592046292925295,
"learning_rate": 1.832729617160047e-05,
"loss": 0.461454439163208,
"memory(GiB)": 76.61,
"step": 665,
"token_acc": 0.9114801444043321,
"train_speed(iter/s)": 0.02976
},
{
"epoch": 0.21439142434302627,
"grad_norm": 0.0656829490109071,
"learning_rate": 1.8298433013508384e-05,
"loss": 0.46404447555541994,
"memory(GiB)": 76.61,
"step": 670,
"token_acc": 0.8516549891278087,
"train_speed(iter/s)": 0.029736
},
{
"epoch": 0.21599136034558616,
"grad_norm": 0.05417998837874903,
"learning_rate": 1.826934609456129e-05,
"loss": 0.47208566665649415,
"memory(GiB)": 76.61,
"step": 675,
"token_acc": 0.8798815733822078,
"train_speed(iter/s)": 0.029718
},
{
"epoch": 0.21759129634814608,
"grad_norm": 0.06917195844649823,
"learning_rate": 1.8240036199065546e-05,
"loss": 0.4724391460418701,
"memory(GiB)": 76.61,
"step": 680,
"token_acc": 0.875845675626257,
"train_speed(iter/s)": 0.029745
},
{
"epoch": 0.21919123235070598,
"grad_norm": 0.055849189404917746,
"learning_rate": 1.8210504117339917e-05,
"loss": 0.463816499710083,
"memory(GiB)": 76.61,
"step": 685,
"token_acc": 0.8841904379268782,
"train_speed(iter/s)": 0.029711
},
{
"epoch": 0.22079116835326587,
"grad_norm": 0.059563786969142496,
"learning_rate": 1.8180750645694236e-05,
"loss": 0.4678086757659912,
"memory(GiB)": 76.61,
"step": 690,
"token_acc": 0.8675231977159172,
"train_speed(iter/s)": 0.029714
},
{
"epoch": 0.22239110435582576,
"grad_norm": 0.05908606421708839,
"learning_rate": 1.8150776586407957e-05,
"loss": 0.46315860748291016,
"memory(GiB)": 76.61,
"step": 695,
"token_acc": 0.8914956011730205,
"train_speed(iter/s)": 0.029731
},
{
"epoch": 0.22399104035838566,
"grad_norm": 0.05617530731492468,
"learning_rate": 1.8120582747708503e-05,
"loss": 0.46682062149047854,
"memory(GiB)": 76.61,
"step": 700,
"token_acc": 0.8805088596092685,
"train_speed(iter/s)": 0.029689
},
{
"epoch": 0.22559097636094555,
"grad_norm": 0.06138477303861948,
"learning_rate": 1.8090169943749477e-05,
"loss": 0.47155141830444336,
"memory(GiB)": 76.61,
"step": 705,
"token_acc": 0.8753766681015928,
"train_speed(iter/s)": 0.029703
},
{
"epoch": 0.22719091236350547,
"grad_norm": 0.07073141016351848,
"learning_rate": 1.8059538994588715e-05,
"loss": 0.45953845977783203,
"memory(GiB)": 76.61,
"step": 710,
"token_acc": 0.8449233877757198,
"train_speed(iter/s)": 0.0297
},
{
"epoch": 0.22879084836606536,
"grad_norm": 0.06266619359839708,
"learning_rate": 1.8028690726166172e-05,
"loss": 0.4604049205780029,
"memory(GiB)": 76.61,
"step": 715,
"token_acc": 0.8688032048072108,
"train_speed(iter/s)": 0.02966
},
{
"epoch": 0.23039078436862526,
"grad_norm": 0.0563660774004587,
"learning_rate": 1.7997625970281652e-05,
"loss": 0.4622708797454834,
"memory(GiB)": 76.61,
"step": 720,
"token_acc": 0.8698216735253772,
"train_speed(iter/s)": 0.029685
},
{
"epoch": 0.23199072037118515,
"grad_norm": 0.06596213612143108,
"learning_rate": 1.796634556457236e-05,
"loss": 0.4681892395019531,
"memory(GiB)": 76.61,
"step": 725,
"token_acc": 0.8842619184376795,
"train_speed(iter/s)": 0.029661
},
{
"epoch": 0.23359065637374504,
"grad_norm": 0.05364579438678848,
"learning_rate": 1.793485035249036e-05,
"loss": 0.46258745193481443,
"memory(GiB)": 76.61,
"step": 730,
"token_acc": 0.8599531615925059,
"train_speed(iter/s)": 0.02965
},
{
"epoch": 0.23519059237630494,
"grad_norm": 0.07509450433159735,
"learning_rate": 1.7903141183279776e-05,
"loss": 0.47242441177368166,
"memory(GiB)": 76.61,
"step": 735,
"token_acc": 0.8404958677685951,
"train_speed(iter/s)": 0.029665
},
{
"epoch": 0.23679052837886486,
"grad_norm": 0.06478313540282635,
"learning_rate": 1.7871218911953942e-05,
"loss": 0.4565444469451904,
"memory(GiB)": 76.61,
"step": 740,
"token_acc": 0.8338650865998177,
"train_speed(iter/s)": 0.029634
},
{
"epoch": 0.23839046438142475,
"grad_norm": 0.06348939893307848,
"learning_rate": 1.7839084399272317e-05,
"loss": 0.4670473575592041,
"memory(GiB)": 76.61,
"step": 745,
"token_acc": 0.8652410477034038,
"train_speed(iter/s)": 0.029638
},
{
"epoch": 0.23999040038398464,
"grad_norm": 0.07434587030241245,
"learning_rate": 1.780673851171728e-05,
"loss": 0.47047910690307615,
"memory(GiB)": 76.61,
"step": 750,
"token_acc": 0.88801504530689,
"train_speed(iter/s)": 0.029638
},
{
"epoch": 0.23999040038398464,
"eval_loss": 0.6746003031730652,
"eval_runtime": 113.2223,
"eval_samples_per_second": 177.421,
"eval_steps_per_second": 0.892,
"eval_token_acc": 0.8668385651547512,
"step": 750
},
{
"epoch": 0.24159033638654454,
"grad_norm": 0.06732795706859432,
"learning_rate": 1.777418212147079e-05,
"loss": 0.46190509796142576,
"memory(GiB)": 76.61,
"step": 755,
"token_acc": 0.8881346728210697,
"train_speed(iter/s)": 0.029543
},
{
"epoch": 0.24319027238910443,
"grad_norm": 0.06836940989947664,
"learning_rate": 1.7741416106390828e-05,
"loss": 0.46283302307128904,
"memory(GiB)": 76.61,
"step": 760,
"token_acc": 0.8831443688586545,
"train_speed(iter/s)": 0.029566
},
{
"epoch": 0.24479020839166432,
"grad_norm": 0.07072489516219096,
"learning_rate": 1.7708441349987753e-05,
"loss": 0.4619740962982178,
"memory(GiB)": 76.61,
"step": 765,
"token_acc": 0.8610668789808917,
"train_speed(iter/s)": 0.0296
},
{
"epoch": 0.24639014439422424,
"grad_norm": 0.07152232857362027,
"learning_rate": 1.767525874140048e-05,
"loss": 0.46694121360778806,
"memory(GiB)": 76.61,
"step": 770,
"token_acc": 0.8397869022869023,
"train_speed(iter/s)": 0.029606
},
{
"epoch": 0.24799008039678413,
"grad_norm": 0.059354056163304685,
"learning_rate": 1.7641869175372493e-05,
"loss": 0.4596868991851807,
"memory(GiB)": 76.61,
"step": 775,
"token_acc": 0.8582827406764961,
"train_speed(iter/s)": 0.029599
},
{
"epoch": 0.24959001639934403,
"grad_norm": 0.0629690289705531,
"learning_rate": 1.7608273552227723e-05,
"loss": 0.4583168029785156,
"memory(GiB)": 76.61,
"step": 780,
"token_acc": 0.8841532106646639,
"train_speed(iter/s)": 0.029639
},
{
"epoch": 0.25118995240190395,
"grad_norm": 0.05810355160479093,
"learning_rate": 1.7574472777846276e-05,
"loss": 0.47337069511413576,
"memory(GiB)": 76.61,
"step": 785,
"token_acc": 0.8676557863501484,
"train_speed(iter/s)": 0.029632
},
{
"epoch": 0.2527898884044638,
"grad_norm": 0.05365185572887828,
"learning_rate": 1.7540467763639994e-05,
"loss": 0.46567063331604003,
"memory(GiB)": 76.61,
"step": 790,
"token_acc": 0.8745288099084545,
"train_speed(iter/s)": 0.029629
},
{
"epoch": 0.25438982440702373,
"grad_norm": 0.054672322658953366,
"learning_rate": 1.7506259426527903e-05,
"loss": 0.47023472785949705,
"memory(GiB)": 76.61,
"step": 795,
"token_acc": 0.874407844001322,
"train_speed(iter/s)": 0.02965
},
{
"epoch": 0.2559897604095836,
"grad_norm": 0.057060955079149434,
"learning_rate": 1.7471848688911465e-05,
"loss": 0.4684537410736084,
"memory(GiB)": 76.61,
"step": 800,
"token_acc": 0.8839382448537378,
"train_speed(iter/s)": 0.029634
},
{
"epoch": 0.2575896964121435,
"grad_norm": 0.06051290772323595,
"learning_rate": 1.7437236478649718e-05,
"loss": 0.46199979782104494,
"memory(GiB)": 76.61,
"step": 805,
"token_acc": 0.8673650919153983,
"train_speed(iter/s)": 0.02966
},
{
"epoch": 0.2591896324147034,
"grad_norm": 0.0643397562387603,
"learning_rate": 1.7402423729034252e-05,
"loss": 0.4548381805419922,
"memory(GiB)": 76.61,
"step": 810,
"token_acc": 0.83125,
"train_speed(iter/s)": 0.029652
},
{
"epoch": 0.2607895684172633,
"grad_norm": 0.065624934571794,
"learning_rate": 1.736741137876405e-05,
"loss": 0.46353764533996583,
"memory(GiB)": 76.61,
"step": 815,
"token_acc": 0.8907202528787537,
"train_speed(iter/s)": 0.029628
},
{
"epoch": 0.2623895044198232,
"grad_norm": 0.053961693017135055,
"learning_rate": 1.7332200371920173e-05,
"loss": 0.46685361862182617,
"memory(GiB)": 76.61,
"step": 820,
"token_acc": 0.8522188711762172,
"train_speed(iter/s)": 0.029672
},
{
"epoch": 0.2639894404223831,
"grad_norm": 0.054388550053431586,
"learning_rate": 1.72967916579403e-05,
"loss": 0.46024084091186523,
"memory(GiB)": 76.61,
"step": 825,
"token_acc": 0.8684630384683567,
"train_speed(iter/s)": 0.02966
},
{
"epoch": 0.265589376424943,
"grad_norm": 0.0583019332597641,
"learning_rate": 1.7261186191593135e-05,
"loss": 0.47214059829711913,
"memory(GiB)": 76.61,
"step": 830,
"token_acc": 0.8717123935666982,
"train_speed(iter/s)": 0.029645
},
{
"epoch": 0.2671893124275029,
"grad_norm": 0.06004272220759217,
"learning_rate": 1.7225384932952655e-05,
"loss": 0.4626835823059082,
"memory(GiB)": 76.61,
"step": 835,
"token_acc": 0.8737211788059246,
"train_speed(iter/s)": 0.02967
},
{
"epoch": 0.2687892484300628,
"grad_norm": 0.05611993161069816,
"learning_rate": 1.7189388847372227e-05,
"loss": 0.46799750328063966,
"memory(GiB)": 76.61,
"step": 840,
"token_acc": 0.8781684382665577,
"train_speed(iter/s)": 0.029642
},
{
"epoch": 0.2703891844326227,
"grad_norm": 0.06345947319153013,
"learning_rate": 1.715319890545857e-05,
"loss": 0.4568619728088379,
"memory(GiB)": 76.61,
"step": 845,
"token_acc": 0.860916969527537,
"train_speed(iter/s)": 0.029655
},
{
"epoch": 0.2719891204351826,
"grad_norm": 0.0592531603954309,
"learning_rate": 1.7116816083045603e-05,
"loss": 0.46942729949951173,
"memory(GiB)": 76.61,
"step": 850,
"token_acc": 0.8726317245194303,
"train_speed(iter/s)": 0.029655
},
{
"epoch": 0.2735890564377425,
"grad_norm": 0.05711267065318382,
"learning_rate": 1.7080241361168108e-05,
"loss": 0.45801239013671874,
"memory(GiB)": 76.61,
"step": 855,
"token_acc": 0.8834167608590344,
"train_speed(iter/s)": 0.02963
},
{
"epoch": 0.27518899244030237,
"grad_norm": 0.05715792257951623,
"learning_rate": 1.704347572603529e-05,
"loss": 0.4675910472869873,
"memory(GiB)": 76.61,
"step": 860,
"token_acc": 0.8361073624231519,
"train_speed(iter/s)": 0.029659
},
{
"epoch": 0.2767889284428623,
"grad_norm": 0.056617536923221766,
"learning_rate": 1.700652016900419e-05,
"loss": 0.467483377456665,
"memory(GiB)": 76.61,
"step": 865,
"token_acc": 0.8753590807532716,
"train_speed(iter/s)": 0.029639
},
{
"epoch": 0.27838886444542216,
"grad_norm": 0.060433939578350394,
"learning_rate": 1.696937568655294e-05,
"loss": 0.46129570007324217,
"memory(GiB)": 76.61,
"step": 870,
"token_acc": 0.8700755748512623,
"train_speed(iter/s)": 0.029622
},
{
"epoch": 0.2799888004479821,
"grad_norm": 0.06826391103956585,
"learning_rate": 1.6932043280253892e-05,
"loss": 0.47449960708618166,
"memory(GiB)": 76.61,
"step": 875,
"token_acc": 0.8767408356010885,
"train_speed(iter/s)": 0.02965
},
{
"epoch": 0.281588736450542,
"grad_norm": 0.060978189753072065,
"learning_rate": 1.689452395674664e-05,
"loss": 0.464243745803833,
"memory(GiB)": 76.61,
"step": 880,
"token_acc": 0.8622170179547228,
"train_speed(iter/s)": 0.029624
},
{
"epoch": 0.28318867245310186,
"grad_norm": 0.0760276206328267,
"learning_rate": 1.6856818727710847e-05,
"loss": 0.4566212177276611,
"memory(GiB)": 76.61,
"step": 885,
"token_acc": 0.8465499485066942,
"train_speed(iter/s)": 0.029618
},
{
"epoch": 0.2847886084556618,
"grad_norm": 0.05693121191664627,
"learning_rate": 1.6818928609838967e-05,
"loss": 0.46042599678039553,
"memory(GiB)": 76.61,
"step": 890,
"token_acc": 0.8798391728891441,
"train_speed(iter/s)": 0.029627
},
{
"epoch": 0.28638854445822165,
"grad_norm": 0.05744826995499506,
"learning_rate": 1.678085462480885e-05,
"loss": 0.4604465961456299,
"memory(GiB)": 76.61,
"step": 895,
"token_acc": 0.8780676542118063,
"train_speed(iter/s)": 0.029599
},
{
"epoch": 0.28798848046078157,
"grad_norm": 0.06271464886952488,
"learning_rate": 1.6742597799256182e-05,
"loss": 0.46231966018676757,
"memory(GiB)": 76.61,
"step": 900,
"token_acc": 0.8866765515780555,
"train_speed(iter/s)": 0.029611
},
{
"epoch": 0.2895884164633415,
"grad_norm": 0.06044356676681803,
"learning_rate": 1.6704159164746797e-05,
"loss": 0.47655544281005857,
"memory(GiB)": 76.61,
"step": 905,
"token_acc": 0.8872944211544663,
"train_speed(iter/s)": 0.029601
},
{
"epoch": 0.29118835246590136,
"grad_norm": 0.05103569816400521,
"learning_rate": 1.6665539757748866e-05,
"loss": 0.4603917121887207,
"memory(GiB)": 76.61,
"step": 910,
"token_acc": 0.8611705475141599,
"train_speed(iter/s)": 0.029574
},
{
"epoch": 0.2927882884684613,
"grad_norm": 0.055811472748585486,
"learning_rate": 1.6626740619604967e-05,
"loss": 0.46213679313659667,
"memory(GiB)": 76.61,
"step": 915,
"token_acc": 0.8148507643775783,
"train_speed(iter/s)": 0.029594
},
{
"epoch": 0.29438822447102114,
"grad_norm": 0.05463929857953068,
"learning_rate": 1.658776279650397e-05,
"loss": 0.4658839702606201,
"memory(GiB)": 76.61,
"step": 920,
"token_acc": 0.8766637089618456,
"train_speed(iter/s)": 0.029577
},
{
"epoch": 0.29598816047358106,
"grad_norm": 0.06343067949686905,
"learning_rate": 1.6548607339452853e-05,
"loss": 0.46423888206481934,
"memory(GiB)": 76.61,
"step": 925,
"token_acc": 0.8785782119115453,
"train_speed(iter/s)": 0.029564
},
{
"epoch": 0.29758809647614093,
"grad_norm": 0.052431934937864355,
"learning_rate": 1.6509275304248366e-05,
"loss": 0.46324734687805175,
"memory(GiB)": 76.61,
"step": 930,
"token_acc": 0.8571011956838729,
"train_speed(iter/s)": 0.02958
},
{
"epoch": 0.29918803247870085,
"grad_norm": 0.059009943510604755,
"learning_rate": 1.6469767751448538e-05,
"loss": 0.46290836334228513,
"memory(GiB)": 76.61,
"step": 935,
"token_acc": 0.8388616290480864,
"train_speed(iter/s)": 0.029556
},
{
"epoch": 0.30078796848126077,
"grad_norm": 0.05160057372757322,
"learning_rate": 1.6430085746344107e-05,
"loss": 0.45898871421813964,
"memory(GiB)": 76.61,
"step": 940,
"token_acc": 0.8690580344123651,
"train_speed(iter/s)": 0.029556
},
{
"epoch": 0.30238790448382064,
"grad_norm": 0.05612231994140208,
"learning_rate": 1.639023035892978e-05,
"loss": 0.4546724796295166,
"memory(GiB)": 76.61,
"step": 945,
"token_acc": 0.876509544215037,
"train_speed(iter/s)": 0.02956
},
{
"epoch": 0.30398784048638056,
"grad_norm": 0.06733149115024578,
"learning_rate": 1.6350202663875385e-05,
"loss": 0.4598522663116455,
"memory(GiB)": 76.61,
"step": 950,
"token_acc": 0.8623452294246177,
"train_speed(iter/s)": 0.029531
},
{
"epoch": 0.3055877764889404,
"grad_norm": 0.05450569676621943,
"learning_rate": 1.6310003740496887e-05,
"loss": 0.4602477550506592,
"memory(GiB)": 76.61,
"step": 955,
"token_acc": 0.8647700701480904,
"train_speed(iter/s)": 0.029548
},
{
"epoch": 0.30718771249150034,
"grad_norm": 0.06736921151917717,
"learning_rate": 1.6269634672727296e-05,
"loss": 0.4589672565460205,
"memory(GiB)": 76.61,
"step": 960,
"token_acc": 0.877502001601281,
"train_speed(iter/s)": 0.029536
},
{
"epoch": 0.30878764849406026,
"grad_norm": 0.06166660436042404,
"learning_rate": 1.6229096549087434e-05,
"loss": 0.4601268291473389,
"memory(GiB)": 76.61,
"step": 965,
"token_acc": 0.8723534201954397,
"train_speed(iter/s)": 0.029518
},
{
"epoch": 0.31038758449662013,
"grad_norm": 0.055128746386822226,
"learning_rate": 1.618839046265658e-05,
"loss": 0.4666788101196289,
"memory(GiB)": 76.61,
"step": 970,
"token_acc": 0.8550563360689943,
"train_speed(iter/s)": 0.029541
},
{
"epoch": 0.31198752049918005,
"grad_norm": 0.056867326711030626,
"learning_rate": 1.614751751104301e-05,
"loss": 0.4646125793457031,
"memory(GiB)": 76.61,
"step": 975,
"token_acc": 0.8651571964234208,
"train_speed(iter/s)": 0.029524
},
{
"epoch": 0.3135874565017399,
"grad_norm": 0.05501107287069041,
"learning_rate": 1.6106478796354382e-05,
"loss": 0.4588280200958252,
"memory(GiB)": 76.61,
"step": 980,
"token_acc": 0.8767766331985918,
"train_speed(iter/s)": 0.029517
},
{
"epoch": 0.31518739250429983,
"grad_norm": 0.08099201898186387,
"learning_rate": 1.6065275425168034e-05,
"loss": 0.4589373111724854,
"memory(GiB)": 76.61,
"step": 985,
"token_acc": 0.8917890157694399,
"train_speed(iter/s)": 0.029526
},
{
"epoch": 0.3167873285068597,
"grad_norm": 0.0522899382710734,
"learning_rate": 1.602390850850113e-05,
"loss": 0.46761279106140136,
"memory(GiB)": 76.61,
"step": 990,
"token_acc": 0.8461229409401366,
"train_speed(iter/s)": 0.029505
},
{
"epoch": 0.3183872645094196,
"grad_norm": 0.05838858698011934,
"learning_rate": 1.5982379161780722e-05,
"loss": 0.44941887855529783,
"memory(GiB)": 76.61,
"step": 995,
"token_acc": 0.8547228871294421,
"train_speed(iter/s)": 0.029511
},
{
"epoch": 0.31998720051197954,
"grad_norm": 0.054930484370324516,
"learning_rate": 1.5940688504813664e-05,
"loss": 0.4591392517089844,
"memory(GiB)": 76.61,
"step": 1000,
"token_acc": 0.8995555555555556,
"train_speed(iter/s)": 0.029505
},
{
"epoch": 0.31998720051197954,
"eval_loss": 0.671963095664978,
"eval_runtime": 110.8694,
"eval_samples_per_second": 181.186,
"eval_steps_per_second": 0.911,
"eval_token_acc": 0.8676077802864524,
"step": 1000
},
{
"epoch": 0.3215871365145394,
"grad_norm": 0.0578985798516978,
"learning_rate": 1.5898837661756405e-05,
"loss": 0.46222972869873047,
"memory(GiB)": 76.61,
"step": 1005,
"token_acc": 0.8840002569208042,
"train_speed(iter/s)": 0.029425
},
{
"epoch": 0.3231870725170993,
"grad_norm": 0.05872050053297838,
"learning_rate": 1.5856827761084698e-05,
"loss": 0.45543718338012695,
"memory(GiB)": 76.61,
"step": 1010,
"token_acc": 0.8753668220265838,
"train_speed(iter/s)": 0.02945
},
{
"epoch": 0.3247870085196592,
"grad_norm": 0.05268695066428434,
"learning_rate": 1.5814659935563165e-05,
"loss": 0.46614727973937986,
"memory(GiB)": 76.61,
"step": 1015,
"token_acc": 0.8792250035355678,
"train_speed(iter/s)": 0.029474
},
{
"epoch": 0.3263869445222191,
"grad_norm": 0.059454673806441594,
"learning_rate": 1.577233532221474e-05,
"loss": 0.45902605056762696,
"memory(GiB)": 76.61,
"step": 1020,
"token_acc": 0.86709886547812,
"train_speed(iter/s)": 0.029475
},
{
"epoch": 0.32798688052477903,
"grad_norm": 0.053728974295076275,
"learning_rate": 1.5729855062290024e-05,
"loss": 0.46491541862487795,
"memory(GiB)": 76.61,
"step": 1025,
"token_acc": 0.8708870261478794,
"train_speed(iter/s)": 0.029469
},
{
"epoch": 0.3295868165273389,
"grad_norm": 0.07030309576814114,
"learning_rate": 1.568722030123651e-05,
"loss": 0.453840970993042,
"memory(GiB)": 76.61,
"step": 1030,
"token_acc": 0.8568111455108359,
"train_speed(iter/s)": 0.029496
},
{
"epoch": 0.3311867525298988,
"grad_norm": 0.07385415365022158,
"learning_rate": 1.5644432188667695e-05,
"loss": 0.45582828521728513,
"memory(GiB)": 76.61,
"step": 1035,
"token_acc": 0.8800162140251317,
"train_speed(iter/s)": 0.029488
},
{
"epoch": 0.3327866885324587,
"grad_norm": 0.05407863995123405,
"learning_rate": 1.5601491878332077e-05,
"loss": 0.4665637969970703,
"memory(GiB)": 76.61,
"step": 1040,
"token_acc": 0.8628481345244351,
"train_speed(iter/s)": 0.029487
},
{
"epoch": 0.3343866245350186,
"grad_norm": 0.05879461372080454,
"learning_rate": 1.5558400528082057e-05,
"loss": 0.4657593250274658,
"memory(GiB)": 76.61,
"step": 1045,
"token_acc": 0.879185119574845,
"train_speed(iter/s)": 0.02951
},
{
"epoch": 0.33598656053757847,
"grad_norm": 0.06618244368029796,
"learning_rate": 1.551515929984271e-05,
"loss": 0.45760574340820315,
"memory(GiB)": 76.61,
"step": 1050,
"token_acc": 0.8899380348185305,
"train_speed(iter/s)": 0.029502
},
{
"epoch": 0.3375864965401384,
"grad_norm": 0.06388796415692906,
"learning_rate": 1.547176935958044e-05,
"loss": 0.46065597534179686,
"memory(GiB)": 76.61,
"step": 1055,
"token_acc": 0.8536853685368537,
"train_speed(iter/s)": 0.029524
},
{
"epoch": 0.3391864325426983,
"grad_norm": 0.05811152365312673,
"learning_rate": 1.5428231877271584e-05,
"loss": 0.46312780380249025,
"memory(GiB)": 76.61,
"step": 1060,
"token_acc": 0.8520375161707633,
"train_speed(iter/s)": 0.029515
},
{
"epoch": 0.3407863685452582,
"grad_norm": 0.05545936328508829,
"learning_rate": 1.538454802687081e-05,
"loss": 0.4615220546722412,
"memory(GiB)": 76.61,
"step": 1065,
"token_acc": 0.8744265080713679,
"train_speed(iter/s)": 0.029504
},
{
"epoch": 0.3423863045478181,
"grad_norm": 0.05964362984731802,
"learning_rate": 1.5340718986279505e-05,
"loss": 0.46706466674804686,
"memory(GiB)": 76.61,
"step": 1070,
"token_acc": 0.8592233009708737,
"train_speed(iter/s)": 0.029536
},
{
"epoch": 0.34398624055037796,
"grad_norm": 0.05356886450328198,
"learning_rate": 1.529674593731399e-05,
"loss": 0.45301499366760256,
"memory(GiB)": 76.61,
"step": 1075,
"token_acc": 0.8575192096597146,
"train_speed(iter/s)": 0.029526
},
{
"epoch": 0.3455861765529379,
"grad_norm": 0.05995962073425321,
"learning_rate": 1.5252630065673662e-05,
"loss": 0.46819314956665037,
"memory(GiB)": 76.61,
"step": 1080,
"token_acc": 0.8875031814711123,
"train_speed(iter/s)": 0.029518
},
{
"epoch": 0.3471861125554978,
"grad_norm": 0.05389432634852101,
"learning_rate": 1.5208372560909031e-05,
"loss": 0.46298394203186033,
"memory(GiB)": 76.61,
"step": 1085,
"token_acc": 0.8872426699937617,
"train_speed(iter/s)": 0.029543
},
{
"epoch": 0.34878604855805767,
"grad_norm": 0.06642390255342462,
"learning_rate": 1.5163974616389621e-05,
"loss": 0.45978522300720215,
"memory(GiB)": 76.61,
"step": 1090,
"token_acc": 0.8246258860593332,
"train_speed(iter/s)": 0.029525
},
{
"epoch": 0.3503859845606176,
"grad_norm": 0.06115184110491886,
"learning_rate": 1.5119437429271813e-05,
"loss": 0.4637304782867432,
"memory(GiB)": 76.61,
"step": 1095,
"token_acc": 0.8666082895504962,
"train_speed(iter/s)": 0.029534
},
{
"epoch": 0.35198592056317746,
"grad_norm": 0.060865150660591956,
"learning_rate": 1.5074762200466557e-05,
"loss": 0.4542848587036133,
"memory(GiB)": 76.61,
"step": 1100,
"token_acc": 0.8913602663035255,
"train_speed(iter/s)": 0.029544
},
{
"epoch": 0.3535858565657374,
"grad_norm": 0.057666943430007674,
"learning_rate": 1.5029950134606991e-05,
"loss": 0.4574248790740967,
"memory(GiB)": 76.61,
"step": 1105,
"token_acc": 0.8634470336597996,
"train_speed(iter/s)": 0.029524
},
{
"epoch": 0.35518579256829724,
"grad_norm": 0.054034554153381265,
"learning_rate": 1.4985002440015959e-05,
"loss": 0.4520272254943848,
"memory(GiB)": 76.61,
"step": 1110,
"token_acc": 0.8674898358680921,
"train_speed(iter/s)": 0.029551
},
{
"epoch": 0.35678572857085716,
"grad_norm": 0.06416854479766453,
"learning_rate": 1.4939920328673422e-05,
"loss": 0.4668846130371094,
"memory(GiB)": 76.61,
"step": 1115,
"token_acc": 0.9170854271356784,
"train_speed(iter/s)": 0.029541
},
{
"epoch": 0.3583856645734171,
"grad_norm": 0.05775941336987237,
"learning_rate": 1.4894705016183803e-05,
"loss": 0.4518620491027832,
"memory(GiB)": 76.61,
"step": 1120,
"token_acc": 0.8672782874617737,
"train_speed(iter/s)": 0.029531
},
{
"epoch": 0.35998560057597695,
"grad_norm": 0.0625175589581686,
"learning_rate": 1.4849357721743169e-05,
"loss": 0.4566941738128662,
"memory(GiB)": 76.61,
"step": 1125,
"token_acc": 0.8505491793163026,
"train_speed(iter/s)": 0.029557
},
{
"epoch": 0.36158553657853687,
"grad_norm": 0.05911529293553411,
"learning_rate": 1.4803879668106393e-05,
"loss": 0.4640664577484131,
"memory(GiB)": 76.61,
"step": 1130,
"token_acc": 0.8772325625117503,
"train_speed(iter/s)": 0.029544
},
{
"epoch": 0.36318547258109674,
"grad_norm": 0.06483783687935218,
"learning_rate": 1.4758272081554168e-05,
"loss": 0.45419878959655763,
"memory(GiB)": 76.61,
"step": 1135,
"token_acc": 0.8594914930223667,
"train_speed(iter/s)": 0.029539
},
{
"epoch": 0.36478540858365666,
"grad_norm": 0.06032730304497941,
"learning_rate": 1.4712536191859934e-05,
"loss": 0.45779004096984866,
"memory(GiB)": 76.61,
"step": 1140,
"token_acc": 0.8938053097345132,
"train_speed(iter/s)": 0.029564
},
{
"epoch": 0.3663853445862166,
"grad_norm": 0.0637380940226065,
"learning_rate": 1.4666673232256738e-05,
"loss": 0.46385722160339354,
"memory(GiB)": 76.61,
"step": 1145,
"token_acc": 0.8621830209481808,
"train_speed(iter/s)": 0.029544
},
{
"epoch": 0.36798528058877644,
"grad_norm": 0.057006770373085346,
"learning_rate": 1.4620684439403962e-05,
"loss": 0.4613553524017334,
"memory(GiB)": 76.61,
"step": 1150,
"token_acc": 0.8831837819873712,
"train_speed(iter/s)": 0.029558
},
{
"epoch": 0.36958521659133636,
"grad_norm": 0.057569299635009126,
"learning_rate": 1.4574571053353987e-05,
"loss": 0.4598341464996338,
"memory(GiB)": 76.61,
"step": 1155,
"token_acc": 0.8825154371140721,
"train_speed(iter/s)": 0.029557
},
{
"epoch": 0.37118515259389623,
"grad_norm": 0.06747695219063263,
"learning_rate": 1.452833431751875e-05,
"loss": 0.4570640563964844,
"memory(GiB)": 76.61,
"step": 1160,
"token_acc": 0.8726823238566132,
"train_speed(iter/s)": 0.029543
},
{
"epoch": 0.37278508859645615,
"grad_norm": 0.05405367649749466,
"learning_rate": 1.448197547863622e-05,
"loss": 0.4516812801361084,
"memory(GiB)": 76.61,
"step": 1165,
"token_acc": 0.8704696273608984,
"train_speed(iter/s)": 0.029568
},
{
"epoch": 0.374385024599016,
"grad_norm": 0.06041157710672601,
"learning_rate": 1.4435495786736796e-05,
"loss": 0.465837287902832,
"memory(GiB)": 76.61,
"step": 1170,
"token_acc": 0.8673412029229904,
"train_speed(iter/s)": 0.029554
},
{
"epoch": 0.37598496060157593,
"grad_norm": 0.05229585247228306,
"learning_rate": 1.438889649510956e-05,
"loss": 0.4427653789520264,
"memory(GiB)": 76.61,
"step": 1175,
"token_acc": 0.8558139534883721,
"train_speed(iter/s)": 0.02954
},
{
"epoch": 0.37758489660413586,
"grad_norm": 0.0547875272797444,
"learning_rate": 1.4342178860268523e-05,
"loss": 0.45673260688781736,
"memory(GiB)": 76.61,
"step": 1180,
"token_acc": 0.880563238622077,
"train_speed(iter/s)": 0.029563
},
{
"epoch": 0.3791848326066957,
"grad_norm": 0.0565328006493161,
"learning_rate": 1.4295344141918734e-05,
"loss": 0.46208748817443845,
"memory(GiB)": 76.61,
"step": 1185,
"token_acc": 0.8671328671328671,
"train_speed(iter/s)": 0.029544
},
{
"epoch": 0.38078476860925564,
"grad_norm": 0.062473905403265834,
"learning_rate": 1.4248393602922299e-05,
"loss": 0.46883163452148435,
"memory(GiB)": 76.61,
"step": 1190,
"token_acc": 0.8412252145605209,
"train_speed(iter/s)": 0.029548
},
{
"epoch": 0.3823847046118155,
"grad_norm": 0.05646151042315891,
"learning_rate": 1.420132850926434e-05,
"loss": 0.45732822418212893,
"memory(GiB)": 76.61,
"step": 1195,
"token_acc": 0.8820655966503839,
"train_speed(iter/s)": 0.02956
},
{
"epoch": 0.3839846406143754,
"grad_norm": 0.052981558367052706,
"learning_rate": 1.4154150130018867e-05,
"loss": 0.45579113960266116,
"memory(GiB)": 76.61,
"step": 1200,
"token_acc": 0.8677085226240233,
"train_speed(iter/s)": 0.029546
},
{
"epoch": 0.38558457661693535,
"grad_norm": 0.052315204322432474,
"learning_rate": 1.4106859737314532e-05,
"loss": 0.45348801612854006,
"memory(GiB)": 76.61,
"step": 1205,
"token_acc": 0.8616187989556136,
"train_speed(iter/s)": 0.029561
},
{
"epoch": 0.3871845126194952,
"grad_norm": 0.05319888084520812,
"learning_rate": 1.4059458606300358e-05,
"loss": 0.45279593467712403,
"memory(GiB)": 76.61,
"step": 1210,
"token_acc": 0.86090645233311,
"train_speed(iter/s)": 0.029565
},
{
"epoch": 0.38878444862205513,
"grad_norm": 0.054475973938428034,
"learning_rate": 1.4011948015111334e-05,
"loss": 0.4616706848144531,
"memory(GiB)": 76.61,
"step": 1215,
"token_acc": 0.8390133684805121,
"train_speed(iter/s)": 0.029549
},
{
"epoch": 0.390384384624615,
"grad_norm": 0.054891067059900926,
"learning_rate": 1.396432924483396e-05,
"loss": 0.4553243637084961,
"memory(GiB)": 76.61,
"step": 1220,
"token_acc": 0.8715350793347353,
"train_speed(iter/s)": 0.029571
},
{
"epoch": 0.3919843206271749,
"grad_norm": 0.06058246643434403,
"learning_rate": 1.3916603579471705e-05,
"loss": 0.47067904472351074,
"memory(GiB)": 76.61,
"step": 1225,
"token_acc": 0.8662144337667232,
"train_speed(iter/s)": 0.029556
},
{
"epoch": 0.3935842566297348,
"grad_norm": 0.05715510214651738,
"learning_rate": 1.3868772305910376e-05,
"loss": 0.46147928237915037,
"memory(GiB)": 76.61,
"step": 1230,
"token_acc": 0.868918375552875,
"train_speed(iter/s)": 0.029548
},
{
"epoch": 0.3951841926322947,
"grad_norm": 0.06593047910666934,
"learning_rate": 1.3820836713883424e-05,
"loss": 0.45935769081115724,
"memory(GiB)": 76.61,
"step": 1235,
"token_acc": 0.8596291476903057,
"train_speed(iter/s)": 0.02957
},
{
"epoch": 0.3967841286348546,
"grad_norm": 0.056071042953882384,
"learning_rate": 1.3772798095937172e-05,
"loss": 0.4495890140533447,
"memory(GiB)": 76.61,
"step": 1240,
"token_acc": 0.8471917163476623,
"train_speed(iter/s)": 0.029553
},
{
"epoch": 0.3983840646374145,
"grad_norm": 0.05810589720196263,
"learning_rate": 1.3724657747395957e-05,
"loss": 0.4619898319244385,
"memory(GiB)": 76.61,
"step": 1245,
"token_acc": 0.8691186216037111,
"train_speed(iter/s)": 0.029561
},
{
"epoch": 0.3999840006399744,
"grad_norm": 0.055604926632171425,
"learning_rate": 1.3676416966327201e-05,
"loss": 0.4587514400482178,
"memory(GiB)": 76.61,
"step": 1250,
"token_acc": 0.8369355461211887,
"train_speed(iter/s)": 0.029564
},
{
"epoch": 0.3999840006399744,
"eval_loss": 0.6690404415130615,
"eval_runtime": 106.3444,
"eval_samples_per_second": 188.896,
"eval_steps_per_second": 0.95,
"eval_token_acc": 0.8683678146748934,
"step": 1250
},
{
"epoch": 0.4015839366425343,
"grad_norm": 0.04782987834900457,
"learning_rate": 1.362807705350641e-05,
"loss": 0.46315851211547854,
"memory(GiB)": 76.61,
"step": 1255,
"token_acc": 0.8767961498796838,
"train_speed(iter/s)": 0.029512
},
{
"epoch": 0.4031838726450942,
"grad_norm": 0.05995996443795485,
"learning_rate": 1.3579639312382105e-05,
"loss": 0.46349530220031737,
"memory(GiB)": 76.61,
"step": 1260,
"token_acc": 0.8588617886178862,
"train_speed(iter/s)": 0.029524
},
{
"epoch": 0.4047838086476541,
"grad_norm": 0.06488882353036057,
"learning_rate": 1.3531105049040667e-05,
"loss": 0.45726447105407714,
"memory(GiB)": 76.61,
"step": 1265,
"token_acc": 0.8802249582003344,
"train_speed(iter/s)": 0.029543
},
{
"epoch": 0.406383744650214,
"grad_norm": 0.05350128050935312,
"learning_rate": 1.3482475572171132e-05,
"loss": 0.4516806125640869,
"memory(GiB)": 76.61,
"step": 1270,
"token_acc": 0.8560765550239234,
"train_speed(iter/s)": 0.029549
},
{
"epoch": 0.4079836806527739,
"grad_norm": 0.05672697687392494,
"learning_rate": 1.3433752193029888e-05,
"loss": 0.46581568717956545,
"memory(GiB)": 76.61,
"step": 1275,
"token_acc": 0.8881742738589211,
"train_speed(iter/s)": 0.029547
},
{
"epoch": 0.40958361665533377,
"grad_norm": 0.0598115330947421,
"learning_rate": 1.3384936225405326e-05,
"loss": 0.46333680152893064,
"memory(GiB)": 76.61,
"step": 1280,
"token_acc": 0.8608710985716804,
"train_speed(iter/s)": 0.029573
},
{
"epoch": 0.4111835526578937,
"grad_norm": 0.05384417907735887,
"learning_rate": 1.333602898558242e-05,
"loss": 0.4611030578613281,
"memory(GiB)": 76.61,
"step": 1285,
"token_acc": 0.8845689770746749,
"train_speed(iter/s)": 0.029567
},
{
"epoch": 0.41278348866045356,
"grad_norm": 0.06043637267465684,
"learning_rate": 1.3287031792307226e-05,
"loss": 0.46013875007629396,
"memory(GiB)": 76.61,
"step": 1290,
"token_acc": 0.870195210303884,
"train_speed(iter/s)": 0.029565
},
{
"epoch": 0.4143834246630135,
"grad_norm": 0.06140603532631629,
"learning_rate": 1.323794596675132e-05,
"loss": 0.45681238174438477,
"memory(GiB)": 76.61,
"step": 1295,
"token_acc": 0.8450012281994596,
"train_speed(iter/s)": 0.029583
},
{
"epoch": 0.4159833606655734,
"grad_norm": 0.062077229851937275,
"learning_rate": 1.318877283247619e-05,
"loss": 0.4490199565887451,
"memory(GiB)": 76.61,
"step": 1300,
"token_acc": 0.89259877573734,
"train_speed(iter/s)": 0.029573
},
{
"epoch": 0.41758329666813326,
"grad_norm": 0.05216177276902916,
"learning_rate": 1.3139513715397521e-05,
"loss": 0.45108351707458494,
"memory(GiB)": 76.61,
"step": 1305,
"token_acc": 0.8547701815372731,
"train_speed(iter/s)": 0.029594
},
{
"epoch": 0.4191832326706932,
"grad_norm": 0.05738628087610287,
"learning_rate": 1.3090169943749475e-05,
"loss": 0.4558550834655762,
"memory(GiB)": 76.61,
"step": 1310,
"token_acc": 0.840696686491079,
"train_speed(iter/s)": 0.029587
},
{
"epoch": 0.42078316867325305,
"grad_norm": 0.05518036740697275,
"learning_rate": 1.304074284804885e-05,
"loss": 0.4631648063659668,
"memory(GiB)": 76.61,
"step": 1315,
"token_acc": 0.8788111708941839,
"train_speed(iter/s)": 0.029578
},
{
"epoch": 0.42238310467581297,
"grad_norm": 0.05902492258138098,
"learning_rate": 1.2991233761059214e-05,
"loss": 0.45921921730041504,
"memory(GiB)": 76.61,
"step": 1320,
"token_acc": 0.866059646344682,
"train_speed(iter/s)": 0.029604
},
{
"epoch": 0.4239830406783729,
"grad_norm": 0.059535437419073044,
"learning_rate": 1.2941644017754964e-05,
"loss": 0.46445517539978026,
"memory(GiB)": 76.61,
"step": 1325,
"token_acc": 0.8831345826235094,
"train_speed(iter/s)": 0.029591
},
{
"epoch": 0.42558297668093276,
"grad_norm": 0.04863893443696892,
"learning_rate": 1.289197495528534e-05,
"loss": 0.45836362838745115,
"memory(GiB)": 76.61,
"step": 1330,
"token_acc": 0.8986429177268872,
"train_speed(iter/s)": 0.029582
},
{
"epoch": 0.4271829126834927,
"grad_norm": 0.05945822860509985,
"learning_rate": 1.284222791293836e-05,
"loss": 0.45783252716064454,
"memory(GiB)": 76.61,
"step": 1335,
"token_acc": 0.8500874125874126,
"train_speed(iter/s)": 0.029596
},
{
"epoch": 0.42878284868605254,
"grad_norm": 0.05989616737178823,
"learning_rate": 1.2792404232104699e-05,
"loss": 0.45293269157409666,
"memory(GiB)": 76.61,
"step": 1340,
"token_acc": 0.8637480798771121,
"train_speed(iter/s)": 0.029584
},
{
"epoch": 0.43038278468861246,
"grad_norm": 0.0586629819404024,
"learning_rate": 1.2742505256241543e-05,
"loss": 0.45876450538635255,
"memory(GiB)": 76.61,
"step": 1345,
"token_acc": 0.8296499119890475,
"train_speed(iter/s)": 0.029588
},
{
"epoch": 0.43198272069117233,
"grad_norm": 0.052924904785980484,
"learning_rate": 1.2692532330836346e-05,
"loss": 0.45821080207824705,
"memory(GiB)": 76.61,
"step": 1350,
"token_acc": 0.8636084374360025,
"train_speed(iter/s)": 0.029594
},
{
"epoch": 0.43358265669373225,
"grad_norm": 0.059304249814977644,
"learning_rate": 1.2642486803370553e-05,
"loss": 0.45485148429870603,
"memory(GiB)": 76.61,
"step": 1355,
"token_acc": 0.8686690223792697,
"train_speed(iter/s)": 0.029579
},
{
"epoch": 0.43518259269629217,
"grad_norm": 0.06253442360689314,
"learning_rate": 1.2592370023283268e-05,
"loss": 0.45198469161987304,
"memory(GiB)": 76.61,
"step": 1360,
"token_acc": 0.8737075332348597,
"train_speed(iter/s)": 0.0296
},
{
"epoch": 0.43678252869885204,
"grad_norm": 0.05314091037792793,
"learning_rate": 1.2542183341934873e-05,
"loss": 0.4516898155212402,
"memory(GiB)": 76.61,
"step": 1365,
"token_acc": 0.8714476021314387,
"train_speed(iter/s)": 0.029596
},
{
"epoch": 0.43838246470141196,
"grad_norm": 0.06014404788689081,
"learning_rate": 1.2491928112570568e-05,
"loss": 0.45399184226989747,
"memory(GiB)": 76.61,
"step": 1370,
"token_acc": 0.8657097288676237,
"train_speed(iter/s)": 0.029583
},
{
"epoch": 0.4399824007039718,
"grad_norm": 0.05910144328100835,
"learning_rate": 1.2441605690283915e-05,
"loss": 0.4607128143310547,
"memory(GiB)": 76.61,
"step": 1375,
"token_acc": 0.8990952307928232,
"train_speed(iter/s)": 0.029603
},
{
"epoch": 0.44158233670653174,
"grad_norm": 0.059073628736854025,
"learning_rate": 1.2391217431980273e-05,
"loss": 0.4515543937683105,
"memory(GiB)": 76.61,
"step": 1380,
"token_acc": 0.9016349860428021,
"train_speed(iter/s)": 0.029591
},
{
"epoch": 0.44318227270909166,
"grad_norm": 0.058358968679540275,
"learning_rate": 1.234076469634022e-05,
"loss": 0.45762925148010253,
"memory(GiB)": 76.61,
"step": 1385,
"token_acc": 0.8919261822376009,
"train_speed(iter/s)": 0.029584
},
{
"epoch": 0.4447822087116515,
"grad_norm": 0.0672513399669503,
"learning_rate": 1.2290248843782915e-05,
"loss": 0.44803729057312014,
"memory(GiB)": 76.61,
"step": 1390,
"token_acc": 0.8975998070196599,
"train_speed(iter/s)": 0.029597
},
{
"epoch": 0.44638214471421145,
"grad_norm": 0.05793114375836921,
"learning_rate": 1.2239671236429413e-05,
"loss": 0.4537235736846924,
"memory(GiB)": 76.61,
"step": 1395,
"token_acc": 0.8839514422541486,
"train_speed(iter/s)": 0.02958
},
{
"epoch": 0.4479820807167713,
"grad_norm": 0.05955306099185102,
"learning_rate": 1.218903323806595e-05,
"loss": 0.4573692798614502,
"memory(GiB)": 76.61,
"step": 1400,
"token_acc": 0.8418099547511312,
"train_speed(iter/s)": 0.029594
},
{
"epoch": 0.44958201671933123,
"grad_norm": 0.058484796569864064,
"learning_rate": 1.2138336214107148e-05,
"loss": 0.44894704818725584,
"memory(GiB)": 76.61,
"step": 1405,
"token_acc": 0.8525200458190149,
"train_speed(iter/s)": 0.029594
},
{
"epoch": 0.4511819527218911,
"grad_norm": 0.05092836798588581,
"learning_rate": 1.2087581531559208e-05,
"loss": 0.45393967628479004,
"memory(GiB)": 76.61,
"step": 1410,
"token_acc": 0.8791390728476821,
"train_speed(iter/s)": 0.02958
},
{
"epoch": 0.452781888724451,
"grad_norm": 0.07033477253264378,
"learning_rate": 1.2036770558983067e-05,
"loss": 0.45307221412658694,
"memory(GiB)": 76.61,
"step": 1415,
"token_acc": 0.8387482900136799,
"train_speed(iter/s)": 0.029599
},
{
"epoch": 0.45438182472701094,
"grad_norm": 0.05966547548288182,
"learning_rate": 1.1985904666457455e-05,
"loss": 0.455959415435791,
"memory(GiB)": 76.61,
"step": 1420,
"token_acc": 0.9042096902303416,
"train_speed(iter/s)": 0.029583
},
{
"epoch": 0.4559817607295708,
"grad_norm": 0.08159145764722696,
"learning_rate": 1.1934985225541998e-05,
"loss": 0.462065601348877,
"memory(GiB)": 76.61,
"step": 1425,
"token_acc": 0.885252444621832,
"train_speed(iter/s)": 0.029573
},
{
"epoch": 0.4575816967321307,
"grad_norm": 0.05540814227664117,
"learning_rate": 1.18840136092402e-05,
"loss": 0.4551572322845459,
"memory(GiB)": 76.61,
"step": 1430,
"token_acc": 0.8559651934966797,
"train_speed(iter/s)": 0.029592
},
{
"epoch": 0.4591816327346906,
"grad_norm": 0.05534004007067895,
"learning_rate": 1.1832991191962435e-05,
"loss": 0.4455368995666504,
"memory(GiB)": 76.61,
"step": 1435,
"token_acc": 0.875560538116592,
"train_speed(iter/s)": 0.029576
},
{
"epoch": 0.4607815687372505,
"grad_norm": 0.058276771895487044,
"learning_rate": 1.1781919349488894e-05,
"loss": 0.4590908527374268,
"memory(GiB)": 76.61,
"step": 1440,
"token_acc": 0.8510418460478733,
"train_speed(iter/s)": 0.029576
},
{
"epoch": 0.46238150473981043,
"grad_norm": 0.05839975543902795,
"learning_rate": 1.1730799458932473e-05,
"loss": 0.462816858291626,
"memory(GiB)": 76.61,
"step": 1445,
"token_acc": 0.9052378085490669,
"train_speed(iter/s)": 0.029586
},
{
"epoch": 0.4639814407423703,
"grad_norm": 0.07084434546926481,
"learning_rate": 1.1679632898701649e-05,
"loss": 0.4550295829772949,
"memory(GiB)": 76.61,
"step": 1450,
"token_acc": 0.8805626598465474,
"train_speed(iter/s)": 0.029572
},
{
"epoch": 0.4655813767449302,
"grad_norm": 0.06519996046237972,
"learning_rate": 1.1628421048463315e-05,
"loss": 0.46291208267211914,
"memory(GiB)": 76.61,
"step": 1455,
"token_acc": 0.8565744600227359,
"train_speed(iter/s)": 0.029581
},
{
"epoch": 0.4671813127474901,
"grad_norm": 0.05799269979733804,
"learning_rate": 1.1577165289105565e-05,
"loss": 0.4474311351776123,
"memory(GiB)": 76.61,
"step": 1460,
"token_acc": 0.8579789309403043,
"train_speed(iter/s)": 0.029568
},
{
"epoch": 0.46878124875005,
"grad_norm": 0.057120675003187855,
"learning_rate": 1.1525867002700484e-05,
"loss": 0.46109714508056643,
"memory(GiB)": 76.61,
"step": 1465,
"token_acc": 0.8752182516587126,
"train_speed(iter/s)": 0.029548
},
{
"epoch": 0.47038118475260987,
"grad_norm": 0.05696370798749074,
"learning_rate": 1.1474527572466847e-05,
"loss": 0.4501948833465576,
"memory(GiB)": 76.61,
"step": 1470,
"token_acc": 0.8529032258064516,
"train_speed(iter/s)": 0.029562
},
{
"epoch": 0.4719811207551698,
"grad_norm": 0.05518112754329221,
"learning_rate": 1.1423148382732854e-05,
"loss": 0.45941987037658694,
"memory(GiB)": 76.61,
"step": 1475,
"token_acc": 0.9009282399143164,
"train_speed(iter/s)": 0.029545
},
{
"epoch": 0.4735810567577297,
"grad_norm": 0.051496444525703684,
"learning_rate": 1.1371730818898785e-05,
"loss": 0.45296878814697267,
"memory(GiB)": 76.61,
"step": 1480,
"token_acc": 0.8814303638644918,
"train_speed(iter/s)": 0.029538
},
{
"epoch": 0.4751809927602896,
"grad_norm": 0.0677105428949175,
"learning_rate": 1.132027626739965e-05,
"loss": 0.45635080337524414,
"memory(GiB)": 76.61,
"step": 1485,
"token_acc": 0.880248833592535,
"train_speed(iter/s)": 0.029546
},
{
"epoch": 0.4767809287628495,
"grad_norm": 0.0673509631098402,
"learning_rate": 1.1268786115667798e-05,
"loss": 0.4614115715026855,
"memory(GiB)": 76.61,
"step": 1490,
"token_acc": 0.8609592251210748,
"train_speed(iter/s)": 0.029525
},
{
"epoch": 0.47838086476540936,
"grad_norm": 0.053337771378298794,
"learning_rate": 1.1217261752095518e-05,
"loss": 0.45500664710998534,
"memory(GiB)": 76.61,
"step": 1495,
"token_acc": 0.8794466403162056,
"train_speed(iter/s)": 0.029522
},
{
"epoch": 0.4799808007679693,
"grad_norm": 0.05429302474155136,
"learning_rate": 1.1165704565997593e-05,
"loss": 0.44763407707214353,
"memory(GiB)": 76.61,
"step": 1500,
"token_acc": 0.8700440528634361,
"train_speed(iter/s)": 0.02952
},
{
"epoch": 0.4799808007679693,
"eval_loss": 0.6668144464492798,
"eval_runtime": 124.2589,
"eval_samples_per_second": 161.662,
"eval_steps_per_second": 0.813,
"eval_token_acc": 0.8694583558206896,
"step": 1500
},
{
"epoch": 0.4815807367705292,
"grad_norm": 0.06146640524587408,
"learning_rate": 1.1114115947573834e-05,
"loss": 0.45711498260498046,
"memory(GiB)": 76.61,
"step": 1505,
"token_acc": 0.8695166967121641,
"train_speed(iter/s)": 0.029461
},
{
"epoch": 0.48318067277308907,
"grad_norm": 0.06183782693437151,
"learning_rate": 1.1062497287871606e-05,
"loss": 0.4499336242675781,
"memory(GiB)": 76.61,
"step": 1510,
"token_acc": 0.8487557381009906,
"train_speed(iter/s)": 0.029468
},
{
"epoch": 0.484780608775649,
"grad_norm": 0.056753269624682155,
"learning_rate": 1.1010849978748314e-05,
"loss": 0.4551094055175781,
"memory(GiB)": 76.61,
"step": 1515,
"token_acc": 0.8579035448045033,
"train_speed(iter/s)": 0.02948
},
{
"epoch": 0.48638054477820886,
"grad_norm": 0.05188962595699218,
"learning_rate": 1.0959175412833869e-05,
"loss": 0.4483503818511963,
"memory(GiB)": 76.61,
"step": 1520,
"token_acc": 0.8649127992905705,
"train_speed(iter/s)": 0.029477
},
{
"epoch": 0.4879804807807688,
"grad_norm": 0.0526414480661873,
"learning_rate": 1.0907474983493144e-05,
"loss": 0.45140752792358396,
"memory(GiB)": 76.61,
"step": 1525,
"token_acc": 0.8700204290091931,
"train_speed(iter/s)": 0.029464
},
{
"epoch": 0.48958041678332864,
"grad_norm": 0.06682159988119828,
"learning_rate": 1.08557500847884e-05,
"loss": 0.4480952262878418,
"memory(GiB)": 76.61,
"step": 1530,
"token_acc": 0.8906385187748745,
"train_speed(iter/s)": 0.029478
},
{
"epoch": 0.49118035278588856,
"grad_norm": 0.06117568492897364,
"learning_rate": 1.080400211144169e-05,
"loss": 0.453688907623291,
"memory(GiB)": 76.61,
"step": 1535,
"token_acc": 0.8130096719135217,
"train_speed(iter/s)": 0.029469
},
{
"epoch": 0.4927802887884485,
"grad_norm": 0.05143211947513191,
"learning_rate": 1.0752232458797262e-05,
"loss": 0.44568753242492676,
"memory(GiB)": 76.61,
"step": 1540,
"token_acc": 0.847257743677181,
"train_speed(iter/s)": 0.029464
},
{
"epoch": 0.49438022479100835,
"grad_norm": 0.05201971134010435,
"learning_rate": 1.070044252278393e-05,
"loss": 0.46500363349914553,
"memory(GiB)": 76.61,
"step": 1545,
"token_acc": 0.84,
"train_speed(iter/s)": 0.029474
},
{
"epoch": 0.49598016079356827,
"grad_norm": 0.05304880581645989,
"learning_rate": 1.064863369987743e-05,
"loss": 0.4501206398010254,
"memory(GiB)": 76.61,
"step": 1550,
"token_acc": 0.8888641920426762,
"train_speed(iter/s)": 0.029465
},
{
"epoch": 0.49758009679612814,
"grad_norm": 0.050584443072610216,
"learning_rate": 1.0596807387062772e-05,
"loss": 0.456621789932251,
"memory(GiB)": 76.61,
"step": 1555,
"token_acc": 0.8793768317137128,
"train_speed(iter/s)": 0.02948
},
{
"epoch": 0.49918003279868806,
"grad_norm": 0.05907676168100355,
"learning_rate": 1.0544964981796563e-05,
"loss": 0.4567122936248779,
"memory(GiB)": 76.61,
"step": 1560,
"token_acc": 0.8505747126436781,
"train_speed(iter/s)": 0.029477
},
{
"epoch": 0.500779968801248,
"grad_norm": 0.055037989511506104,
"learning_rate": 1.0493107881969335e-05,
"loss": 0.44720020294189455,
"memory(GiB)": 76.61,
"step": 1565,
"token_acc": 0.8853304383227032,
"train_speed(iter/s)": 0.029466
},
{
"epoch": 0.5023799048038079,
"grad_norm": 0.0597376748229471,
"learning_rate": 1.0441237485867845e-05,
"loss": 0.4492997169494629,
"memory(GiB)": 76.61,
"step": 1570,
"token_acc": 0.8809347181008902,
"train_speed(iter/s)": 0.029489
},
{
"epoch": 0.5039798408063677,
"grad_norm": 0.060265741182571844,
"learning_rate": 1.0389355192137379e-05,
"loss": 0.4525942325592041,
"memory(GiB)": 76.61,
"step": 1575,
"token_acc": 0.8839541547277937,
"train_speed(iter/s)": 0.029481
},
{
"epoch": 0.5055797768089276,
"grad_norm": 0.06015007204584338,
"learning_rate": 1.0337462399744025e-05,
"loss": 0.4606604099273682,
"memory(GiB)": 76.61,
"step": 1580,
"token_acc": 0.8439696373348328,
"train_speed(iter/s)": 0.029471
},
{
"epoch": 0.5071797128114875,
"grad_norm": 0.0539606724017438,
"learning_rate": 1.0285560507936962e-05,
"loss": 0.46471481323242186,
"memory(GiB)": 76.61,
"step": 1585,
"token_acc": 0.8212732305258995,
"train_speed(iter/s)": 0.029486
},
{
"epoch": 0.5087796488140475,
"grad_norm": 0.0588254805138369,
"learning_rate": 1.0233650916210736e-05,
"loss": 0.45154604911804197,
"memory(GiB)": 76.61,
"step": 1590,
"token_acc": 0.883357041251778,
"train_speed(iter/s)": 0.029474
},
{
"epoch": 0.5103795848166074,
"grad_norm": 0.06409304780777438,
"learning_rate": 1.0181735024267504e-05,
"loss": 0.45000271797180175,
"memory(GiB)": 76.61,
"step": 1595,
"token_acc": 0.8340197693574959,
"train_speed(iter/s)": 0.029485
},
{
"epoch": 0.5119795208191672,
"grad_norm": 0.058200459243944895,
"learning_rate": 1.012981423197931e-05,
"loss": 0.4608008861541748,
"memory(GiB)": 76.61,
"step": 1600,
"token_acc": 0.8793913904007917,
"train_speed(iter/s)": 0.029484
},
{
"epoch": 0.5135794568217271,
"grad_norm": 0.052541653818392466,
"learning_rate": 1.007788993935033e-05,
"loss": 0.45448942184448243,
"memory(GiB)": 76.61,
"step": 1605,
"token_acc": 0.8615229110512129,
"train_speed(iter/s)": 0.029472
},
{
"epoch": 0.515179392824287,
"grad_norm": 0.06526426191856917,
"learning_rate": 1.002596354647912e-05,
"loss": 0.45614914894104003,
"memory(GiB)": 76.61,
"step": 1610,
"token_acc": 0.871312462372065,
"train_speed(iter/s)": 0.029489
},
{
"epoch": 0.516779328826847,
"grad_norm": 0.05743481751682084,
"learning_rate": 9.974036453520881e-06,
"loss": 0.447450590133667,
"memory(GiB)": 76.61,
"step": 1615,
"token_acc": 0.8760352658295485,
"train_speed(iter/s)": 0.02948
},
{
"epoch": 0.5183792648294068,
"grad_norm": 0.06577378911981274,
"learning_rate": 9.922110060649672e-06,
"loss": 0.45809640884399416,
"memory(GiB)": 76.61,
"step": 1620,
"token_acc": 0.9048299514146899,
"train_speed(iter/s)": 0.029468
},
{
"epoch": 0.5199792008319667,
"grad_norm": 0.05180626448607971,
"learning_rate": 9.870185768020694e-06,
"loss": 0.4360641002655029,
"memory(GiB)": 76.61,
"step": 1625,
"token_acc": 0.890797148412184,
"train_speed(iter/s)": 0.029481
},
{
"epoch": 0.5215791368345266,
"grad_norm": 0.048795347699578454,
"learning_rate": 9.818264975732497e-06,
"loss": 0.4505919933319092,
"memory(GiB)": 76.61,
"step": 1630,
"token_acc": 0.8830155979202773,
"train_speed(iter/s)": 0.029462
},
{
"epoch": 0.5231790728370865,
"grad_norm": 0.054325754690003863,
"learning_rate": 9.766349083789266e-06,
"loss": 0.4518167495727539,
"memory(GiB)": 76.61,
"step": 1635,
"token_acc": 0.8740636704119851,
"train_speed(iter/s)": 0.029458
},
{
"epoch": 0.5247790088396465,
"grad_norm": 0.05473270131153146,
"learning_rate": 9.71443949206304e-06,
"loss": 0.4629377841949463,
"memory(GiB)": 76.61,
"step": 1640,
"token_acc": 0.8927648578811369,
"train_speed(iter/s)": 0.029469
},
{
"epoch": 0.5263789448422063,
"grad_norm": 0.05962553624327487,
"learning_rate": 9.662537600255979e-06,
"loss": 0.4535552501678467,
"memory(GiB)": 76.61,
"step": 1645,
"token_acc": 0.8980960623106881,
"train_speed(iter/s)": 0.029457
},
{
"epoch": 0.5279788808447662,
"grad_norm": 0.06367541172972058,
"learning_rate": 9.610644807862625e-06,
"loss": 0.44418978691101074,
"memory(GiB)": 76.61,
"step": 1650,
"token_acc": 0.8769371011850501,
"train_speed(iter/s)": 0.029468
},
{
"epoch": 0.5295788168473261,
"grad_norm": 0.05288367644088033,
"learning_rate": 9.558762514132157e-06,
"loss": 0.4513704299926758,
"memory(GiB)": 76.61,
"step": 1655,
"token_acc": 0.8576478906434126,
"train_speed(iter/s)": 0.029464
},
{
"epoch": 0.531178752849886,
"grad_norm": 0.054919719691940275,
"learning_rate": 9.506892118030668e-06,
"loss": 0.4454075336456299,
"memory(GiB)": 76.61,
"step": 1660,
"token_acc": 0.8535078688042359,
"train_speed(iter/s)": 0.029456
},
{
"epoch": 0.532778688852446,
"grad_norm": 0.0561391497524031,
"learning_rate": 9.455035018203439e-06,
"loss": 0.4459484100341797,
"memory(GiB)": 76.61,
"step": 1665,
"token_acc": 0.8793124922157181,
"train_speed(iter/s)": 0.029471
},
{
"epoch": 0.5343786248550058,
"grad_norm": 0.051526152917333715,
"learning_rate": 9.40319261293723e-06,
"loss": 0.4593966484069824,
"memory(GiB)": 76.61,
"step": 1670,
"token_acc": 0.8957568638966378,
"train_speed(iter/s)": 0.029466
},
{
"epoch": 0.5359785608575657,
"grad_norm": 0.05336577516465571,
"learning_rate": 9.351366300122569e-06,
"loss": 0.45195541381835935,
"memory(GiB)": 76.61,
"step": 1675,
"token_acc": 0.8254120659305488,
"train_speed(iter/s)": 0.029459
},
{
"epoch": 0.5375784968601256,
"grad_norm": 0.05605931671991975,
"learning_rate": 9.299557477216073e-06,
"loss": 0.4473400115966797,
"memory(GiB)": 76.61,
"step": 1680,
"token_acc": 0.8684433164128595,
"train_speed(iter/s)": 0.029474
},
{
"epoch": 0.5391784328626855,
"grad_norm": 0.05284398220938548,
"learning_rate": 9.247767541202738e-06,
"loss": 0.4539934158325195,
"memory(GiB)": 76.61,
"step": 1685,
"token_acc": 0.8787515006002401,
"train_speed(iter/s)": 0.029458
},
{
"epoch": 0.5407783688652454,
"grad_norm": 0.06074778175058899,
"learning_rate": 9.195997888558312e-06,
"loss": 0.4540121078491211,
"memory(GiB)": 76.61,
"step": 1690,
"token_acc": 0.882076702321941,
"train_speed(iter/s)": 0.029458
},
{
"epoch": 0.5423783048678052,
"grad_norm": 0.06072504661929311,
"learning_rate": 9.144249915211605e-06,
"loss": 0.45176243782043457,
"memory(GiB)": 76.61,
"step": 1695,
"token_acc": 0.8652057386094908,
"train_speed(iter/s)": 0.029465
},
{
"epoch": 0.5439782408703652,
"grad_norm": 0.058552695609385315,
"learning_rate": 9.092525016506858e-06,
"loss": 0.4491862773895264,
"memory(GiB)": 76.61,
"step": 1700,
"token_acc": 0.8822588020118884,
"train_speed(iter/s)": 0.02945
},
{
"epoch": 0.5455781768729251,
"grad_norm": 0.056892490634495974,
"learning_rate": 9.040824587166136e-06,
"loss": 0.45043745040893557,
"memory(GiB)": 76.61,
"step": 1705,
"token_acc": 0.8825789923142613,
"train_speed(iter/s)": 0.029461
},
{
"epoch": 0.547178112875485,
"grad_norm": 0.05885692671807609,
"learning_rate": 8.98915002125169e-06,
"loss": 0.4475353240966797,
"memory(GiB)": 76.61,
"step": 1710,
"token_acc": 0.8721031538595574,
"train_speed(iter/s)": 0.029454
},
{
"epoch": 0.5487780488780449,
"grad_norm": 0.060276094585736115,
"learning_rate": 8.9375027121284e-06,
"loss": 0.4502556800842285,
"memory(GiB)": 76.61,
"step": 1715,
"token_acc": 0.8562842259917189,
"train_speed(iter/s)": 0.029449
},
{
"epoch": 0.5503779848806047,
"grad_norm": 0.06782068962590707,
"learning_rate": 8.885884052426168e-06,
"loss": 0.4532322883605957,
"memory(GiB)": 76.61,
"step": 1720,
"token_acc": 0.8593545573484518,
"train_speed(iter/s)": 0.029466
},
{
"epoch": 0.5519779208831647,
"grad_norm": 0.06070839045848377,
"learning_rate": 8.83429543400241e-06,
"loss": 0.45258092880249023,
"memory(GiB)": 76.61,
"step": 1725,
"token_acc": 0.8751242791807516,
"train_speed(iter/s)": 0.029452
},
{
"epoch": 0.5535778568857246,
"grad_norm": 0.049979952181739715,
"learning_rate": 8.78273824790448e-06,
"loss": 0.4340657234191895,
"memory(GiB)": 76.61,
"step": 1730,
"token_acc": 0.8650843222985634,
"train_speed(iter/s)": 0.029451
},
{
"epoch": 0.5551777928882845,
"grad_norm": 0.059124658124222323,
"learning_rate": 8.731213884332205e-06,
"loss": 0.43556828498840333,
"memory(GiB)": 76.61,
"step": 1735,
"token_acc": 0.8524390243902439,
"train_speed(iter/s)": 0.029459
},
{
"epoch": 0.5567777288908443,
"grad_norm": 0.05228309031135195,
"learning_rate": 8.679723732600355e-06,
"loss": 0.4483633041381836,
"memory(GiB)": 76.61,
"step": 1740,
"token_acc": 0.9039166284928997,
"train_speed(iter/s)": 0.029445
},
{
"epoch": 0.5583776648934042,
"grad_norm": 0.05659321921396489,
"learning_rate": 8.628269181101216e-06,
"loss": 0.45377864837646487,
"memory(GiB)": 76.61,
"step": 1745,
"token_acc": 0.8812897628687102,
"train_speed(iter/s)": 0.029449
},
{
"epoch": 0.5599776008959642,
"grad_norm": 0.0610469666222746,
"learning_rate": 8.576851617267151e-06,
"loss": 0.4495216369628906,
"memory(GiB)": 76.61,
"step": 1750,
"token_acc": 0.8734145104008117,
"train_speed(iter/s)": 0.029452
},
{
"epoch": 0.5599776008959642,
"eval_loss": 0.6640093922615051,
"eval_runtime": 114.9985,
"eval_samples_per_second": 174.681,
"eval_steps_per_second": 0.878,
"eval_token_acc": 0.8701128116616424,
"step": 1750
},
{
"epoch": 0.5615775368985241,
"grad_norm": 0.061306770256929585,
"learning_rate": 8.525472427533156e-06,
"loss": 0.44908857345581055,
"memory(GiB)": 77.63,
"step": 1755,
"token_acc": 0.8715457946180765,
"train_speed(iter/s)": 0.029409
},
{
"epoch": 0.563177472901084,
"grad_norm": 0.05235639827008535,
"learning_rate": 8.474132997299521e-06,
"loss": 0.4579316139221191,
"memory(GiB)": 77.63,
"step": 1760,
"token_acc": 0.8922923256201098,
"train_speed(iter/s)": 0.029422
},
{
"epoch": 0.5647774089036438,
"grad_norm": 0.051281007426132216,
"learning_rate": 8.422834710894434e-06,
"loss": 0.45467004776000974,
"memory(GiB)": 77.63,
"step": 1765,
"token_acc": 0.903878366189924,
"train_speed(iter/s)": 0.029438
},
{
"epoch": 0.5663773449062037,
"grad_norm": 0.05049109520782513,
"learning_rate": 8.371578951536689e-06,
"loss": 0.45294957160949706,
"memory(GiB)": 77.63,
"step": 1770,
"token_acc": 0.8928110202324581,
"train_speed(iter/s)": 0.029439
},
{
"epoch": 0.5679772809087636,
"grad_norm": 0.04946427707777728,
"learning_rate": 8.320367101298351e-06,
"loss": 0.4473431587219238,
"memory(GiB)": 77.63,
"step": 1775,
"token_acc": 0.8723599632690542,
"train_speed(iter/s)": 0.029439
},
{
"epoch": 0.5695772169113236,
"grad_norm": 0.053606352244487274,
"learning_rate": 8.26920054106753e-06,
"loss": 0.4495864391326904,
"memory(GiB)": 77.63,
"step": 1780,
"token_acc": 0.8844315111203492,
"train_speed(iter/s)": 0.029459
},
{
"epoch": 0.5711771529138835,
"grad_norm": 0.05525614374940963,
"learning_rate": 8.218080650511107e-06,
"loss": 0.44890499114990234,
"memory(GiB)": 77.63,
"step": 1785,
"token_acc": 0.8749736453721273,
"train_speed(iter/s)": 0.02946
},
{
"epoch": 0.5727770889164433,
"grad_norm": 0.05882148265537131,
"learning_rate": 8.167008808037568e-06,
"loss": 0.44676194190979,
"memory(GiB)": 77.63,
"step": 1790,
"token_acc": 0.8807511737089202,
"train_speed(iter/s)": 0.029457
},
{
"epoch": 0.5743770249190032,
"grad_norm": 0.048821121641334515,
"learning_rate": 8.115986390759805e-06,
"loss": 0.4417415142059326,
"memory(GiB)": 77.63,
"step": 1795,
"token_acc": 0.8531673379714391,
"train_speed(iter/s)": 0.029469
},
{
"epoch": 0.5759769609215631,
"grad_norm": 0.054949264031140505,
"learning_rate": 8.065014774458004e-06,
"loss": 0.46439437866210936,
"memory(GiB)": 77.63,
"step": 1800,
"token_acc": 0.8333022213925705,
"train_speed(iter/s)": 0.029467
},
{
"epoch": 0.5775768969241231,
"grad_norm": 0.059507220518762304,
"learning_rate": 8.014095333542548e-06,
"loss": 0.4539642333984375,
"memory(GiB)": 77.63,
"step": 1805,
"token_acc": 0.8577178858942948,
"train_speed(iter/s)": 0.029483
},
{
"epoch": 0.579176832926683,
"grad_norm": 0.05302143027350534,
"learning_rate": 7.963229441016938e-06,
"loss": 0.4606470108032227,
"memory(GiB)": 77.63,
"step": 1810,
"token_acc": 0.8760775862068966,
"train_speed(iter/s)": 0.029484
},
{
"epoch": 0.5807767689292428,
"grad_norm": 0.0699581228289572,
"learning_rate": 7.912418468440794e-06,
"loss": 0.4488551139831543,
"memory(GiB)": 77.63,
"step": 1815,
"token_acc": 0.8892276422764228,
"train_speed(iter/s)": 0.029481
},
{
"epoch": 0.5823767049318027,
"grad_norm": 0.053456667148895104,
"learning_rate": 7.861663785892857e-06,
"loss": 0.45035881996154786,
"memory(GiB)": 77.63,
"step": 1820,
"token_acc": 0.8806643202815662,
"train_speed(iter/s)": 0.029498
},
{
"epoch": 0.5839766409343626,
"grad_norm": 0.05451209338463787,
"learning_rate": 7.810966761934053e-06,
"loss": 0.44800753593444825,
"memory(GiB)": 77.63,
"step": 1825,
"token_acc": 0.8771571298819255,
"train_speed(iter/s)": 0.029496
},
{
"epoch": 0.5855765769369226,
"grad_norm": 0.05912934985203241,
"learning_rate": 7.760328763570589e-06,
"loss": 0.4499057769775391,
"memory(GiB)": 77.63,
"step": 1830,
"token_acc": 0.868710326675956,
"train_speed(iter/s)": 0.029487
},
{
"epoch": 0.5871765129394825,
"grad_norm": 0.052841905445767515,
"learning_rate": 7.709751156217088e-06,
"loss": 0.4497323989868164,
"memory(GiB)": 77.63,
"step": 1835,
"token_acc": 0.8117094325984822,
"train_speed(iter/s)": 0.029501
},
{
"epoch": 0.5887764489420423,
"grad_norm": 0.060076953422732254,
"learning_rate": 7.659235303659784e-06,
"loss": 0.4582187652587891,
"memory(GiB)": 77.63,
"step": 1840,
"token_acc": 0.8795674258561363,
"train_speed(iter/s)": 0.029492
},
{
"epoch": 0.5903763849446022,
"grad_norm": 0.06307528499562465,
"learning_rate": 7.608782568019729e-06,
"loss": 0.4430552005767822,
"memory(GiB)": 77.63,
"step": 1845,
"token_acc": 0.8452444922084901,
"train_speed(iter/s)": 0.029498
},
{
"epoch": 0.5919763209471621,
"grad_norm": 0.05378691938628143,
"learning_rate": 7.558394309716088e-06,
"loss": 0.459810209274292,
"memory(GiB)": 77.63,
"step": 1850,
"token_acc": 0.8506092736192435,
"train_speed(iter/s)": 0.029503
},
{
"epoch": 0.593576256949722,
"grad_norm": 0.0586506530143339,
"learning_rate": 7.508071887429433e-06,
"loss": 0.46239190101623534,
"memory(GiB)": 77.63,
"step": 1855,
"token_acc": 0.9115304709141274,
"train_speed(iter/s)": 0.029495
},
{
"epoch": 0.5951761929522819,
"grad_norm": 0.053290473896441634,
"learning_rate": 7.4578166580651335e-06,
"loss": 0.4524221897125244,
"memory(GiB)": 77.63,
"step": 1860,
"token_acc": 0.8817879571481345,
"train_speed(iter/s)": 0.029508
},
{
"epoch": 0.5967761289548418,
"grad_norm": 0.051901913358510056,
"learning_rate": 7.4076299767167325e-06,
"loss": 0.4579325675964355,
"memory(GiB)": 77.63,
"step": 1865,
"token_acc": 0.8617401668653158,
"train_speed(iter/s)": 0.029506
},
{
"epoch": 0.5983760649574017,
"grad_norm": 0.05256077511072294,
"learning_rate": 7.35751319662945e-06,
"loss": 0.45406513214111327,
"memory(GiB)": 77.63,
"step": 1870,
"token_acc": 0.8924402944873406,
"train_speed(iter/s)": 0.029497
},
{
"epoch": 0.5999760009599616,
"grad_norm": 0.056121622843709036,
"learning_rate": 7.307467669163655e-06,
"loss": 0.450104284286499,
"memory(GiB)": 77.63,
"step": 1875,
"token_acc": 0.8646184340931615,
"train_speed(iter/s)": 0.02951
},
{
"epoch": 0.6015759369625215,
"grad_norm": 0.051068951060234354,
"learning_rate": 7.25749474375846e-06,
"loss": 0.45695791244506834,
"memory(GiB)": 77.63,
"step": 1880,
"token_acc": 0.9112655568126717,
"train_speed(iter/s)": 0.029503
},
{
"epoch": 0.6031758729650813,
"grad_norm": 0.05120698584703106,
"learning_rate": 7.207595767895303e-06,
"loss": 0.4460740089416504,
"memory(GiB)": 77.63,
"step": 1885,
"token_acc": 0.8637192342752963,
"train_speed(iter/s)": 0.029499
},
{
"epoch": 0.6047758089676413,
"grad_norm": 0.05826366701259215,
"learning_rate": 7.157772087061645e-06,
"loss": 0.4498391628265381,
"memory(GiB)": 77.63,
"step": 1890,
"token_acc": 0.8602477214302408,
"train_speed(iter/s)": 0.029509
},
{
"epoch": 0.6063757449702012,
"grad_norm": 0.05454678875604061,
"learning_rate": 7.108025044714661e-06,
"loss": 0.44768247604370115,
"memory(GiB)": 77.63,
"step": 1895,
"token_acc": 0.8998014357721094,
"train_speed(iter/s)": 0.029496
},
{
"epoch": 0.6079756809727611,
"grad_norm": 0.04862560763785379,
"learning_rate": 7.058355982245038e-06,
"loss": 0.44283151626586914,
"memory(GiB)": 77.63,
"step": 1900,
"token_acc": 0.8749580958766343,
"train_speed(iter/s)": 0.029501
},
{
"epoch": 0.609575616975321,
"grad_norm": 0.05390239428952395,
"learning_rate": 7.00876623894079e-06,
"loss": 0.4445077419281006,
"memory(GiB)": 77.63,
"step": 1905,
"token_acc": 0.8588156123822341,
"train_speed(iter/s)": 0.029504
},
{
"epoch": 0.6111755529778808,
"grad_norm": 0.052917745372876655,
"learning_rate": 6.959257151951153e-06,
"loss": 0.45001955032348634,
"memory(GiB)": 77.63,
"step": 1910,
"token_acc": 0.8768155911013054,
"train_speed(iter/s)": 0.029494
},
{
"epoch": 0.6127754889804408,
"grad_norm": 0.05432256049056495,
"learning_rate": 6.909830056250527e-06,
"loss": 0.44944238662719727,
"memory(GiB)": 77.63,
"step": 1915,
"token_acc": 0.8941244909831297,
"train_speed(iter/s)": 0.029507
},
{
"epoch": 0.6143754249830007,
"grad_norm": 0.05852297407331436,
"learning_rate": 6.860486284602479e-06,
"loss": 0.4477729797363281,
"memory(GiB)": 77.63,
"step": 1920,
"token_acc": 0.8854845719252499,
"train_speed(iter/s)": 0.029501
},
{
"epoch": 0.6159753609855606,
"grad_norm": 0.05474007805899836,
"learning_rate": 6.8112271675238154e-06,
"loss": 0.4501204013824463,
"memory(GiB)": 77.63,
"step": 1925,
"token_acc": 0.8803290949887809,
"train_speed(iter/s)": 0.029496
},
{
"epoch": 0.6175752969881205,
"grad_norm": 0.05545012433641634,
"learning_rate": 6.762054033248681e-06,
"loss": 0.44565958976745607,
"memory(GiB)": 77.63,
"step": 1930,
"token_acc": 0.8480542195015304,
"train_speed(iter/s)": 0.029507
},
{
"epoch": 0.6191752329906803,
"grad_norm": 0.05495247298953925,
"learning_rate": 6.712968207692778e-06,
"loss": 0.44170804023742677,
"memory(GiB)": 77.63,
"step": 1935,
"token_acc": 0.8709073900841908,
"train_speed(iter/s)": 0.029498
},
{
"epoch": 0.6207751689932403,
"grad_norm": 0.05792014047889592,
"learning_rate": 6.663971014417585e-06,
"loss": 0.4454016208648682,
"memory(GiB)": 77.63,
"step": 1940,
"token_acc": 0.8606651376146789,
"train_speed(iter/s)": 0.0295
},
{
"epoch": 0.6223751049958002,
"grad_norm": 0.04853659131630362,
"learning_rate": 6.615063774594677e-06,
"loss": 0.4387532711029053,
"memory(GiB)": 77.63,
"step": 1945,
"token_acc": 0.8920454545454546,
"train_speed(iter/s)": 0.029507
},
{
"epoch": 0.6239750409983601,
"grad_norm": 0.05266495974136303,
"learning_rate": 6.566247806970119e-06,
"loss": 0.4472493171691895,
"memory(GiB)": 77.63,
"step": 1950,
"token_acc": 0.848505251817937,
"train_speed(iter/s)": 0.029497
},
{
"epoch": 0.62557497700092,
"grad_norm": 0.054994759694813,
"learning_rate": 6.5175244278288705e-06,
"loss": 0.44487895965576174,
"memory(GiB)": 77.63,
"step": 1955,
"token_acc": 0.8689320388349514,
"train_speed(iter/s)": 0.029506
},
{
"epoch": 0.6271749130034798,
"grad_norm": 0.057067387083368365,
"learning_rate": 6.468894950959336e-06,
"loss": 0.4466127395629883,
"memory(GiB)": 77.63,
"step": 1960,
"token_acc": 0.846737755286463,
"train_speed(iter/s)": 0.0295
},
{
"epoch": 0.6287748490060397,
"grad_norm": 0.05417940634204734,
"learning_rate": 6.420360687617897e-06,
"loss": 0.44883151054382325,
"memory(GiB)": 77.63,
"step": 1965,
"token_acc": 0.8795967892477132,
"train_speed(iter/s)": 0.02949
},
{
"epoch": 0.6303747850085997,
"grad_norm": 0.05176488752108695,
"learning_rate": 6.3719229464935915e-06,
"loss": 0.4542849063873291,
"memory(GiB)": 77.63,
"step": 1970,
"token_acc": 0.8568893191352049,
"train_speed(iter/s)": 0.029503
},
{
"epoch": 0.6319747210111596,
"grad_norm": 0.04867135924369273,
"learning_rate": 6.323583033672799e-06,
"loss": 0.44331774711608884,
"memory(GiB)": 77.63,
"step": 1975,
"token_acc": 0.8647865559204172,
"train_speed(iter/s)": 0.029491
},
{
"epoch": 0.6335746570137194,
"grad_norm": 0.06076783884358601,
"learning_rate": 6.275342252604044e-06,
"loss": 0.44751858711242676,
"memory(GiB)": 77.63,
"step": 1980,
"token_acc": 0.871765773944621,
"train_speed(iter/s)": 0.029486
},
{
"epoch": 0.6351745930162793,
"grad_norm": 0.0520886449098567,
"learning_rate": 6.22720190406283e-06,
"loss": 0.46150927543640136,
"memory(GiB)": 77.63,
"step": 1985,
"token_acc": 0.8921661480178595,
"train_speed(iter/s)": 0.029497
},
{
"epoch": 0.6367745290188392,
"grad_norm": 0.058090405193780774,
"learning_rate": 6.179163286116581e-06,
"loss": 0.44019436836242676,
"memory(GiB)": 77.63,
"step": 1990,
"token_acc": 0.9157033805888768,
"train_speed(iter/s)": 0.029488
},
{
"epoch": 0.6383744650213992,
"grad_norm": 0.057472120727550105,
"learning_rate": 6.13122769408963e-06,
"loss": 0.4466409683227539,
"memory(GiB)": 77.63,
"step": 1995,
"token_acc": 0.8608313968499871,
"train_speed(iter/s)": 0.029492
},
{
"epoch": 0.6399744010239591,
"grad_norm": 0.05665485079826101,
"learning_rate": 6.083396420528298e-06,
"loss": 0.451153039932251,
"memory(GiB)": 77.63,
"step": 2000,
"token_acc": 0.8910367046369808,
"train_speed(iter/s)": 0.029496
},
{
"epoch": 0.6399744010239591,
"eval_loss": 0.6622327566146851,
"eval_runtime": 115.9166,
"eval_samples_per_second": 173.297,
"eval_steps_per_second": 0.871,
"eval_token_acc": 0.8709580958089251,
"step": 2000
},
{
"epoch": 0.6415743370265189,
"grad_norm": 0.05247711719910443,
"learning_rate": 6.0356707551660434e-06,
"loss": 0.45055346488952636,
"memory(GiB)": 77.63,
"step": 2005,
"token_acc": 0.8961562482257424,
"train_speed(iter/s)": 0.029453
},
{
"epoch": 0.6431742730290788,
"grad_norm": 0.0533769803289562,
"learning_rate": 5.988051984888668e-06,
"loss": 0.4436792373657227,
"memory(GiB)": 77.63,
"step": 2010,
"token_acc": 0.8894836272040302,
"train_speed(iter/s)": 0.029463
},
{
"epoch": 0.6447742090316387,
"grad_norm": 0.05428299581856707,
"learning_rate": 5.940541393699646e-06,
"loss": 0.44562363624572754,
"memory(GiB)": 77.63,
"step": 2015,
"token_acc": 0.8804031789106416,
"train_speed(iter/s)": 0.029477
},
{
"epoch": 0.6463741450341987,
"grad_norm": 0.055930584071511934,
"learning_rate": 5.893140262685469e-06,
"loss": 0.4412201404571533,
"memory(GiB)": 77.63,
"step": 2020,
"token_acc": 0.8791348600508906,
"train_speed(iter/s)": 0.029482
},
{
"epoch": 0.6479740810367586,
"grad_norm": 0.06077731970466293,
"learning_rate": 5.845849869981137e-06,
"loss": 0.44964237213134767,
"memory(GiB)": 77.63,
"step": 2025,
"token_acc": 0.8710450018908358,
"train_speed(iter/s)": 0.029484
},
{
"epoch": 0.6495740170393184,
"grad_norm": 0.05824848510516177,
"learning_rate": 5.7986714907356614e-06,
"loss": 0.4586543083190918,
"memory(GiB)": 77.63,
"step": 2030,
"token_acc": 0.8852591792656588,
"train_speed(iter/s)": 0.029498
},
{
"epoch": 0.6511739530418783,
"grad_norm": 0.06066761562869553,
"learning_rate": 5.751606397077703e-06,
"loss": 0.44632205963134763,
"memory(GiB)": 77.63,
"step": 2035,
"token_acc": 0.8871352785145888,
"train_speed(iter/s)": 0.029494
},
{
"epoch": 0.6527738890444382,
"grad_norm": 0.055201144436432543,
"learning_rate": 5.704655858081268e-06,
"loss": 0.43164916038513185,
"memory(GiB)": 77.63,
"step": 2040,
"token_acc": 0.8937977909940527,
"train_speed(iter/s)": 0.029496
},
{
"epoch": 0.6543738250469981,
"grad_norm": 0.05987076771844116,
"learning_rate": 5.6578211397314765e-06,
"loss": 0.4560856819152832,
"memory(GiB)": 77.63,
"step": 2045,
"token_acc": 0.8462914545204349,
"train_speed(iter/s)": 0.029507
},
{
"epoch": 0.6559737610495581,
"grad_norm": 0.05549604363093839,
"learning_rate": 5.611103504890444e-06,
"loss": 0.44809746742248535,
"memory(GiB)": 77.63,
"step": 2050,
"token_acc": 0.8783254279232832,
"train_speed(iter/s)": 0.029503
},
{
"epoch": 0.6575736970521179,
"grad_norm": 0.060605872447174955,
"learning_rate": 5.564504213263205e-06,
"loss": 0.43492536544799804,
"memory(GiB)": 77.63,
"step": 2055,
"token_acc": 0.8383036405886909,
"train_speed(iter/s)": 0.029515
},
{
"epoch": 0.6591736330546778,
"grad_norm": 0.05003885513998493,
"learning_rate": 5.5180245213637785e-06,
"loss": 0.44741315841674806,
"memory(GiB)": 77.63,
"step": 2060,
"token_acc": 0.8784857874174862,
"train_speed(iter/s)": 0.029515
},
{
"epoch": 0.6607735690572377,
"grad_norm": 0.053248832099036005,
"learning_rate": 5.4716656824812505e-06,
"loss": 0.4469279766082764,
"memory(GiB)": 77.63,
"step": 2065,
"token_acc": 0.8853107344632768,
"train_speed(iter/s)": 0.029507
},
{
"epoch": 0.6623735050597976,
"grad_norm": 0.050937098304756526,
"learning_rate": 5.425428946646016e-06,
"loss": 0.44948582649230956,
"memory(GiB)": 77.63,
"step": 2070,
"token_acc": 0.8934210526315789,
"train_speed(iter/s)": 0.029526
},
{
"epoch": 0.6639734410623576,
"grad_norm": 0.050018563981163396,
"learning_rate": 5.379315560596038e-06,
"loss": 0.4475410461425781,
"memory(GiB)": 77.63,
"step": 2075,
"token_acc": 0.8478816857555876,
"train_speed(iter/s)": 0.029518
},
{
"epoch": 0.6655733770649174,
"grad_norm": 0.05745892430696422,
"learning_rate": 5.333326767743263e-06,
"loss": 0.45008225440979005,
"memory(GiB)": 77.63,
"step": 2080,
"token_acc": 0.8264099454214675,
"train_speed(iter/s)": 0.029515
},
{
"epoch": 0.6671733130674773,
"grad_norm": 0.05536507134607956,
"learning_rate": 5.287463808140069e-06,
"loss": 0.4393789291381836,
"memory(GiB)": 77.63,
"step": 2085,
"token_acc": 0.8450008816787162,
"train_speed(iter/s)": 0.029527
},
{
"epoch": 0.6687732490700372,
"grad_norm": 0.06142641017026178,
"learning_rate": 5.241727918445836e-06,
"loss": 0.4437687873840332,
"memory(GiB)": 77.63,
"step": 2090,
"token_acc": 0.8837277242185217,
"train_speed(iter/s)": 0.02952
},
{
"epoch": 0.6703731850725971,
"grad_norm": 0.05426196603270913,
"learning_rate": 5.1961203318936116e-06,
"loss": 0.4427367687225342,
"memory(GiB)": 77.63,
"step": 2095,
"token_acc": 0.856048805815161,
"train_speed(iter/s)": 0.029524
},
{
"epoch": 0.6719731210751569,
"grad_norm": 0.054169398190345976,
"learning_rate": 5.1506422782568345e-06,
"loss": 0.4520686626434326,
"memory(GiB)": 77.63,
"step": 2100,
"token_acc": 0.8747133027522935,
"train_speed(iter/s)": 0.029529
},
{
"epoch": 0.6735730570777169,
"grad_norm": 0.054436537230257924,
"learning_rate": 5.105294983816203e-06,
"loss": 0.44482645988464353,
"memory(GiB)": 77.63,
"step": 2105,
"token_acc": 0.8637377049180328,
"train_speed(iter/s)": 0.029517
},
{
"epoch": 0.6751729930802768,
"grad_norm": 0.05860088154390529,
"learning_rate": 5.060079671326577e-06,
"loss": 0.44719686508178713,
"memory(GiB)": 77.63,
"step": 2110,
"token_acc": 0.8593150866058442,
"train_speed(iter/s)": 0.029529
},
{
"epoch": 0.6767729290828367,
"grad_norm": 0.05264024149284518,
"learning_rate": 5.014997559984045e-06,
"loss": 0.43972039222717285,
"memory(GiB)": 77.63,
"step": 2115,
"token_acc": 0.8533221194280909,
"train_speed(iter/s)": 0.029526
},
{
"epoch": 0.6783728650853966,
"grad_norm": 0.0534652970629265,
"learning_rate": 4.970049865393009e-06,
"loss": 0.4468375205993652,
"memory(GiB)": 77.63,
"step": 2120,
"token_acc": 0.8628782287822878,
"train_speed(iter/s)": 0.029518
},
{
"epoch": 0.6799728010879564,
"grad_norm": 0.05246927821047006,
"learning_rate": 4.925237799533445e-06,
"loss": 0.4498266696929932,
"memory(GiB)": 77.63,
"step": 2125,
"token_acc": 0.9048205760049284,
"train_speed(iter/s)": 0.029533
},
{
"epoch": 0.6815727370905164,
"grad_norm": 0.05104237083350841,
"learning_rate": 4.880562570728188e-06,
"loss": 0.4389338970184326,
"memory(GiB)": 77.63,
"step": 2130,
"token_acc": 0.8844444444444445,
"train_speed(iter/s)": 0.029525
},
{
"epoch": 0.6831726730930763,
"grad_norm": 0.05297787940328326,
"learning_rate": 4.836025383610382e-06,
"loss": 0.4495584487915039,
"memory(GiB)": 77.63,
"step": 2135,
"token_acc": 0.8647426233038984,
"train_speed(iter/s)": 0.029524
},
{
"epoch": 0.6847726090956362,
"grad_norm": 0.05092547333787791,
"learning_rate": 4.791627439090975e-06,
"loss": 0.4421692848205566,
"memory(GiB)": 77.63,
"step": 2140,
"token_acc": 0.8828041384231181,
"train_speed(iter/s)": 0.029534
},
{
"epoch": 0.6863725450981961,
"grad_norm": 0.053418572817851825,
"learning_rate": 4.74736993432634e-06,
"loss": 0.44208922386169436,
"memory(GiB)": 77.63,
"step": 2145,
"token_acc": 0.8888520238885202,
"train_speed(iter/s)": 0.029522
},
{
"epoch": 0.6879724811007559,
"grad_norm": 0.053760421496406786,
"learning_rate": 4.703254062686017e-06,
"loss": 0.4469425201416016,
"memory(GiB)": 77.63,
"step": 2150,
"token_acc": 0.8764145324597975,
"train_speed(iter/s)": 0.02953
},
{
"epoch": 0.6895724171033158,
"grad_norm": 0.05280613332203591,
"learning_rate": 4.6592810137205e-06,
"loss": 0.45023741722106936,
"memory(GiB)": 77.63,
"step": 2155,
"token_acc": 0.8968010517090271,
"train_speed(iter/s)": 0.029531
},
{
"epoch": 0.6911723531058758,
"grad_norm": 0.05438679052798784,
"learning_rate": 4.615451973129196e-06,
"loss": 0.4470167636871338,
"memory(GiB)": 77.63,
"step": 2160,
"token_acc": 0.8761696818465378,
"train_speed(iter/s)": 0.02952
},
{
"epoch": 0.6927722891084357,
"grad_norm": 0.056426544102266905,
"learning_rate": 4.571768122728421e-06,
"loss": 0.4486443042755127,
"memory(GiB)": 77.63,
"step": 2165,
"token_acc": 0.8781996587030717,
"train_speed(iter/s)": 0.029533
},
{
"epoch": 0.6943722251109956,
"grad_norm": 0.05337656902490804,
"learning_rate": 4.528230640419562e-06,
"loss": 0.4497722625732422,
"memory(GiB)": 77.63,
"step": 2170,
"token_acc": 0.8571428571428571,
"train_speed(iter/s)": 0.029522
},
{
"epoch": 0.6959721611135554,
"grad_norm": 0.054129658100736014,
"learning_rate": 4.4848407001572945e-06,
"loss": 0.44121665954589845,
"memory(GiB)": 77.63,
"step": 2175,
"token_acc": 0.8674278464954012,
"train_speed(iter/s)": 0.029518
},
{
"epoch": 0.6975720971161153,
"grad_norm": 0.05332136363084243,
"learning_rate": 4.441599471917946e-06,
"loss": 0.43872866630554197,
"memory(GiB)": 77.63,
"step": 2180,
"token_acc": 0.8575067664384652,
"train_speed(iter/s)": 0.029529
},
{
"epoch": 0.6991720331186753,
"grad_norm": 0.06093731322456081,
"learning_rate": 4.398508121667925e-06,
"loss": 0.42902402877807616,
"memory(GiB)": 77.63,
"step": 2185,
"token_acc": 0.8526187576126675,
"train_speed(iter/s)": 0.029519
},
{
"epoch": 0.7007719691212352,
"grad_norm": 0.05959591977220614,
"learning_rate": 4.355567811332311e-06,
"loss": 0.44504075050354003,
"memory(GiB)": 77.63,
"step": 2190,
"token_acc": 0.8604511878618487,
"train_speed(iter/s)": 0.029519
},
{
"epoch": 0.702371905123795,
"grad_norm": 0.052994813328955795,
"learning_rate": 4.312779698763493e-06,
"loss": 0.4408130168914795,
"memory(GiB)": 77.63,
"step": 2195,
"token_acc": 0.8787728847105394,
"train_speed(iter/s)": 0.029526
},
{
"epoch": 0.7039718411263549,
"grad_norm": 0.05402320661185779,
"learning_rate": 4.270144937709981e-06,
"loss": 0.4396658897399902,
"memory(GiB)": 77.63,
"step": 2200,
"token_acc": 0.8677248677248677,
"train_speed(iter/s)": 0.029515
},
{
"epoch": 0.7055717771289148,
"grad_norm": 0.053270864469091045,
"learning_rate": 4.227664677785264e-06,
"loss": 0.4493250846862793,
"memory(GiB)": 77.63,
"step": 2205,
"token_acc": 0.8801270542742715,
"train_speed(iter/s)": 0.029523
},
{
"epoch": 0.7071717131314748,
"grad_norm": 0.06248819511930574,
"learning_rate": 4.1853400644368395e-06,
"loss": 0.44740095138549807,
"memory(GiB)": 77.63,
"step": 2210,
"token_acc": 0.8988747408942849,
"train_speed(iter/s)": 0.029514
},
{
"epoch": 0.7087716491340347,
"grad_norm": 0.05807018681919018,
"learning_rate": 4.143172238915302e-06,
"loss": 0.4508991241455078,
"memory(GiB)": 77.63,
"step": 2215,
"token_acc": 0.8739803562510404,
"train_speed(iter/s)": 0.029506
},
{
"epoch": 0.7103715851365945,
"grad_norm": 0.05376168037907147,
"learning_rate": 4.101162338243595e-06,
"loss": 0.4486696243286133,
"memory(GiB)": 77.63,
"step": 2220,
"token_acc": 0.8750778169744761,
"train_speed(iter/s)": 0.029514
},
{
"epoch": 0.7119715211391544,
"grad_norm": 0.05460944158847028,
"learning_rate": 4.059311495186338e-06,
"loss": 0.4484865188598633,
"memory(GiB)": 77.63,
"step": 2225,
"token_acc": 0.8524350054924936,
"train_speed(iter/s)": 0.029506
},
{
"epoch": 0.7135714571417143,
"grad_norm": 0.05716955035585288,
"learning_rate": 4.017620838219276e-06,
"loss": 0.44258599281311034,
"memory(GiB)": 77.63,
"step": 2230,
"token_acc": 0.8597191629955947,
"train_speed(iter/s)": 0.029504
},
{
"epoch": 0.7151713931442742,
"grad_norm": 0.05984913995816041,
"learning_rate": 3.9760914914988716e-06,
"loss": 0.4547589778900146,
"memory(GiB)": 77.63,
"step": 2235,
"token_acc": 0.8679617117117117,
"train_speed(iter/s)": 0.029511
},
{
"epoch": 0.7167713291468342,
"grad_norm": 0.05686589162715874,
"learning_rate": 3.93472457483197e-06,
"loss": 0.4416301727294922,
"memory(GiB)": 77.63,
"step": 2240,
"token_acc": 0.826577064816822,
"train_speed(iter/s)": 0.029498
},
{
"epoch": 0.718371265149394,
"grad_norm": 0.05780707586931182,
"learning_rate": 3.893521203645618e-06,
"loss": 0.45052361488342285,
"memory(GiB)": 77.63,
"step": 2245,
"token_acc": 0.8836182062608028,
"train_speed(iter/s)": 0.0295
},
{
"epoch": 0.7199712011519539,
"grad_norm": 0.049110615928360885,
"learning_rate": 3.852482488956992e-06,
"loss": 0.4427218437194824,
"memory(GiB)": 77.63,
"step": 2250,
"token_acc": 0.8621255642183012,
"train_speed(iter/s)": 0.029501
},
{
"epoch": 0.7199712011519539,
"eval_loss": 0.659950852394104,
"eval_runtime": 108.3142,
"eval_samples_per_second": 185.461,
"eval_steps_per_second": 0.932,
"eval_token_acc": 0.8716289458342705,
"step": 2250
},
{
"epoch": 0.7215711371545138,
"grad_norm": 0.04899882607919235,
"learning_rate": 3.8116095373434204e-06,
"loss": 0.4487879753112793,
"memory(GiB)": 77.63,
"step": 2255,
"token_acc": 0.8912671818368324,
"train_speed(iter/s)": 0.029473
},
{
"epoch": 0.7231710731570737,
"grad_norm": 0.05080548488435112,
"learning_rate": 3.7709034509125706e-06,
"loss": 0.44452829360961915,
"memory(GiB)": 77.63,
"step": 2260,
"token_acc": 0.8442477876106195,
"train_speed(iter/s)": 0.029482
},
{
"epoch": 0.7247710091596337,
"grad_norm": 0.048986009146357284,
"learning_rate": 3.7303653272727057e-06,
"loss": 0.4472095012664795,
"memory(GiB)": 77.63,
"step": 2265,
"token_acc": 0.870567815521944,
"train_speed(iter/s)": 0.029495
},
{
"epoch": 0.7263709451621935,
"grad_norm": 0.05152412916361422,
"learning_rate": 3.689996259503116e-06,
"loss": 0.440493106842041,
"memory(GiB)": 77.63,
"step": 2270,
"token_acc": 0.8795436455293181,
"train_speed(iter/s)": 0.029499
},
{
"epoch": 0.7279708811647534,
"grad_norm": 0.055480142184644934,
"learning_rate": 3.6497973361246153e-06,
"loss": 0.4417555809020996,
"memory(GiB)": 77.63,
"step": 2275,
"token_acc": 0.8660460713158725,
"train_speed(iter/s)": 0.029499
},
{
"epoch": 0.7295708171673133,
"grad_norm": 0.05625540509082736,
"learning_rate": 3.609769641070221e-06,
"loss": 0.4407214164733887,
"memory(GiB)": 77.63,
"step": 2280,
"token_acc": 0.8890608875128999,
"train_speed(iter/s)": 0.029513
},
{
"epoch": 0.7311707531698732,
"grad_norm": 0.05002588428206622,
"learning_rate": 3.569914253655896e-06,
"loss": 0.4413386344909668,
"memory(GiB)": 77.63,
"step": 2285,
"token_acc": 0.8921049390319005,
"train_speed(iter/s)": 0.029513
},
{
"epoch": 0.7327706891724332,
"grad_norm": 0.05088814815973685,
"learning_rate": 3.530232248551466e-06,
"loss": 0.4507819652557373,
"memory(GiB)": 77.63,
"step": 2290,
"token_acc": 0.8278411830895355,
"train_speed(iter/s)": 0.029513
},
{
"epoch": 0.734370625174993,
"grad_norm": 0.05399937134620822,
"learning_rate": 3.4907246957516416e-06,
"loss": 0.4447961330413818,
"memory(GiB)": 77.63,
"step": 2295,
"token_acc": 0.8888263967004124,
"train_speed(iter/s)": 0.029522
},
{
"epoch": 0.7359705611775529,
"grad_norm": 0.06200035405309708,
"learning_rate": 3.4513926605471504e-06,
"loss": 0.45868444442749023,
"memory(GiB)": 77.63,
"step": 2300,
"token_acc": 0.8513141426783479,
"train_speed(iter/s)": 0.02952
},
{
"epoch": 0.7375704971801128,
"grad_norm": 0.055876255912378235,
"learning_rate": 3.412237203496036e-06,
"loss": 0.4431456089019775,
"memory(GiB)": 77.63,
"step": 2305,
"token_acc": 0.8651997041420119,
"train_speed(iter/s)": 0.029532
},
{
"epoch": 0.7391704331826727,
"grad_norm": 0.06032844036632358,
"learning_rate": 3.3732593803950354e-06,
"loss": 0.4452229976654053,
"memory(GiB)": 77.63,
"step": 2310,
"token_acc": 0.8915232899706252,
"train_speed(iter/s)": 0.02953
},
{
"epoch": 0.7407703691852325,
"grad_norm": 0.05255216039270682,
"learning_rate": 3.3344602422511343e-06,
"loss": 0.4414207458496094,
"memory(GiB)": 77.63,
"step": 2315,
"token_acc": 0.8901802257032171,
"train_speed(iter/s)": 0.029522
},
{
"epoch": 0.7423703051877925,
"grad_norm": 0.05535966142690852,
"learning_rate": 3.2958408352532055e-06,
"loss": 0.43938393592834474,
"memory(GiB)": 77.63,
"step": 2320,
"token_acc": 0.8354077253218885,
"train_speed(iter/s)": 0.029536
},
{
"epoch": 0.7439702411903524,
"grad_norm": 0.0505418855319798,
"learning_rate": 3.257402200743821e-06,
"loss": 0.44445362091064455,
"memory(GiB)": 77.63,
"step": 2325,
"token_acc": 0.864262790258637,
"train_speed(iter/s)": 0.02953
},
{
"epoch": 0.7455701771929123,
"grad_norm": 0.0563107101835597,
"learning_rate": 3.2191453751911505e-06,
"loss": 0.45569453239440916,
"memory(GiB)": 77.63,
"step": 2330,
"token_acc": 0.8784313725490196,
"train_speed(iter/s)": 0.029523
},
{
"epoch": 0.7471701131954722,
"grad_norm": 0.05000152007266613,
"learning_rate": 3.1810713901610367e-06,
"loss": 0.4395348072052002,
"memory(GiB)": 77.63,
"step": 2335,
"token_acc": 0.8867111781175964,
"train_speed(iter/s)": 0.029536
},
{
"epoch": 0.748770049198032,
"grad_norm": 0.057169590375126145,
"learning_rate": 3.1431812722891598e-06,
"loss": 0.4397278785705566,
"memory(GiB)": 77.63,
"step": 2340,
"token_acc": 0.8577532891037895,
"train_speed(iter/s)": 0.029529
},
{
"epoch": 0.750369985200592,
"grad_norm": 0.05730745865195846,
"learning_rate": 3.1054760432533626e-06,
"loss": 0.45998029708862304,
"memory(GiB)": 77.63,
"step": 2345,
"token_acc": 0.8845755097339016,
"train_speed(iter/s)": 0.029534
},
{
"epoch": 0.7519699212031519,
"grad_norm": 0.05180470840824953,
"learning_rate": 3.0679567197461135e-06,
"loss": 0.45008273124694825,
"memory(GiB)": 77.63,
"step": 2350,
"token_acc": 0.8394425931535898,
"train_speed(iter/s)": 0.029537
},
{
"epoch": 0.7535698572057118,
"grad_norm": 0.06025883780673481,
"learning_rate": 3.0306243134470668e-06,
"loss": 0.4444745540618896,
"memory(GiB)": 77.63,
"step": 2355,
"token_acc": 0.889631386074585,
"train_speed(iter/s)": 0.02953
},
{
"epoch": 0.7551697932082717,
"grad_norm": 0.05199872680450009,
"learning_rate": 2.993479830995815e-06,
"loss": 0.451768159866333,
"memory(GiB)": 77.63,
"step": 2360,
"token_acc": 0.8736520199581522,
"train_speed(iter/s)": 0.029541
},
{
"epoch": 0.7567697292108315,
"grad_norm": 0.05489027404588469,
"learning_rate": 2.9565242739647115e-06,
"loss": 0.4442115306854248,
"memory(GiB)": 77.63,
"step": 2365,
"token_acc": 0.8865552903739061,
"train_speed(iter/s)": 0.029538
},
{
"epoch": 0.7583696652133914,
"grad_norm": 0.06334021131539457,
"learning_rate": 2.919758638831893e-06,
"loss": 0.4570741653442383,
"memory(GiB)": 77.63,
"step": 2370,
"token_acc": 0.8652606912712361,
"train_speed(iter/s)": 0.029531
},
{
"epoch": 0.7599696012159514,
"grad_norm": 0.053831666314624105,
"learning_rate": 2.8831839169543998e-06,
"loss": 0.44495415687561035,
"memory(GiB)": 77.63,
"step": 2375,
"token_acc": 0.8756476683937824,
"train_speed(iter/s)": 0.029541
},
{
"epoch": 0.7615695372185113,
"grad_norm": 0.0527583973457582,
"learning_rate": 2.84680109454143e-06,
"loss": 0.4472104549407959,
"memory(GiB)": 77.63,
"step": 2380,
"token_acc": 0.8725328947368421,
"train_speed(iter/s)": 0.029536
},
{
"epoch": 0.7631694732210712,
"grad_norm": 0.058941021305098804,
"learning_rate": 2.810611152627777e-06,
"loss": 0.4499720573425293,
"memory(GiB)": 77.63,
"step": 2385,
"token_acc": 0.8632213889794588,
"train_speed(iter/s)": 0.029535
},
{
"epoch": 0.764769409223631,
"grad_norm": 0.05393008170855123,
"learning_rate": 2.774615067047346e-06,
"loss": 0.43872222900390623,
"memory(GiB)": 77.63,
"step": 2390,
"token_acc": 0.8742202234150588,
"train_speed(iter/s)": 0.02954
},
{
"epoch": 0.7663693452261909,
"grad_norm": 0.0556417645809335,
"learning_rate": 2.738813808406866e-06,
"loss": 0.4399220943450928,
"memory(GiB)": 77.63,
"step": 2395,
"token_acc": 0.8997599039615847,
"train_speed(iter/s)": 0.029531
},
{
"epoch": 0.7679692812287509,
"grad_norm": 0.05379432234404155,
"learning_rate": 2.7032083420597e-06,
"loss": 0.4382453441619873,
"memory(GiB)": 77.63,
"step": 2400,
"token_acc": 0.8875784668061633,
"train_speed(iter/s)": 0.029541
},
{
"epoch": 0.7695692172313108,
"grad_norm": 0.05806842163630925,
"learning_rate": 2.667799628079829e-06,
"loss": 0.44454326629638674,
"memory(GiB)": 77.63,
"step": 2405,
"token_acc": 0.8880662020905923,
"train_speed(iter/s)": 0.029535
},
{
"epoch": 0.7711691532338707,
"grad_norm": 0.06148704112133217,
"learning_rate": 2.6325886212359496e-06,
"loss": 0.43945813179016113,
"memory(GiB)": 77.63,
"step": 2410,
"token_acc": 0.8767772511848341,
"train_speed(iter/s)": 0.029527
},
{
"epoch": 0.7727690892364305,
"grad_norm": 0.056530065759685846,
"learning_rate": 2.5975762709657506e-06,
"loss": 0.4438450813293457,
"memory(GiB)": 77.63,
"step": 2415,
"token_acc": 0.8570975416336241,
"train_speed(iter/s)": 0.029538
},
{
"epoch": 0.7743690252389904,
"grad_norm": 0.056107845444701834,
"learning_rate": 2.5627635213502832e-06,
"loss": 0.43836054801940916,
"memory(GiB)": 77.63,
"step": 2420,
"token_acc": 0.8966822253059165,
"train_speed(iter/s)": 0.029532
},
{
"epoch": 0.7759689612415503,
"grad_norm": 0.05796065696017405,
"learning_rate": 2.528151311088537e-06,
"loss": 0.4400279998779297,
"memory(GiB)": 77.63,
"step": 2425,
"token_acc": 0.8552805280528053,
"train_speed(iter/s)": 0.029526
},
{
"epoch": 0.7775688972441103,
"grad_norm": 0.05418546630146028,
"learning_rate": 2.4937405734720964e-06,
"loss": 0.44541444778442385,
"memory(GiB)": 77.63,
"step": 2430,
"token_acc": 0.8620764552562988,
"train_speed(iter/s)": 0.029532
},
{
"epoch": 0.7791688332466701,
"grad_norm": 0.05654250831277805,
"learning_rate": 2.459532236360007e-06,
"loss": 0.43491110801696775,
"memory(GiB)": 77.63,
"step": 2435,
"token_acc": 0.8570184983677911,
"train_speed(iter/s)": 0.029522
},
{
"epoch": 0.78076876924923,
"grad_norm": 0.05526372242621089,
"learning_rate": 2.4255272221537295e-06,
"loss": 0.4378859043121338,
"memory(GiB)": 77.63,
"step": 2440,
"token_acc": 0.8631236857197476,
"train_speed(iter/s)": 0.029526
},
{
"epoch": 0.7823687052517899,
"grad_norm": 0.05404315424969483,
"learning_rate": 2.391726447772279e-06,
"loss": 0.45857391357421873,
"memory(GiB)": 77.63,
"step": 2445,
"token_acc": 0.8634816932081122,
"train_speed(iter/s)": 0.029527
},
{
"epoch": 0.7839686412543498,
"grad_norm": 0.05765113554061621,
"learning_rate": 2.3581308246275103e-06,
"loss": 0.4473139762878418,
"memory(GiB)": 77.63,
"step": 2450,
"token_acc": 0.8979846898922044,
"train_speed(iter/s)": 0.029518
},
{
"epoch": 0.7855685772569098,
"grad_norm": 0.058353221842389495,
"learning_rate": 2.324741258599521e-06,
"loss": 0.44444866180419923,
"memory(GiB)": 77.63,
"step": 2455,
"token_acc": 0.8648913576213038,
"train_speed(iter/s)": 0.029527
},
{
"epoch": 0.7871685132594696,
"grad_norm": 0.05309289512529438,
"learning_rate": 2.29155865001225e-06,
"loss": 0.43857607841491697,
"memory(GiB)": 77.63,
"step": 2460,
"token_acc": 0.894580549368968,
"train_speed(iter/s)": 0.029521
},
{
"epoch": 0.7887684492620295,
"grad_norm": 0.0540412312893473,
"learning_rate": 2.2585838936091753e-06,
"loss": 0.43953213691711424,
"memory(GiB)": 77.63,
"step": 2465,
"token_acc": 0.8868672731513879,
"train_speed(iter/s)": 0.029515
},
{
"epoch": 0.7903683852645894,
"grad_norm": 0.05880520800742855,
"learning_rate": 2.225817878529214e-06,
"loss": 0.4457580089569092,
"memory(GiB)": 77.63,
"step": 2470,
"token_acc": 0.8630282437884901,
"train_speed(iter/s)": 0.029525
},
{
"epoch": 0.7919683212671493,
"grad_norm": 0.058328885138964066,
"learning_rate": 2.1932614882827196e-06,
"loss": 0.4424918174743652,
"memory(GiB)": 77.63,
"step": 2475,
"token_acc": 0.8814697747925722,
"train_speed(iter/s)": 0.029517
},
{
"epoch": 0.7935682572697093,
"grad_norm": 0.05685263085809571,
"learning_rate": 2.160915600727688e-06,
"loss": 0.43921732902526855,
"memory(GiB)": 77.63,
"step": 2480,
"token_acc": 0.913681738109219,
"train_speed(iter/s)": 0.029516
},
{
"epoch": 0.7951681932722691,
"grad_norm": 0.056639845561812056,
"learning_rate": 2.1287810880460636e-06,
"loss": 0.44060502052307127,
"memory(GiB)": 77.63,
"step": 2485,
"token_acc": 0.8829075425790754,
"train_speed(iter/s)": 0.02952
},
{
"epoch": 0.796768129274829,
"grad_norm": 0.05230960490349676,
"learning_rate": 2.0968588167202265e-06,
"loss": 0.43935480117797854,
"memory(GiB)": 77.63,
"step": 2490,
"token_acc": 0.8856997455470738,
"train_speed(iter/s)": 0.029511
},
{
"epoch": 0.7983680652773889,
"grad_norm": 0.05305183045142263,
"learning_rate": 2.0651496475096455e-06,
"loss": 0.4360368728637695,
"memory(GiB)": 77.63,
"step": 2495,
"token_acc": 0.8394655704008221,
"train_speed(iter/s)": 0.029517
},
{
"epoch": 0.7999680012799488,
"grad_norm": 0.05620012484228566,
"learning_rate": 2.03365443542764e-06,
"loss": 0.44507203102111814,
"memory(GiB)": 77.63,
"step": 2500,
"token_acc": 0.8857098429482195,
"train_speed(iter/s)": 0.029515
},
{
"epoch": 0.7999680012799488,
"eval_loss": 0.6586322784423828,
"eval_runtime": 105.1966,
"eval_samples_per_second": 190.957,
"eval_steps_per_second": 0.96,
"eval_token_acc": 0.8721292963419328,
"step": 2500
},
{
"epoch": 0.8015679372825087,
"grad_norm": 0.05541548293022976,
"learning_rate": 2.0023740297183536e-06,
"loss": 0.44654192924499514,
"memory(GiB)": 77.63,
"step": 2505,
"token_acc": 0.8819252077562327,
"train_speed(iter/s)": 0.029489
},
{
"epoch": 0.8031678732850686,
"grad_norm": 0.057374199419424524,
"learning_rate": 1.971309273833828e-06,
"loss": 0.44596128463745116,
"memory(GiB)": 77.63,
"step": 2510,
"token_acc": 0.855553561815898,
"train_speed(iter/s)": 0.029494
},
{
"epoch": 0.8047678092876285,
"grad_norm": 0.05297595418967434,
"learning_rate": 1.940461005411288e-06,
"loss": 0.45099148750305174,
"memory(GiB)": 77.63,
"step": 2515,
"token_acc": 0.8958361962347121,
"train_speed(iter/s)": 0.029503
},
{
"epoch": 0.8063677452901884,
"grad_norm": 0.056714681257095015,
"learning_rate": 1.9098300562505266e-06,
"loss": 0.4423669338226318,
"memory(GiB)": 77.63,
"step": 2520,
"token_acc": 0.879045996592845,
"train_speed(iter/s)": 0.029507
},
{
"epoch": 0.8079676812927483,
"grad_norm": 0.05868047347236466,
"learning_rate": 1.8794172522915022e-06,
"loss": 0.4462554931640625,
"memory(GiB)": 77.63,
"step": 2525,
"token_acc": 0.8811320754716981,
"train_speed(iter/s)": 0.029507
},
{
"epoch": 0.8095676172953082,
"grad_norm": 0.04860202133053876,
"learning_rate": 1.849223413592046e-06,
"loss": 0.4488513946533203,
"memory(GiB)": 77.63,
"step": 2530,
"token_acc": 0.8654490616621984,
"train_speed(iter/s)": 0.029517
},
{
"epoch": 0.811167553297868,
"grad_norm": 0.055005201314914515,
"learning_rate": 1.8192493543057676e-06,
"loss": 0.45094904899597166,
"memory(GiB)": 77.63,
"step": 2535,
"token_acc": 0.8800874078120732,
"train_speed(iter/s)": 0.029516
},
{
"epoch": 0.812767489300428,
"grad_norm": 0.058350813115584405,
"learning_rate": 1.7894958826600884e-06,
"loss": 0.4489152908325195,
"memory(GiB)": 77.63,
"step": 2540,
"token_acc": 0.8784655623365301,
"train_speed(iter/s)": 0.02952
},
{
"epoch": 0.8143674253029879,
"grad_norm": 0.05243037926696882,
"learning_rate": 1.7599638009344566e-06,
"loss": 0.4506648063659668,
"memory(GiB)": 77.63,
"step": 2545,
"token_acc": 0.8711162255466053,
"train_speed(iter/s)": 0.029531
},
{
"epoch": 0.8159673613055478,
"grad_norm": 0.055874525296942985,
"learning_rate": 1.730653905438714e-06,
"loss": 0.451121187210083,
"memory(GiB)": 77.63,
"step": 2550,
"token_acc": 0.875531914893617,
"train_speed(iter/s)": 0.029527
},
{
"epoch": 0.8175672973081076,
"grad_norm": 0.052382476873803714,
"learning_rate": 1.701566986491614e-06,
"loss": 0.43659415245056155,
"memory(GiB)": 77.63,
"step": 2555,
"token_acc": 0.8824301518844928,
"train_speed(iter/s)": 0.029539
},
{
"epoch": 0.8191672333106675,
"grad_norm": 0.05679998544269888,
"learning_rate": 1.672703828399529e-06,
"loss": 0.44143290519714357,
"memory(GiB)": 77.63,
"step": 2560,
"token_acc": 0.9194786645241921,
"train_speed(iter/s)": 0.029538
},
{
"epoch": 0.8207671693132275,
"grad_norm": 0.05177151194489443,
"learning_rate": 1.6440652094352838e-06,
"loss": 0.44036478996276857,
"memory(GiB)": 77.63,
"step": 2565,
"token_acc": 0.86801315171442,
"train_speed(iter/s)": 0.029534
},
{
"epoch": 0.8223671053157874,
"grad_norm": 0.047985617439266506,
"learning_rate": 1.6156519018171856e-06,
"loss": 0.44090909957885743,
"memory(GiB)": 77.63,
"step": 2570,
"token_acc": 0.8926761055759482,
"train_speed(iter/s)": 0.029546
},
{
"epoch": 0.8239670413183473,
"grad_norm": 0.06022869670658613,
"learning_rate": 1.587464671688187e-06,
"loss": 0.4480876922607422,
"memory(GiB)": 77.63,
"step": 2575,
"token_acc": 0.8749468913751594,
"train_speed(iter/s)": 0.029542
},
{
"epoch": 0.8255669773209071,
"grad_norm": 0.05591525813204934,
"learning_rate": 1.5595042790952442e-06,
"loss": 0.4516183853149414,
"memory(GiB)": 77.63,
"step": 2580,
"token_acc": 0.8408594319009468,
"train_speed(iter/s)": 0.029539
},
{
"epoch": 0.827166913323467,
"grad_norm": 0.05315795994218538,
"learning_rate": 1.5317714779688076e-06,
"loss": 0.44116387367248533,
"memory(GiB)": 77.63,
"step": 2585,
"token_acc": 0.8697747394374089,
"train_speed(iter/s)": 0.029549
},
{
"epoch": 0.828766849326027,
"grad_norm": 0.054322305095737905,
"learning_rate": 1.5042670161024975e-06,
"loss": 0.4457075119018555,
"memory(GiB)": 77.63,
"step": 2590,
"token_acc": 0.8946940985381701,
"train_speed(iter/s)": 0.029543
},
{
"epoch": 0.8303667853285869,
"grad_norm": 0.05625328099370699,
"learning_rate": 1.4769916351329495e-06,
"loss": 0.4413478851318359,
"memory(GiB)": 77.63,
"step": 2595,
"token_acc": 0.8992218637312583,
"train_speed(iter/s)": 0.029548
},
{
"epoch": 0.8319667213311468,
"grad_norm": 0.055362561635933255,
"learning_rate": 1.4499460705198e-06,
"loss": 0.4511932373046875,
"memory(GiB)": 77.63,
"step": 2600,
"token_acc": 0.8438552188552189,
"train_speed(iter/s)": 0.029552
},
{
"epoch": 0.8335666573337066,
"grad_norm": 0.052438473367666195,
"learning_rate": 1.4231310515258745e-06,
"loss": 0.441973352432251,
"memory(GiB)": 77.63,
"step": 2605,
"token_acc": 0.8753952017853822,
"train_speed(iter/s)": 0.029546
},
{
"epoch": 0.8351665933362665,
"grad_norm": 0.05274974730016364,
"learning_rate": 1.396547301197504e-06,
"loss": 0.4393311977386475,
"memory(GiB)": 77.63,
"step": 2610,
"token_acc": 0.8518848700967906,
"train_speed(iter/s)": 0.029557
},
{
"epoch": 0.8367665293388264,
"grad_norm": 0.05853294004614818,
"learning_rate": 1.3701955363450447e-06,
"loss": 0.4380232810974121,
"memory(GiB)": 77.63,
"step": 2615,
"token_acc": 0.8570597362296354,
"train_speed(iter/s)": 0.029554
},
{
"epoch": 0.8383664653413864,
"grad_norm": 0.05410978790127522,
"learning_rate": 1.3440764675235384e-06,
"loss": 0.4373164653778076,
"memory(GiB)": 77.63,
"step": 2620,
"token_acc": 0.8798353909465021,
"train_speed(iter/s)": 0.029552
},
{
"epoch": 0.8399664013439463,
"grad_norm": 0.048967955063799855,
"learning_rate": 1.3181907990135624e-06,
"loss": 0.4333020210266113,
"memory(GiB)": 77.63,
"step": 2625,
"token_acc": 0.8836341008089608,
"train_speed(iter/s)": 0.029564
},
{
"epoch": 0.8415663373465061,
"grad_norm": 0.048274089580157754,
"learning_rate": 1.2925392288022299e-06,
"loss": 0.4414947509765625,
"memory(GiB)": 77.63,
"step": 2630,
"token_acc": 0.8760546404178385,
"train_speed(iter/s)": 0.02956
},
{
"epoch": 0.843166273349066,
"grad_norm": 0.053684823491607934,
"learning_rate": 1.267122448564374e-06,
"loss": 0.44922800064086915,
"memory(GiB)": 77.63,
"step": 2635,
"token_acc": 0.8554064052425748,
"train_speed(iter/s)": 0.029558
},
{
"epoch": 0.8447662093516259,
"grad_norm": 0.05262528572569429,
"learning_rate": 1.2419411436439021e-06,
"loss": 0.4328805923461914,
"memory(GiB)": 77.63,
"step": 2640,
"token_acc": 0.8400081317340923,
"train_speed(iter/s)": 0.029565
},
{
"epoch": 0.8463661453541859,
"grad_norm": 0.05549273276976771,
"learning_rate": 1.2169959930353049e-06,
"loss": 0.4460554599761963,
"memory(GiB)": 77.63,
"step": 2645,
"token_acc": 0.8804424157303371,
"train_speed(iter/s)": 0.029559
},
{
"epoch": 0.8479660813567458,
"grad_norm": 0.05180537064402683,
"learning_rate": 1.1922876693653584e-06,
"loss": 0.4503427505493164,
"memory(GiB)": 77.63,
"step": 2650,
"token_acc": 0.8934362934362934,
"train_speed(iter/s)": 0.029564
},
{
"epoch": 0.8495660173593056,
"grad_norm": 0.05011013702559624,
"learning_rate": 1.1678168388749788e-06,
"loss": 0.4415099620819092,
"memory(GiB)": 77.63,
"step": 2655,
"token_acc": 0.8995949690897463,
"train_speed(iter/s)": 0.029566
},
{
"epoch": 0.8511659533618655,
"grad_norm": 0.057170743989864214,
"learning_rate": 1.1435841614012666e-06,
"loss": 0.44884433746337893,
"memory(GiB)": 77.63,
"step": 2660,
"token_acc": 0.854816112084063,
"train_speed(iter/s)": 0.02956
},
{
"epoch": 0.8527658893644254,
"grad_norm": 0.054388272142607975,
"learning_rate": 1.1195902903597023e-06,
"loss": 0.439667797088623,
"memory(GiB)": 77.63,
"step": 2665,
"token_acc": 0.8846260387811634,
"train_speed(iter/s)": 0.02957
},
{
"epoch": 0.8543658253669854,
"grad_norm": 0.05039207218424233,
"learning_rate": 1.0958358727265438e-06,
"loss": 0.4384475231170654,
"memory(GiB)": 77.63,
"step": 2670,
"token_acc": 0.8525793222533995,
"train_speed(iter/s)": 0.029565
},
{
"epoch": 0.8559657613695452,
"grad_norm": 0.0543645938545219,
"learning_rate": 1.0723215490213635e-06,
"loss": 0.4338691711425781,
"memory(GiB)": 77.63,
"step": 2675,
"token_acc": 0.853824495541999,
"train_speed(iter/s)": 0.029558
},
{
"epoch": 0.8575656973721051,
"grad_norm": 0.0579168704227633,
"learning_rate": 1.0490479532897946e-06,
"loss": 0.458463716506958,
"memory(GiB)": 77.63,
"step": 2680,
"token_acc": 0.867092866756393,
"train_speed(iter/s)": 0.029566
},
{
"epoch": 0.859165633374665,
"grad_norm": 0.04996147776053655,
"learning_rate": 1.0260157130864178e-06,
"loss": 0.43754091262817385,
"memory(GiB)": 77.63,
"step": 2685,
"token_acc": 0.8611873713109128,
"train_speed(iter/s)": 0.029558
},
{
"epoch": 0.8607655693772249,
"grad_norm": 0.053537353272037014,
"learning_rate": 1.0032254494578519e-06,
"loss": 0.44204487800598147,
"memory(GiB)": 77.63,
"step": 2690,
"token_acc": 0.8575780654988576,
"train_speed(iter/s)": 0.029558
},
{
"epoch": 0.8623655053797848,
"grad_norm": 0.055119127073411836,
"learning_rate": 9.806777769260034e-07,
"loss": 0.4500781536102295,
"memory(GiB)": 77.63,
"step": 2695,
"token_acc": 0.8872294372294373,
"train_speed(iter/s)": 0.029562
},
{
"epoch": 0.8639654413823447,
"grad_norm": 0.0558713442289911,
"learning_rate": 9.583733034714982e-07,
"loss": 0.43947248458862304,
"memory(GiB)": 77.63,
"step": 2700,
"token_acc": 0.8926744522729466,
"train_speed(iter/s)": 0.029555
},
{
"epoch": 0.8655653773849046,
"grad_norm": 0.0552196144062876,
"learning_rate": 9.363126305172831e-07,
"loss": 0.4443229675292969,
"memory(GiB)": 77.63,
"step": 2705,
"token_acc": 0.9038031319910514,
"train_speed(iter/s)": 0.029561
},
{
"epoch": 0.8671653133874645,
"grad_norm": 0.055723745057826034,
"learning_rate": 9.144963529124163e-07,
"loss": 0.42942004203796386,
"memory(GiB)": 77.63,
"step": 2710,
"token_acc": 0.8680161943319838,
"train_speed(iter/s)": 0.029557
},
{
"epoch": 0.8687652493900244,
"grad_norm": 0.0589746686821641,
"learning_rate": 8.929250589160166e-07,
"loss": 0.4397599220275879,
"memory(GiB)": 77.63,
"step": 2715,
"token_acc": 0.8713878713878714,
"train_speed(iter/s)": 0.029552
},
{
"epoch": 0.8703651853925843,
"grad_norm": 0.04909314017257213,
"learning_rate": 8.715993301814174e-07,
"loss": 0.44155421257019045,
"memory(GiB)": 77.63,
"step": 2720,
"token_acc": 0.8710053650571495,
"train_speed(iter/s)": 0.029561
},
{
"epoch": 0.8719651213951441,
"grad_norm": 0.05047518453544575,
"learning_rate": 8.505197417404687e-07,
"loss": 0.43677616119384766,
"memory(GiB)": 77.63,
"step": 2725,
"token_acc": 0.8809886575249704,
"train_speed(iter/s)": 0.029556
},
{
"epoch": 0.8735650573977041,
"grad_norm": 0.05102151204327215,
"learning_rate": 8.296868619880372e-07,
"loss": 0.44188566207885743,
"memory(GiB)": 77.63,
"step": 2730,
"token_acc": 0.8547172833573602,
"train_speed(iter/s)": 0.029553
},
{
"epoch": 0.875164993400264,
"grad_norm": 0.04729834705444575,
"learning_rate": 8.091012526666797e-07,
"loss": 0.4441237926483154,
"memory(GiB)": 77.63,
"step": 2735,
"token_acc": 0.8537975972307066,
"train_speed(iter/s)": 0.029561
},
{
"epoch": 0.8767649294028239,
"grad_norm": 0.047668539210598965,
"learning_rate": 7.887634688515e-07,
"loss": 0.4462736129760742,
"memory(GiB)": 77.63,
"step": 2740,
"token_acc": 0.903437815975733,
"train_speed(iter/s)": 0.029554
},
{
"epoch": 0.8783648654053838,
"grad_norm": 0.052216823887528664,
"learning_rate": 7.686740589351704e-07,
"loss": 0.44857120513916016,
"memory(GiB)": 77.63,
"step": 2745,
"token_acc": 0.8033573141486811,
"train_speed(iter/s)": 0.029556
},
{
"epoch": 0.8799648014079436,
"grad_norm": 0.055862979558343906,
"learning_rate": 7.488335646131628e-07,
"loss": 0.44959425926208496,
"memory(GiB)": 77.63,
"step": 2750,
"token_acc": 0.8605054151624548,
"train_speed(iter/s)": 0.029558
},
{
"epoch": 0.8799648014079436,
"eval_loss": 0.6577034592628479,
"eval_runtime": 106.4875,
"eval_samples_per_second": 188.642,
"eval_steps_per_second": 0.948,
"eval_token_acc": 0.8722958612553617,
"step": 2750
},
{
"epoch": 0.8815647374105036,
"grad_norm": 0.051528350081992934,
"learning_rate": 7.292425208691212e-07,
"loss": 0.43878631591796874,
"memory(GiB)": 77.63,
"step": 2755,
"token_acc": 0.8812832745626772,
"train_speed(iter/s)": 0.029532
},
{
"epoch": 0.8831646734130635,
"grad_norm": 0.05310175482611414,
"learning_rate": 7.099014559604556e-07,
"loss": 0.45635418891906737,
"memory(GiB)": 77.63,
"step": 2760,
"token_acc": 0.8894999360532038,
"train_speed(iter/s)": 0.029537
},
{
"epoch": 0.8847646094156234,
"grad_norm": 0.04975738369955541,
"learning_rate": 6.908108914040823e-07,
"loss": 0.4421397686004639,
"memory(GiB)": 77.63,
"step": 2765,
"token_acc": 0.9070493575117089,
"train_speed(iter/s)": 0.029548
},
{
"epoch": 0.8863645454181833,
"grad_norm": 0.053564472486824076,
"learning_rate": 6.71971341962373e-07,
"loss": 0.4513510227203369,
"memory(GiB)": 77.63,
"step": 2770,
"token_acc": 0.8660589060308556,
"train_speed(iter/s)": 0.02955
},
{
"epoch": 0.8879644814207431,
"grad_norm": 0.06332925713320489,
"learning_rate": 6.53383315629268e-07,
"loss": 0.4404273509979248,
"memory(GiB)": 77.63,
"step": 2775,
"token_acc": 0.8507806501151779,
"train_speed(iter/s)": 0.029546
},
{
"epoch": 0.889564417423303,
"grad_norm": 0.063294227744794,
"learning_rate": 6.350473136165836e-07,
"loss": 0.4379493236541748,
"memory(GiB)": 77.63,
"step": 2780,
"token_acc": 0.8879898461050294,
"train_speed(iter/s)": 0.029561
},
{
"epoch": 0.891164353425863,
"grad_norm": 0.05151642870451994,
"learning_rate": 6.169638303404912e-07,
"loss": 0.4380655765533447,
"memory(GiB)": 77.63,
"step": 2785,
"token_acc": 0.8904059040590406,
"train_speed(iter/s)": 0.02956
},
{
"epoch": 0.8927642894284229,
"grad_norm": 0.05406168921762394,
"learning_rate": 5.991333534081878e-07,
"loss": 0.4479250907897949,
"memory(GiB)": 77.63,
"step": 2790,
"token_acc": 0.8831118813787792,
"train_speed(iter/s)": 0.02956
},
{
"epoch": 0.8943642254309827,
"grad_norm": 0.054911478200183335,
"learning_rate": 5.815563636047539e-07,
"loss": 0.43634886741638185,
"memory(GiB)": 77.63,
"step": 2795,
"token_acc": 0.8714865708931917,
"train_speed(iter/s)": 0.02957
},
{
"epoch": 0.8959641614335426,
"grad_norm": 0.05257509941727236,
"learning_rate": 5.64233334880181e-07,
"loss": 0.44048466682434084,
"memory(GiB)": 77.63,
"step": 2800,
"token_acc": 0.891296869625043,
"train_speed(iter/s)": 0.029568
},
{
"epoch": 0.8975640974361025,
"grad_norm": 0.05532206374459385,
"learning_rate": 5.471647343365982e-07,
"loss": 0.44726853370666503,
"memory(GiB)": 77.63,
"step": 2805,
"token_acc": 0.8828892005610098,
"train_speed(iter/s)": 0.029576
},
{
"epoch": 0.8991640334386625,
"grad_norm": 0.05107208389411162,
"learning_rate": 5.303510222156716e-07,
"loss": 0.4470540523529053,
"memory(GiB)": 77.63,
"step": 2810,
"token_acc": 0.8489765812281025,
"train_speed(iter/s)": 0.029574
},
{
"epoch": 0.9007639694412224,
"grad_norm": 0.055452171728558604,
"learning_rate": 5.137926518862013e-07,
"loss": 0.4417248249053955,
"memory(GiB)": 77.63,
"step": 2815,
"token_acc": 0.8739084132055378,
"train_speed(iter/s)": 0.029569
},
{
"epoch": 0.9023639054437822,
"grad_norm": 0.04866974543895629,
"learning_rate": 4.974900698318885e-07,
"loss": 0.4414045810699463,
"memory(GiB)": 77.63,
"step": 2820,
"token_acc": 0.8659420289855072,
"train_speed(iter/s)": 0.029581
},
{
"epoch": 0.9039638414463421,
"grad_norm": 0.05275403785935388,
"learning_rate": 4.814437156393048e-07,
"loss": 0.4543337821960449,
"memory(GiB)": 77.63,
"step": 2825,
"token_acc": 0.8396122896854425,
"train_speed(iter/s)": 0.029574
},
{
"epoch": 0.905563777448902,
"grad_norm": 0.05372217346495556,
"learning_rate": 4.656540219860317e-07,
"loss": 0.45271754264831543,
"memory(GiB)": 77.63,
"step": 2830,
"token_acc": 0.8707617789520036,
"train_speed(iter/s)": 0.029571
},
{
"epoch": 0.907163713451462,
"grad_norm": 0.05715427837146615,
"learning_rate": 4.501214146289956e-07,
"loss": 0.4418344497680664,
"memory(GiB)": 77.63,
"step": 2835,
"token_acc": 0.8788416882939489,
"train_speed(iter/s)": 0.02958
},
{
"epoch": 0.9087636494540219,
"grad_norm": 0.04986985198768239,
"learning_rate": 4.3484631239299356e-07,
"loss": 0.4437891960144043,
"memory(GiB)": 77.63,
"step": 2840,
"token_acc": 0.8431597023468803,
"train_speed(iter/s)": 0.029572
},
{
"epoch": 0.9103635854565817,
"grad_norm": 0.04999552116510165,
"learning_rate": 4.198291271593924e-07,
"loss": 0.44283204078674315,
"memory(GiB)": 77.63,
"step": 2845,
"token_acc": 0.8843727072633896,
"train_speed(iter/s)": 0.029575
},
{
"epoch": 0.9119635214591416,
"grad_norm": 0.047844581858855956,
"learning_rate": 4.0507026385502747e-07,
"loss": 0.4449836254119873,
"memory(GiB)": 77.63,
"step": 2850,
"token_acc": 0.9034812490661885,
"train_speed(iter/s)": 0.029576
},
{
"epoch": 0.9135634574617015,
"grad_norm": 0.05675006848739315,
"learning_rate": 3.9057012044127817e-07,
"loss": 0.44204154014587405,
"memory(GiB)": 77.63,
"step": 2855,
"token_acc": 0.8655569782330346,
"train_speed(iter/s)": 0.029569
},
{
"epoch": 0.9151633934642615,
"grad_norm": 0.054077703664588216,
"learning_rate": 3.7632908790334656e-07,
"loss": 0.4383398532867432,
"memory(GiB)": 77.63,
"step": 2860,
"token_acc": 0.8899396929824561,
"train_speed(iter/s)": 0.029578
},
{
"epoch": 0.9167633294668214,
"grad_norm": 0.055142279260510525,
"learning_rate": 3.6234755023970447e-07,
"loss": 0.4388674259185791,
"memory(GiB)": 77.63,
"step": 2865,
"token_acc": 0.8406133828996283,
"train_speed(iter/s)": 0.029574
},
{
"epoch": 0.9183632654693812,
"grad_norm": 0.05242909357272202,
"learning_rate": 3.4862588445174985e-07,
"loss": 0.44350008964538573,
"memory(GiB)": 77.63,
"step": 2870,
"token_acc": 0.8773854961832062,
"train_speed(iter/s)": 0.029568
},
{
"epoch": 0.9199632014719411,
"grad_norm": 0.051431062939426,
"learning_rate": 3.3516446053363015e-07,
"loss": 0.43948516845703123,
"memory(GiB)": 77.63,
"step": 2875,
"token_acc": 0.8812238692512353,
"train_speed(iter/s)": 0.029578
},
{
"epoch": 0.921563137474501,
"grad_norm": 0.04842108410960974,
"learning_rate": 3.219636414622751e-07,
"loss": 0.44395694732666013,
"memory(GiB)": 77.63,
"step": 2880,
"token_acc": 0.872349158571624,
"train_speed(iter/s)": 0.029573
},
{
"epoch": 0.923163073477061,
"grad_norm": 0.050854557927068264,
"learning_rate": 3.090237831876053e-07,
"loss": 0.4437469482421875,
"memory(GiB)": 77.63,
"step": 2885,
"token_acc": 0.8500566251415629,
"train_speed(iter/s)": 0.029572
},
{
"epoch": 0.9247630094796209,
"grad_norm": 0.05289158935619814,
"learning_rate": 2.9634523462293005e-07,
"loss": 0.439394474029541,
"memory(GiB)": 77.63,
"step": 2890,
"token_acc": 0.9039268013724743,
"train_speed(iter/s)": 0.029576
},
{
"epoch": 0.9263629454821807,
"grad_norm": 0.0532633695315608,
"learning_rate": 2.839283376355506e-07,
"loss": 0.4414195537567139,
"memory(GiB)": 77.63,
"step": 2895,
"token_acc": 0.8979206049149339,
"train_speed(iter/s)": 0.029567
},
{
"epoch": 0.9279628814847406,
"grad_norm": 0.05197494947423765,
"learning_rate": 2.717734270375272e-07,
"loss": 0.4303572177886963,
"memory(GiB)": 77.63,
"step": 2900,
"token_acc": 0.8932318992654774,
"train_speed(iter/s)": 0.029573
},
{
"epoch": 0.9295628174873005,
"grad_norm": 0.05715199054523632,
"learning_rate": 2.5988083057666534e-07,
"loss": 0.4488718032836914,
"memory(GiB)": 77.63,
"step": 2905,
"token_acc": 0.8734599589322382,
"train_speed(iter/s)": 0.029572
},
{
"epoch": 0.9311627534898604,
"grad_norm": 0.052180972248620894,
"learning_rate": 2.4825086892766745e-07,
"loss": 0.44499683380126953,
"memory(GiB)": 77.63,
"step": 2910,
"token_acc": 0.8798773215198501,
"train_speed(iter/s)": 0.029568
},
{
"epoch": 0.9327626894924202,
"grad_norm": 0.05195036988843101,
"learning_rate": 2.3688385568349515e-07,
"loss": 0.4348268508911133,
"memory(GiB)": 77.63,
"step": 2915,
"token_acc": 0.8380835380835381,
"train_speed(iter/s)": 0.029578
},
{
"epoch": 0.9343626254949802,
"grad_norm": 0.05340475761749026,
"learning_rate": 2.2578009734690264e-07,
"loss": 0.4533662796020508,
"memory(GiB)": 77.63,
"step": 2920,
"token_acc": 0.8854700854700854,
"train_speed(iter/s)": 0.029569
},
{
"epoch": 0.9359625614975401,
"grad_norm": 0.047434143218971526,
"learning_rate": 2.1493989332218468e-07,
"loss": 0.4382183074951172,
"memory(GiB)": 77.63,
"step": 2925,
"token_acc": 0.880750496121234,
"train_speed(iter/s)": 0.029565
},
{
"epoch": 0.9375624975001,
"grad_norm": 0.052509681961836426,
"learning_rate": 2.043635359070928e-07,
"loss": 0.44708704948425293,
"memory(GiB)": 77.63,
"step": 2930,
"token_acc": 0.8942012598933936,
"train_speed(iter/s)": 0.029576
},
{
"epoch": 0.9391624335026599,
"grad_norm": 0.0531539858574761,
"learning_rate": 1.9405131028495838e-07,
"loss": 0.45058341026306153,
"memory(GiB)": 77.63,
"step": 2935,
"token_acc": 0.8614694335389792,
"train_speed(iter/s)": 0.029569
},
{
"epoch": 0.9407623695052197,
"grad_norm": 0.05931701393554437,
"learning_rate": 1.8400349451700438e-07,
"loss": 0.44367156028747556,
"memory(GiB)": 77.63,
"step": 2940,
"token_acc": 0.881156184096266,
"train_speed(iter/s)": 0.029573
},
{
"epoch": 0.9423623055077797,
"grad_norm": 0.05380572562718686,
"learning_rate": 1.742203595348435e-07,
"loss": 0.4424111843109131,
"memory(GiB)": 77.63,
"step": 2945,
"token_acc": 0.8769617074701821,
"train_speed(iter/s)": 0.029576
},
{
"epoch": 0.9439622415103396,
"grad_norm": 0.06156259021000458,
"learning_rate": 1.6470216913317628e-07,
"loss": 0.4509577751159668,
"memory(GiB)": 77.63,
"step": 2950,
"token_acc": 0.8544532947139754,
"train_speed(iter/s)": 0.029567
},
{
"epoch": 0.9455621775128995,
"grad_norm": 0.050697077683688974,
"learning_rate": 1.5544917996267562e-07,
"loss": 0.44117283821105957,
"memory(GiB)": 77.63,
"step": 2955,
"token_acc": 0.8515226026101759,
"train_speed(iter/s)": 0.029575
},
{
"epoch": 0.9471621135154594,
"grad_norm": 0.05205296295346333,
"learning_rate": 1.464616415230702e-07,
"loss": 0.4488182067871094,
"memory(GiB)": 77.63,
"step": 2960,
"token_acc": 0.874439461883408,
"train_speed(iter/s)": 0.029569
},
{
"epoch": 0.9487620495180192,
"grad_norm": 0.047632029637579856,
"learning_rate": 1.3773979615640976e-07,
"loss": 0.4415272235870361,
"memory(GiB)": 77.63,
"step": 2965,
"token_acc": 0.8889883616830797,
"train_speed(iter/s)": 0.029564
},
{
"epoch": 0.9503619855205792,
"grad_norm": 0.051226023380966074,
"learning_rate": 1.292838790405393e-07,
"loss": 0.4453396797180176,
"memory(GiB)": 77.63,
"step": 2970,
"token_acc": 0.8701866977829639,
"train_speed(iter/s)": 0.029572
},
{
"epoch": 0.9519619215231391,
"grad_norm": 0.05296626405711913,
"learning_rate": 1.2109411818274851e-07,
"loss": 0.44417614936828614,
"memory(GiB)": 77.63,
"step": 2975,
"token_acc": 0.8997547959036493,
"train_speed(iter/s)": 0.029565
},
{
"epoch": 0.953561857525699,
"grad_norm": 0.0535756814263424,
"learning_rate": 1.1317073441363458e-07,
"loss": 0.444796085357666,
"memory(GiB)": 77.63,
"step": 2980,
"token_acc": 0.9101887677336147,
"train_speed(iter/s)": 0.029563
},
{
"epoch": 0.9551617935282589,
"grad_norm": 0.05005027675979017,
"learning_rate": 1.055139413811379e-07,
"loss": 0.45203323364257814,
"memory(GiB)": 77.63,
"step": 2985,
"token_acc": 0.8569892473118279,
"train_speed(iter/s)": 0.029568
},
{
"epoch": 0.9567617295308187,
"grad_norm": 0.04934807615166247,
"learning_rate": 9.812394554478355e-08,
"loss": 0.43912034034729003,
"memory(GiB)": 77.63,
"step": 2990,
"token_acc": 0.8557346268189642,
"train_speed(iter/s)": 0.029559
},
{
"epoch": 0.9583616655333786,
"grad_norm": 0.051618825055470385,
"learning_rate": 9.10009461701189e-08,
"loss": 0.4506105899810791,
"memory(GiB)": 77.63,
"step": 2995,
"token_acc": 0.7809948032665182,
"train_speed(iter/s)": 0.029565
},
{
"epoch": 0.9599616015359386,
"grad_norm": 0.054833054855342726,
"learning_rate": 8.41451353233369e-08,
"loss": 0.442844820022583,
"memory(GiB)": 77.63,
"step": 3000,
"token_acc": 0.8733862959285005,
"train_speed(iter/s)": 0.029563
},
{
"epoch": 0.9599616015359386,
"eval_loss": 0.6573547720909119,
"eval_runtime": 106.0877,
"eval_samples_per_second": 189.353,
"eval_steps_per_second": 0.952,
"eval_token_acc": 0.8724322608695082,
"step": 3000
}
],
"logging_steps": 5,
"max_steps": 3125,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.944159044825703e+20,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}