| { | |
| "best_metric": 0.65735477, | |
| "best_model_checkpoint": "/cpfs04/shared/rlproject/zhangqi/model_garden/0709_intern2b_v7-1-part15-19-resize-decay/v0-20250710-072707/checkpoint-3000", | |
| "epoch": 0.9599616015359386, | |
| "eval_steps": 250, | |
| "global_step": 3000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0003199872005119795, | |
| "grad_norm": 0.06346331978668461, | |
| "learning_rate": 2.0000000000000002e-07, | |
| "loss": 0.4898327589035034, | |
| "memory(GiB)": 37.39, | |
| "step": 1, | |
| "token_acc": 0.8589147286821706, | |
| "train_speed(iter/s)": 0.017141 | |
| }, | |
| { | |
| "epoch": 0.0015999360025598975, | |
| "grad_norm": 0.06701772816278752, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 0.479672372341156, | |
| "memory(GiB)": 57.22, | |
| "step": 5, | |
| "token_acc": 0.8759901666211418, | |
| "train_speed(iter/s)": 0.029212 | |
| }, | |
| { | |
| "epoch": 0.003199872005119795, | |
| "grad_norm": 0.0577242460197625, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 0.4767899990081787, | |
| "memory(GiB)": 57.22, | |
| "step": 10, | |
| "token_acc": 0.8920853213584058, | |
| "train_speed(iter/s)": 0.030985 | |
| }, | |
| { | |
| "epoch": 0.004799808007679693, | |
| "grad_norm": 0.05378464715409511, | |
| "learning_rate": 3e-06, | |
| "loss": 0.46750531196594236, | |
| "memory(GiB)": 57.22, | |
| "step": 15, | |
| "token_acc": 0.846137292877125, | |
| "train_speed(iter/s)": 0.030693 | |
| }, | |
| { | |
| "epoch": 0.00639974401023959, | |
| "grad_norm": 0.05482863789544427, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.47017059326171873, | |
| "memory(GiB)": 57.22, | |
| "step": 20, | |
| "token_acc": 0.8701638201463925, | |
| "train_speed(iter/s)": 0.031839 | |
| }, | |
| { | |
| "epoch": 0.007999680012799487, | |
| "grad_norm": 0.05575025715512655, | |
| "learning_rate": 5e-06, | |
| "loss": 0.46974716186523435, | |
| "memory(GiB)": 57.22, | |
| "step": 25, | |
| "token_acc": 0.8939419941650936, | |
| "train_speed(iter/s)": 0.031203 | |
| }, | |
| { | |
| "epoch": 0.009599616015359386, | |
| "grad_norm": 0.05007200892621057, | |
| "learning_rate": 6e-06, | |
| "loss": 0.4700496196746826, | |
| "memory(GiB)": 57.22, | |
| "step": 30, | |
| "token_acc": 0.872761844398669, | |
| "train_speed(iter/s)": 0.030738 | |
| }, | |
| { | |
| "epoch": 0.011199552017919284, | |
| "grad_norm": 0.06284973201247342, | |
| "learning_rate": 7e-06, | |
| "loss": 0.477018404006958, | |
| "memory(GiB)": 57.22, | |
| "step": 35, | |
| "token_acc": 0.8738273921200751, | |
| "train_speed(iter/s)": 0.031491 | |
| }, | |
| { | |
| "epoch": 0.01279948802047918, | |
| "grad_norm": 0.051858307829802386, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 0.46584124565124513, | |
| "memory(GiB)": 57.22, | |
| "step": 40, | |
| "token_acc": 0.8791312559017942, | |
| "train_speed(iter/s)": 0.031035 | |
| }, | |
| { | |
| "epoch": 0.014399424023039079, | |
| "grad_norm": 0.05024484287076301, | |
| "learning_rate": 9e-06, | |
| "loss": 0.4685808658599854, | |
| "memory(GiB)": 57.22, | |
| "step": 45, | |
| "token_acc": 0.8846325167037862, | |
| "train_speed(iter/s)": 0.031503 | |
| }, | |
| { | |
| "epoch": 0.015999360025598975, | |
| "grad_norm": 0.06025612278216295, | |
| "learning_rate": 1e-05, | |
| "loss": 0.4697974681854248, | |
| "memory(GiB)": 57.22, | |
| "step": 50, | |
| "token_acc": 0.854153041203401, | |
| "train_speed(iter/s)": 0.031549 | |
| }, | |
| { | |
| "epoch": 0.017599296028158875, | |
| "grad_norm": 0.05171252494611451, | |
| "learning_rate": 1.1000000000000001e-05, | |
| "loss": 0.4622661113739014, | |
| "memory(GiB)": 73.06, | |
| "step": 55, | |
| "token_acc": 0.8281821878812525, | |
| "train_speed(iter/s)": 0.031181 | |
| }, | |
| { | |
| "epoch": 0.01919923203071877, | |
| "grad_norm": 0.06238294030934267, | |
| "learning_rate": 1.2e-05, | |
| "loss": 0.47000856399536134, | |
| "memory(GiB)": 73.06, | |
| "step": 60, | |
| "token_acc": 0.8703465982028241, | |
| "train_speed(iter/s)": 0.031617 | |
| }, | |
| { | |
| "epoch": 0.020799168033278668, | |
| "grad_norm": 0.05576663204628902, | |
| "learning_rate": 1.3000000000000001e-05, | |
| "loss": 0.4765446186065674, | |
| "memory(GiB)": 73.06, | |
| "step": 65, | |
| "token_acc": 0.867019517036057, | |
| "train_speed(iter/s)": 0.031138 | |
| }, | |
| { | |
| "epoch": 0.022399104035838568, | |
| "grad_norm": 0.054406694286476175, | |
| "learning_rate": 1.4e-05, | |
| "loss": 0.47959036827087403, | |
| "memory(GiB)": 73.06, | |
| "step": 70, | |
| "token_acc": 0.8927940657011657, | |
| "train_speed(iter/s)": 0.030807 | |
| }, | |
| { | |
| "epoch": 0.023999040038398464, | |
| "grad_norm": 0.05458413282309297, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 0.45742173194885255, | |
| "memory(GiB)": 73.06, | |
| "step": 75, | |
| "token_acc": 0.8916037316748112, | |
| "train_speed(iter/s)": 0.031204 | |
| }, | |
| { | |
| "epoch": 0.02559897604095836, | |
| "grad_norm": 0.05986601718141533, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 0.46221466064453126, | |
| "memory(GiB)": 73.06, | |
| "step": 80, | |
| "token_acc": 0.8513287970214405, | |
| "train_speed(iter/s)": 0.030907 | |
| }, | |
| { | |
| "epoch": 0.02719891204351826, | |
| "grad_norm": 0.0610597448330617, | |
| "learning_rate": 1.7e-05, | |
| "loss": 0.4636178493499756, | |
| "memory(GiB)": 73.06, | |
| "step": 85, | |
| "token_acc": 0.8755391250770179, | |
| "train_speed(iter/s)": 0.030852 | |
| }, | |
| { | |
| "epoch": 0.028798848046078157, | |
| "grad_norm": 0.0537272421077024, | |
| "learning_rate": 1.8e-05, | |
| "loss": 0.4674212455749512, | |
| "memory(GiB)": 73.06, | |
| "step": 90, | |
| "token_acc": 0.8330635370295427, | |
| "train_speed(iter/s)": 0.03096 | |
| }, | |
| { | |
| "epoch": 0.030398784048638054, | |
| "grad_norm": 0.05874793146275866, | |
| "learning_rate": 1.9e-05, | |
| "loss": 0.459043550491333, | |
| "memory(GiB)": 73.06, | |
| "step": 95, | |
| "token_acc": 0.8686557483379277, | |
| "train_speed(iter/s)": 0.030744 | |
| }, | |
| { | |
| "epoch": 0.03199872005119795, | |
| "grad_norm": 0.0635697375742028, | |
| "learning_rate": 2e-05, | |
| "loss": 0.4696988582611084, | |
| "memory(GiB)": 73.06, | |
| "step": 100, | |
| "token_acc": 0.8658624414836155, | |
| "train_speed(iter/s)": 0.030967 | |
| }, | |
| { | |
| "epoch": 0.03359865605375785, | |
| "grad_norm": 0.05212754941929234, | |
| "learning_rate": 1.9999865178850847e-05, | |
| "loss": 0.47245235443115235, | |
| "memory(GiB)": 73.06, | |
| "step": 105, | |
| "token_acc": 0.8773064185459707, | |
| "train_speed(iter/s)": 0.030857 | |
| }, | |
| { | |
| "epoch": 0.03519859205631775, | |
| "grad_norm": 0.0584305183084298, | |
| "learning_rate": 1.999946071903873e-05, | |
| "loss": 0.4617309093475342, | |
| "memory(GiB)": 73.06, | |
| "step": 110, | |
| "token_acc": 0.8634434872058634, | |
| "train_speed(iter/s)": 0.030617 | |
| }, | |
| { | |
| "epoch": 0.03679852805887764, | |
| "grad_norm": 0.0594956422290189, | |
| "learning_rate": 1.9998786631469602e-05, | |
| "loss": 0.4737922191619873, | |
| "memory(GiB)": 73.06, | |
| "step": 115, | |
| "token_acc": 0.8586429725363489, | |
| "train_speed(iter/s)": 0.030917 | |
| }, | |
| { | |
| "epoch": 0.03839846406143754, | |
| "grad_norm": 0.05542732612540538, | |
| "learning_rate": 1.999784293431971e-05, | |
| "loss": 0.46903514862060547, | |
| "memory(GiB)": 73.06, | |
| "step": 120, | |
| "token_acc": 0.8490970309152127, | |
| "train_speed(iter/s)": 0.030707 | |
| }, | |
| { | |
| "epoch": 0.03999840006399744, | |
| "grad_norm": 0.06279644130355853, | |
| "learning_rate": 1.9996629653035128e-05, | |
| "loss": 0.47660508155822756, | |
| "memory(GiB)": 73.06, | |
| "step": 125, | |
| "token_acc": 0.8797399783315276, | |
| "train_speed(iter/s)": 0.03059 | |
| }, | |
| { | |
| "epoch": 0.041598336066557336, | |
| "grad_norm": 0.0589281224878424, | |
| "learning_rate": 1.999514682033104e-05, | |
| "loss": 0.47182955741882326, | |
| "memory(GiB)": 73.06, | |
| "step": 130, | |
| "token_acc": 0.8641215106732348, | |
| "train_speed(iter/s)": 0.030792 | |
| }, | |
| { | |
| "epoch": 0.043198272069117236, | |
| "grad_norm": 0.06201767062726238, | |
| "learning_rate": 1.99933944761909e-05, | |
| "loss": 0.4760914325714111, | |
| "memory(GiB)": 73.06, | |
| "step": 135, | |
| "token_acc": 0.856929955290611, | |
| "train_speed(iter/s)": 0.030572 | |
| }, | |
| { | |
| "epoch": 0.044798208071677136, | |
| "grad_norm": 0.06842512509591067, | |
| "learning_rate": 1.999137266786531e-05, | |
| "loss": 0.4673017501831055, | |
| "memory(GiB)": 73.06, | |
| "step": 140, | |
| "token_acc": 0.8725868725868726, | |
| "train_speed(iter/s)": 0.030583 | |
| }, | |
| { | |
| "epoch": 0.04639814407423703, | |
| "grad_norm": 0.06206860242026748, | |
| "learning_rate": 1.998908144987078e-05, | |
| "loss": 0.4650784969329834, | |
| "memory(GiB)": 73.06, | |
| "step": 145, | |
| "token_acc": 0.8592991206195039, | |
| "train_speed(iter/s)": 0.030589 | |
| }, | |
| { | |
| "epoch": 0.04799808007679693, | |
| "grad_norm": 0.07443005443822684, | |
| "learning_rate": 1.9986520883988233e-05, | |
| "loss": 0.46671695709228517, | |
| "memory(GiB)": 73.06, | |
| "step": 150, | |
| "token_acc": 0.8422288642186165, | |
| "train_speed(iter/s)": 0.030331 | |
| }, | |
| { | |
| "epoch": 0.04959801607935683, | |
| "grad_norm": 0.06105984389773331, | |
| "learning_rate": 1.9983691039261358e-05, | |
| "loss": 0.47356271743774414, | |
| "memory(GiB)": 73.06, | |
| "step": 155, | |
| "token_acc": 0.8814565604591332, | |
| "train_speed(iter/s)": 0.030476 | |
| }, | |
| { | |
| "epoch": 0.05119795208191672, | |
| "grad_norm": 0.08001789963695773, | |
| "learning_rate": 1.998059199199474e-05, | |
| "loss": 0.4710524559020996, | |
| "memory(GiB)": 73.06, | |
| "step": 160, | |
| "token_acc": 0.8710629921259843, | |
| "train_speed(iter/s)": 0.03042 | |
| }, | |
| { | |
| "epoch": 0.05279788808447662, | |
| "grad_norm": 0.05874066660842649, | |
| "learning_rate": 1.9977223825751802e-05, | |
| "loss": 0.46933708190917967, | |
| "memory(GiB)": 73.06, | |
| "step": 165, | |
| "token_acc": 0.894688221709007, | |
| "train_speed(iter/s)": 0.030265 | |
| }, | |
| { | |
| "epoch": 0.05439782408703652, | |
| "grad_norm": 0.06544476253513323, | |
| "learning_rate": 1.997358663135255e-05, | |
| "loss": 0.46097607612609864, | |
| "memory(GiB)": 73.06, | |
| "step": 170, | |
| "token_acc": 0.8832285384319261, | |
| "train_speed(iter/s)": 0.030391 | |
| }, | |
| { | |
| "epoch": 0.055997760089596414, | |
| "grad_norm": 0.052612734205809274, | |
| "learning_rate": 1.9969680506871138e-05, | |
| "loss": 0.4674376010894775, | |
| "memory(GiB)": 73.06, | |
| "step": 175, | |
| "token_acc": 0.8822400558269365, | |
| "train_speed(iter/s)": 0.030298 | |
| }, | |
| { | |
| "epoch": 0.057597696092156314, | |
| "grad_norm": 0.05831524002340728, | |
| "learning_rate": 1.9965505557633188e-05, | |
| "loss": 0.47021942138671874, | |
| "memory(GiB)": 73.06, | |
| "step": 180, | |
| "token_acc": 0.8729036501150937, | |
| "train_speed(iter/s)": 0.030222 | |
| }, | |
| { | |
| "epoch": 0.059197632094716214, | |
| "grad_norm": 0.05558463298745032, | |
| "learning_rate": 1.9961061896213006e-05, | |
| "loss": 0.4707474708557129, | |
| "memory(GiB)": 73.06, | |
| "step": 185, | |
| "token_acc": 0.8507278220269157, | |
| "train_speed(iter/s)": 0.030322 | |
| }, | |
| { | |
| "epoch": 0.06079756809727611, | |
| "grad_norm": 0.05835065786169716, | |
| "learning_rate": 1.9956349642430494e-05, | |
| "loss": 0.4792951583862305, | |
| "memory(GiB)": 73.06, | |
| "step": 190, | |
| "token_acc": 0.8596896665566194, | |
| "train_speed(iter/s)": 0.030221 | |
| }, | |
| { | |
| "epoch": 0.06239750409983601, | |
| "grad_norm": 0.055138234334700054, | |
| "learning_rate": 1.9951368923347945e-05, | |
| "loss": 0.4755210876464844, | |
| "memory(GiB)": 73.06, | |
| "step": 195, | |
| "token_acc": 0.8501170960187353, | |
| "train_speed(iter/s)": 0.030259 | |
| }, | |
| { | |
| "epoch": 0.0639974401023959, | |
| "grad_norm": 0.06763946666899583, | |
| "learning_rate": 1.9946119873266615e-05, | |
| "loss": 0.4560092926025391, | |
| "memory(GiB)": 73.06, | |
| "step": 200, | |
| "token_acc": 0.8891928864569083, | |
| "train_speed(iter/s)": 0.030234 | |
| }, | |
| { | |
| "epoch": 0.0655973761049558, | |
| "grad_norm": 0.056758126149685124, | |
| "learning_rate": 1.9940602633723097e-05, | |
| "loss": 0.470977258682251, | |
| "memory(GiB)": 73.06, | |
| "step": 205, | |
| "token_acc": 0.8635224424698109, | |
| "train_speed(iter/s)": 0.030086 | |
| }, | |
| { | |
| "epoch": 0.0671973121075157, | |
| "grad_norm": 0.06089496841996932, | |
| "learning_rate": 1.99348173534855e-05, | |
| "loss": 0.4699739456176758, | |
| "memory(GiB)": 73.06, | |
| "step": 210, | |
| "token_acc": 0.8621679827709978, | |
| "train_speed(iter/s)": 0.03019 | |
| }, | |
| { | |
| "epoch": 0.06879724811007559, | |
| "grad_norm": 0.05802383504947049, | |
| "learning_rate": 1.9928764188549462e-05, | |
| "loss": 0.46386079788208007, | |
| "memory(GiB)": 73.06, | |
| "step": 215, | |
| "token_acc": 0.8754250939681403, | |
| "train_speed(iter/s)": 0.030114 | |
| }, | |
| { | |
| "epoch": 0.0703971841126355, | |
| "grad_norm": 0.05515703750577632, | |
| "learning_rate": 1.9922443302133906e-05, | |
| "loss": 0.4679898262023926, | |
| "memory(GiB)": 73.06, | |
| "step": 220, | |
| "token_acc": 0.88738807102747, | |
| "train_speed(iter/s)": 0.030007 | |
| }, | |
| { | |
| "epoch": 0.07199712011519539, | |
| "grad_norm": 0.05967474480415036, | |
| "learning_rate": 1.9915854864676665e-05, | |
| "loss": 0.47310919761657716, | |
| "memory(GiB)": 73.06, | |
| "step": 225, | |
| "token_acc": 0.8383795309168444, | |
| "train_speed(iter/s)": 0.030071 | |
| }, | |
| { | |
| "epoch": 0.07359705611775529, | |
| "grad_norm": 0.06076387420670948, | |
| "learning_rate": 1.990899905382988e-05, | |
| "loss": 0.4678232192993164, | |
| "memory(GiB)": 73.06, | |
| "step": 230, | |
| "token_acc": 0.8767123287671232, | |
| "train_speed(iter/s)": 0.029948 | |
| }, | |
| { | |
| "epoch": 0.07519699212031519, | |
| "grad_norm": 0.05635803811030448, | |
| "learning_rate": 1.9901876054455217e-05, | |
| "loss": 0.4821170330047607, | |
| "memory(GiB)": 73.06, | |
| "step": 235, | |
| "token_acc": 0.8841222879684418, | |
| "train_speed(iter/s)": 0.029913 | |
| }, | |
| { | |
| "epoch": 0.07679692812287509, | |
| "grad_norm": 0.053886114557468945, | |
| "learning_rate": 1.9894486058618863e-05, | |
| "loss": 0.46213107109069823, | |
| "memory(GiB)": 73.06, | |
| "step": 240, | |
| "token_acc": 0.8886558627264061, | |
| "train_speed(iter/s)": 0.02992 | |
| }, | |
| { | |
| "epoch": 0.07839686412543498, | |
| "grad_norm": 0.06048992108753748, | |
| "learning_rate": 1.9886829265586368e-05, | |
| "loss": 0.4749046802520752, | |
| "memory(GiB)": 73.06, | |
| "step": 245, | |
| "token_acc": 0.8758281279575999, | |
| "train_speed(iter/s)": 0.0298 | |
| }, | |
| { | |
| "epoch": 0.07999680012799489, | |
| "grad_norm": 0.06662896613700448, | |
| "learning_rate": 1.9878905881817254e-05, | |
| "loss": 0.47487664222717285, | |
| "memory(GiB)": 73.06, | |
| "step": 250, | |
| "token_acc": 0.8239743295897318, | |
| "train_speed(iter/s)": 0.029855 | |
| }, | |
| { | |
| "epoch": 0.07999680012799489, | |
| "eval_loss": 0.6802101731300354, | |
| "eval_runtime": 108.8605, | |
| "eval_samples_per_second": 184.53, | |
| "eval_steps_per_second": 0.928, | |
| "eval_token_acc": 0.8656411339267154, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.08159673613055478, | |
| "grad_norm": 0.058967589577093804, | |
| "learning_rate": 1.9870716120959462e-05, | |
| "loss": 0.4691306591033936, | |
| "memory(GiB)": 73.24, | |
| "step": 255, | |
| "token_acc": 0.8612697569398327, | |
| "train_speed(iter/s)": 0.029606 | |
| }, | |
| { | |
| "epoch": 0.08319667213311467, | |
| "grad_norm": 0.057353651690814994, | |
| "learning_rate": 1.986226020384359e-05, | |
| "loss": 0.46143622398376466, | |
| "memory(GiB)": 73.24, | |
| "step": 260, | |
| "token_acc": 0.8685547371094742, | |
| "train_speed(iter/s)": 0.029681 | |
| }, | |
| { | |
| "epoch": 0.08479660813567458, | |
| "grad_norm": 0.05409688809510523, | |
| "learning_rate": 1.9853538358476933e-05, | |
| "loss": 0.4704445838928223, | |
| "memory(GiB)": 73.24, | |
| "step": 265, | |
| "token_acc": 0.8804637020144431, | |
| "train_speed(iter/s)": 0.02979 | |
| }, | |
| { | |
| "epoch": 0.08639654413823447, | |
| "grad_norm": 0.06968473514476099, | |
| "learning_rate": 1.9844550820037326e-05, | |
| "loss": 0.4717890739440918, | |
| "memory(GiB)": 73.24, | |
| "step": 270, | |
| "token_acc": 0.8638003174145145, | |
| "train_speed(iter/s)": 0.029834 | |
| }, | |
| { | |
| "epoch": 0.08799648014079436, | |
| "grad_norm": 0.06009720175343309, | |
| "learning_rate": 1.9835297830866827e-05, | |
| "loss": 0.4709662437438965, | |
| "memory(GiB)": 73.24, | |
| "step": 275, | |
| "token_acc": 0.8634590377113134, | |
| "train_speed(iter/s)": 0.029835 | |
| }, | |
| { | |
| "epoch": 0.08959641614335427, | |
| "grad_norm": 0.058778539356308675, | |
| "learning_rate": 1.9825779640465157e-05, | |
| "loss": 0.47084336280822753, | |
| "memory(GiB)": 73.24, | |
| "step": 280, | |
| "token_acc": 0.9203691779351793, | |
| "train_speed(iter/s)": 0.029942 | |
| }, | |
| { | |
| "epoch": 0.09119635214591416, | |
| "grad_norm": 0.054325246749067864, | |
| "learning_rate": 1.9815996505483e-05, | |
| "loss": 0.4666774749755859, | |
| "memory(GiB)": 73.24, | |
| "step": 285, | |
| "token_acc": 0.8521723454119344, | |
| "train_speed(iter/s)": 0.029906 | |
| }, | |
| { | |
| "epoch": 0.09279628814847406, | |
| "grad_norm": 0.058698263071843435, | |
| "learning_rate": 1.9805948689715043e-05, | |
| "loss": 0.45826416015625, | |
| "memory(GiB)": 73.24, | |
| "step": 290, | |
| "token_acc": 0.8421138211382114, | |
| "train_speed(iter/s)": 0.029895 | |
| }, | |
| { | |
| "epoch": 0.09439622415103396, | |
| "grad_norm": 0.05517972536097747, | |
| "learning_rate": 1.979563646409291e-05, | |
| "loss": 0.47627692222595214, | |
| "memory(GiB)": 73.24, | |
| "step": 295, | |
| "token_acc": 0.8784122999686226, | |
| "train_speed(iter/s)": 0.029966 | |
| }, | |
| { | |
| "epoch": 0.09599616015359386, | |
| "grad_norm": 0.06223926082468345, | |
| "learning_rate": 1.9785060106677818e-05, | |
| "loss": 0.4711057186126709, | |
| "memory(GiB)": 73.24, | |
| "step": 300, | |
| "token_acc": 0.876372039283651, | |
| "train_speed(iter/s)": 0.029911 | |
| }, | |
| { | |
| "epoch": 0.09759609615615375, | |
| "grad_norm": 0.060678733702642235, | |
| "learning_rate": 1.97742199026531e-05, | |
| "loss": 0.46833024024963377, | |
| "memory(GiB)": 73.24, | |
| "step": 305, | |
| "token_acc": 0.8586995355484102, | |
| "train_speed(iter/s)": 0.029998 | |
| }, | |
| { | |
| "epoch": 0.09919603215871366, | |
| "grad_norm": 0.06117494885421727, | |
| "learning_rate": 1.9763116144316506e-05, | |
| "loss": 0.4692807197570801, | |
| "memory(GiB)": 73.24, | |
| "step": 310, | |
| "token_acc": 0.8383072793304911, | |
| "train_speed(iter/s)": 0.029983 | |
| }, | |
| { | |
| "epoch": 0.10079596816127355, | |
| "grad_norm": 0.059512004342169564, | |
| "learning_rate": 1.9751749131072335e-05, | |
| "loss": 0.462421178817749, | |
| "memory(GiB)": 73.24, | |
| "step": 315, | |
| "token_acc": 0.865073787772312, | |
| "train_speed(iter/s)": 0.029959 | |
| }, | |
| { | |
| "epoch": 0.10239590416383344, | |
| "grad_norm": 0.05759903892800583, | |
| "learning_rate": 1.9740119169423337e-05, | |
| "loss": 0.4749638080596924, | |
| "memory(GiB)": 73.24, | |
| "step": 320, | |
| "token_acc": 0.8657438292194797, | |
| "train_speed(iter/s)": 0.030064 | |
| }, | |
| { | |
| "epoch": 0.10399584016639335, | |
| "grad_norm": 0.05512670495542287, | |
| "learning_rate": 1.9728226572962474e-05, | |
| "loss": 0.48053979873657227, | |
| "memory(GiB)": 73.24, | |
| "step": 325, | |
| "token_acc": 0.9068181818181819, | |
| "train_speed(iter/s)": 0.03004 | |
| }, | |
| { | |
| "epoch": 0.10559577616895324, | |
| "grad_norm": 0.05723038100011267, | |
| "learning_rate": 1.9716071662364454e-05, | |
| "loss": 0.47551665306091306, | |
| "memory(GiB)": 73.24, | |
| "step": 330, | |
| "token_acc": 0.8362432269717038, | |
| "train_speed(iter/s)": 0.030003 | |
| }, | |
| { | |
| "epoch": 0.10719571217151314, | |
| "grad_norm": 0.057638605082885846, | |
| "learning_rate": 1.970365476537707e-05, | |
| "loss": 0.4652701854705811, | |
| "memory(GiB)": 73.24, | |
| "step": 335, | |
| "token_acc": 0.8735049401976079, | |
| "train_speed(iter/s)": 0.030082 | |
| }, | |
| { | |
| "epoch": 0.10879564817407304, | |
| "grad_norm": 0.05903871731521889, | |
| "learning_rate": 1.9690976216812397e-05, | |
| "loss": 0.4698742389678955, | |
| "memory(GiB)": 73.24, | |
| "step": 340, | |
| "token_acc": 0.8620361560418649, | |
| "train_speed(iter/s)": 0.030027 | |
| }, | |
| { | |
| "epoch": 0.11039558417663294, | |
| "grad_norm": 0.053856521964694516, | |
| "learning_rate": 1.9678036358537726e-05, | |
| "loss": 0.4701416015625, | |
| "memory(GiB)": 73.24, | |
| "step": 345, | |
| "token_acc": 0.8708435421771089, | |
| "train_speed(iter/s)": 0.03002 | |
| }, | |
| { | |
| "epoch": 0.11199552017919283, | |
| "grad_norm": 0.05586893539038131, | |
| "learning_rate": 1.966483553946637e-05, | |
| "loss": 0.47447028160095217, | |
| "memory(GiB)": 73.24, | |
| "step": 350, | |
| "token_acc": 0.8617533718689788, | |
| "train_speed(iter/s)": 0.030041 | |
| }, | |
| { | |
| "epoch": 0.11359545618175274, | |
| "grad_norm": 0.052599438001953325, | |
| "learning_rate": 1.9651374115548255e-05, | |
| "loss": 0.4637298583984375, | |
| "memory(GiB)": 73.24, | |
| "step": 355, | |
| "token_acc": 0.8874341610233258, | |
| "train_speed(iter/s)": 0.029967 | |
| }, | |
| { | |
| "epoch": 0.11519539218431263, | |
| "grad_norm": 0.05804143123407663, | |
| "learning_rate": 1.9637652449760297e-05, | |
| "loss": 0.4660144329071045, | |
| "memory(GiB)": 73.24, | |
| "step": 360, | |
| "token_acc": 0.8349885408708938, | |
| "train_speed(iter/s)": 0.030034 | |
| }, | |
| { | |
| "epoch": 0.11679532818687252, | |
| "grad_norm": 0.06055547849970778, | |
| "learning_rate": 1.9623670912096656e-05, | |
| "loss": 0.4716383934020996, | |
| "memory(GiB)": 73.24, | |
| "step": 365, | |
| "token_acc": 0.8751012473675684, | |
| "train_speed(iter/s)": 0.02998 | |
| }, | |
| { | |
| "epoch": 0.11839526418943243, | |
| "grad_norm": 0.058520598293842735, | |
| "learning_rate": 1.9609429879558726e-05, | |
| "loss": 0.46298699378967284, | |
| "memory(GiB)": 73.24, | |
| "step": 370, | |
| "token_acc": 0.8553921568627451, | |
| "train_speed(iter/s)": 0.029931 | |
| }, | |
| { | |
| "epoch": 0.11999520019199232, | |
| "grad_norm": 0.058584318589478955, | |
| "learning_rate": 1.9594929736144978e-05, | |
| "loss": 0.4756875514984131, | |
| "memory(GiB)": 73.24, | |
| "step": 375, | |
| "token_acc": 0.8618346545866364, | |
| "train_speed(iter/s)": 0.030006 | |
| }, | |
| { | |
| "epoch": 0.12159513619455221, | |
| "grad_norm": 0.05966533070217228, | |
| "learning_rate": 1.958017087284061e-05, | |
| "loss": 0.4596414089202881, | |
| "memory(GiB)": 73.24, | |
| "step": 380, | |
| "token_acc": 0.8836156297165856, | |
| "train_speed(iter/s)": 0.029968 | |
| }, | |
| { | |
| "epoch": 0.12319507219711212, | |
| "grad_norm": 0.06510894340277039, | |
| "learning_rate": 1.9565153687607006e-05, | |
| "loss": 0.4687026023864746, | |
| "memory(GiB)": 73.24, | |
| "step": 385, | |
| "token_acc": 0.8694005270092227, | |
| "train_speed(iter/s)": 0.029956 | |
| }, | |
| { | |
| "epoch": 0.12479500819967201, | |
| "grad_norm": 0.05180588304383506, | |
| "learning_rate": 1.9549878585371006e-05, | |
| "loss": 0.4649878978729248, | |
| "memory(GiB)": 73.24, | |
| "step": 390, | |
| "token_acc": 0.8677233429394813, | |
| "train_speed(iter/s)": 0.029999 | |
| }, | |
| { | |
| "epoch": 0.1263949442022319, | |
| "grad_norm": 0.060875424512666344, | |
| "learning_rate": 1.9534345978013972e-05, | |
| "loss": 0.47073874473571775, | |
| "memory(GiB)": 73.24, | |
| "step": 395, | |
| "token_acc": 0.8484663512894858, | |
| "train_speed(iter/s)": 0.029947 | |
| }, | |
| { | |
| "epoch": 0.1279948802047918, | |
| "grad_norm": 0.0571374353277554, | |
| "learning_rate": 1.9518556284360696e-05, | |
| "loss": 0.4666412353515625, | |
| "memory(GiB)": 73.24, | |
| "step": 400, | |
| "token_acc": 0.8975701436434421, | |
| "train_speed(iter/s)": 0.029998 | |
| }, | |
| { | |
| "epoch": 0.1295948162073517, | |
| "grad_norm": 0.05215050598306155, | |
| "learning_rate": 1.9502509930168113e-05, | |
| "loss": 0.4628121376037598, | |
| "memory(GiB)": 73.24, | |
| "step": 405, | |
| "token_acc": 0.8816677696889477, | |
| "train_speed(iter/s)": 0.029966 | |
| }, | |
| { | |
| "epoch": 0.1311947522099116, | |
| "grad_norm": 0.07947570193916972, | |
| "learning_rate": 1.9486207348113803e-05, | |
| "loss": 0.4593012809753418, | |
| "memory(GiB)": 73.24, | |
| "step": 410, | |
| "token_acc": 0.8692473832862602, | |
| "train_speed(iter/s)": 0.029911 | |
| }, | |
| { | |
| "epoch": 0.1327946882124715, | |
| "grad_norm": 0.07262611466641217, | |
| "learning_rate": 1.946964897778433e-05, | |
| "loss": 0.47004990577697753, | |
| "memory(GiB)": 73.24, | |
| "step": 415, | |
| "token_acc": 0.8736337958983176, | |
| "train_speed(iter/s)": 0.029969 | |
| }, | |
| { | |
| "epoch": 0.1343946242150314, | |
| "grad_norm": 0.053754461298334506, | |
| "learning_rate": 1.9452835265663404e-05, | |
| "loss": 0.4695271015167236, | |
| "memory(GiB)": 73.24, | |
| "step": 420, | |
| "token_acc": 0.8747993579454254, | |
| "train_speed(iter/s)": 0.029901 | |
| }, | |
| { | |
| "epoch": 0.1359945602175913, | |
| "grad_norm": 0.0742051800083311, | |
| "learning_rate": 1.9435766665119823e-05, | |
| "loss": 0.47011446952819824, | |
| "memory(GiB)": 73.24, | |
| "step": 425, | |
| "token_acc": 0.8356736242884251, | |
| "train_speed(iter/s)": 0.029856 | |
| }, | |
| { | |
| "epoch": 0.13759449622015119, | |
| "grad_norm": 0.06429200177825628, | |
| "learning_rate": 1.941844363639525e-05, | |
| "loss": 0.476796817779541, | |
| "memory(GiB)": 73.24, | |
| "step": 430, | |
| "token_acc": 0.869019972131909, | |
| "train_speed(iter/s)": 0.0299 | |
| }, | |
| { | |
| "epoch": 0.13919443222271108, | |
| "grad_norm": 0.06544854557851852, | |
| "learning_rate": 1.9400866646591816e-05, | |
| "loss": 0.4666853904724121, | |
| "memory(GiB)": 73.24, | |
| "step": 435, | |
| "token_acc": 0.8204667863554758, | |
| "train_speed(iter/s)": 0.029847 | |
| }, | |
| { | |
| "epoch": 0.140794368225271, | |
| "grad_norm": 0.0546565929911768, | |
| "learning_rate": 1.9383036169659513e-05, | |
| "loss": 0.4738778591156006, | |
| "memory(GiB)": 73.24, | |
| "step": 440, | |
| "token_acc": 0.8605809128630706, | |
| "train_speed(iter/s)": 0.029855 | |
| }, | |
| { | |
| "epoch": 0.1423943042278309, | |
| "grad_norm": 0.06789336848906298, | |
| "learning_rate": 1.936495268638342e-05, | |
| "loss": 0.47726120948791506, | |
| "memory(GiB)": 73.24, | |
| "step": 445, | |
| "token_acc": 0.8404369243949454, | |
| "train_speed(iter/s)": 0.029875 | |
| }, | |
| { | |
| "epoch": 0.14399424023039079, | |
| "grad_norm": 0.049909982274150465, | |
| "learning_rate": 1.934661668437073e-05, | |
| "loss": 0.47165632247924805, | |
| "memory(GiB)": 73.24, | |
| "step": 450, | |
| "token_acc": 0.848471615720524, | |
| "train_speed(iter/s)": 0.029826 | |
| }, | |
| { | |
| "epoch": 0.14559417623295068, | |
| "grad_norm": 0.057441474731933166, | |
| "learning_rate": 1.932802865803763e-05, | |
| "loss": 0.4703391075134277, | |
| "memory(GiB)": 73.24, | |
| "step": 455, | |
| "token_acc": 0.8466442358774571, | |
| "train_speed(iter/s)": 0.029877 | |
| }, | |
| { | |
| "epoch": 0.14719411223551057, | |
| "grad_norm": 0.07263904251491092, | |
| "learning_rate": 1.930918910859592e-05, | |
| "loss": 0.467697811126709, | |
| "memory(GiB)": 73.24, | |
| "step": 460, | |
| "token_acc": 0.8491142333536957, | |
| "train_speed(iter/s)": 0.02984 | |
| }, | |
| { | |
| "epoch": 0.14879404823807046, | |
| "grad_norm": 0.06769237623086669, | |
| "learning_rate": 1.9290098544039546e-05, | |
| "loss": 0.46541628837585447, | |
| "memory(GiB)": 73.24, | |
| "step": 465, | |
| "token_acc": 0.8555353301340394, | |
| "train_speed(iter/s)": 0.02978 | |
| }, | |
| { | |
| "epoch": 0.15039398424063039, | |
| "grad_norm": 0.06751583633556477, | |
| "learning_rate": 1.927075747913088e-05, | |
| "loss": 0.47134056091308596, | |
| "memory(GiB)": 73.24, | |
| "step": 470, | |
| "token_acc": 0.8708000507163687, | |
| "train_speed(iter/s)": 0.029832 | |
| }, | |
| { | |
| "epoch": 0.15199392024319028, | |
| "grad_norm": 0.0539492567012165, | |
| "learning_rate": 1.9251166435386837e-05, | |
| "loss": 0.4645866394042969, | |
| "memory(GiB)": 73.24, | |
| "step": 475, | |
| "token_acc": 0.849832526981764, | |
| "train_speed(iter/s)": 0.029779 | |
| }, | |
| { | |
| "epoch": 0.15359385624575017, | |
| "grad_norm": 0.06038706866556876, | |
| "learning_rate": 1.923132594106483e-05, | |
| "loss": 0.46890692710876464, | |
| "memory(GiB)": 73.24, | |
| "step": 480, | |
| "token_acc": 0.8665925514174542, | |
| "train_speed(iter/s)": 0.02976 | |
| }, | |
| { | |
| "epoch": 0.15519379224831006, | |
| "grad_norm": 0.05215840717634863, | |
| "learning_rate": 1.92112365311485e-05, | |
| "loss": 0.46829919815063475, | |
| "memory(GiB)": 73.24, | |
| "step": 485, | |
| "token_acc": 0.861963565228023, | |
| "train_speed(iter/s)": 0.029794 | |
| }, | |
| { | |
| "epoch": 0.15679372825086996, | |
| "grad_norm": 0.06554142579397569, | |
| "learning_rate": 1.919089874733332e-05, | |
| "loss": 0.4702622413635254, | |
| "memory(GiB)": 73.24, | |
| "step": 490, | |
| "token_acc": 0.8809186723297153, | |
| "train_speed(iter/s)": 0.029747 | |
| }, | |
| { | |
| "epoch": 0.15839366425342985, | |
| "grad_norm": 0.0601172563145885, | |
| "learning_rate": 1.9170313138011964e-05, | |
| "loss": 0.46490135192871096, | |
| "memory(GiB)": 73.24, | |
| "step": 495, | |
| "token_acc": 0.8890911637025627, | |
| "train_speed(iter/s)": 0.02977 | |
| }, | |
| { | |
| "epoch": 0.15999360025598977, | |
| "grad_norm": 0.05924399402367875, | |
| "learning_rate": 1.9149480258259535e-05, | |
| "loss": 0.46698894500732424, | |
| "memory(GiB)": 73.24, | |
| "step": 500, | |
| "token_acc": 0.8781434114096853, | |
| "train_speed(iter/s)": 0.029766 | |
| }, | |
| { | |
| "epoch": 0.15999360025598977, | |
| "eval_loss": 0.677643895149231, | |
| "eval_runtime": 109.3458, | |
| "eval_samples_per_second": 183.711, | |
| "eval_steps_per_second": 0.924, | |
| "eval_token_acc": 0.8661408286670019, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.16159353625854966, | |
| "grad_norm": 0.05118070522939682, | |
| "learning_rate": 1.9128400669818586e-05, | |
| "loss": 0.4606743812561035, | |
| "memory(GiB)": 73.24, | |
| "step": 505, | |
| "token_acc": 0.8727327237295758, | |
| "train_speed(iter/s)": 0.029628 | |
| }, | |
| { | |
| "epoch": 0.16319347226110956, | |
| "grad_norm": 0.05904937387674259, | |
| "learning_rate": 1.9107074941083987e-05, | |
| "loss": 0.47115492820739746, | |
| "memory(GiB)": 73.24, | |
| "step": 510, | |
| "token_acc": 0.8801781737193763, | |
| "train_speed(iter/s)": 0.029663 | |
| }, | |
| { | |
| "epoch": 0.16479340826366945, | |
| "grad_norm": 0.061211680590962145, | |
| "learning_rate": 1.9085503647087588e-05, | |
| "loss": 0.46154184341430665, | |
| "memory(GiB)": 73.24, | |
| "step": 515, | |
| "token_acc": 0.8573438874230431, | |
| "train_speed(iter/s)": 0.029714 | |
| }, | |
| { | |
| "epoch": 0.16639334426622934, | |
| "grad_norm": 0.05461804298242196, | |
| "learning_rate": 1.906368736948272e-05, | |
| "loss": 0.46891465187072756, | |
| "memory(GiB)": 73.24, | |
| "step": 520, | |
| "token_acc": 0.8665508756694167, | |
| "train_speed(iter/s)": 0.029721 | |
| }, | |
| { | |
| "epoch": 0.16799328026878924, | |
| "grad_norm": 0.059072521440841075, | |
| "learning_rate": 1.9041626696528503e-05, | |
| "loss": 0.4666083812713623, | |
| "memory(GiB)": 73.24, | |
| "step": 525, | |
| "token_acc": 0.8742783835792175, | |
| "train_speed(iter/s)": 0.029735 | |
| }, | |
| { | |
| "epoch": 0.16959321627134916, | |
| "grad_norm": 0.06762878495647719, | |
| "learning_rate": 1.9019322223073997e-05, | |
| "loss": 0.4684437274932861, | |
| "memory(GiB)": 73.24, | |
| "step": 530, | |
| "token_acc": 0.8906074591493077, | |
| "train_speed(iter/s)": 0.029782 | |
| }, | |
| { | |
| "epoch": 0.17119315227390905, | |
| "grad_norm": 0.05741557316745661, | |
| "learning_rate": 1.899677455054215e-05, | |
| "loss": 0.4690097332000732, | |
| "memory(GiB)": 73.24, | |
| "step": 535, | |
| "token_acc": 0.8231878958479943, | |
| "train_speed(iter/s)": 0.029785 | |
| }, | |
| { | |
| "epoch": 0.17279308827646894, | |
| "grad_norm": 0.049026865135578496, | |
| "learning_rate": 1.8973984286913584e-05, | |
| "loss": 0.469140625, | |
| "memory(GiB)": 73.24, | |
| "step": 540, | |
| "token_acc": 0.8849415539766216, | |
| "train_speed(iter/s)": 0.029789 | |
| }, | |
| { | |
| "epoch": 0.17439302427902884, | |
| "grad_norm": 0.059746465018255104, | |
| "learning_rate": 1.895095204671021e-05, | |
| "loss": 0.4646149158477783, | |
| "memory(GiB)": 73.24, | |
| "step": 545, | |
| "token_acc": 0.8944385405596883, | |
| "train_speed(iter/s)": 0.029813 | |
| }, | |
| { | |
| "epoch": 0.17599296028158873, | |
| "grad_norm": 0.049833714934798115, | |
| "learning_rate": 1.892767845097864e-05, | |
| "loss": 0.47077240943908694, | |
| "memory(GiB)": 73.24, | |
| "step": 550, | |
| "token_acc": 0.8640860961638605, | |
| "train_speed(iter/s)": 0.029794 | |
| }, | |
| { | |
| "epoch": 0.17759289628414862, | |
| "grad_norm": 0.06593845007149325, | |
| "learning_rate": 1.890416412727346e-05, | |
| "loss": 0.46265759468078616, | |
| "memory(GiB)": 73.24, | |
| "step": 555, | |
| "token_acc": 0.8249895412076419, | |
| "train_speed(iter/s)": 0.02984 | |
| }, | |
| { | |
| "epoch": 0.17919283228670854, | |
| "grad_norm": 0.058254003445636866, | |
| "learning_rate": 1.88804097096403e-05, | |
| "loss": 0.459829044342041, | |
| "memory(GiB)": 73.24, | |
| "step": 560, | |
| "token_acc": 0.8835873095178616, | |
| "train_speed(iter/s)": 0.029842 | |
| }, | |
| { | |
| "epoch": 0.18079276828926844, | |
| "grad_norm": 0.07335953644753283, | |
| "learning_rate": 1.8856415838598738e-05, | |
| "loss": 0.45765042304992676, | |
| "memory(GiB)": 73.24, | |
| "step": 565, | |
| "token_acc": 0.8755007210382951, | |
| "train_speed(iter/s)": 0.029818 | |
| }, | |
| { | |
| "epoch": 0.18239270429182833, | |
| "grad_norm": 0.06659181547700674, | |
| "learning_rate": 1.8832183161125026e-05, | |
| "loss": 0.4609128475189209, | |
| "memory(GiB)": 73.24, | |
| "step": 570, | |
| "token_acc": 0.8344311377245509, | |
| "train_speed(iter/s)": 0.029871 | |
| }, | |
| { | |
| "epoch": 0.18399264029438822, | |
| "grad_norm": 0.05836437871791382, | |
| "learning_rate": 1.8807712330634645e-05, | |
| "loss": 0.4691438674926758, | |
| "memory(GiB)": 73.24, | |
| "step": 575, | |
| "token_acc": 0.8848027659908848, | |
| "train_speed(iter/s)": 0.029828 | |
| }, | |
| { | |
| "epoch": 0.18559257629694811, | |
| "grad_norm": 0.05735059462858394, | |
| "learning_rate": 1.87830040069647e-05, | |
| "loss": 0.4602513790130615, | |
| "memory(GiB)": 73.24, | |
| "step": 580, | |
| "token_acc": 0.8959147903465012, | |
| "train_speed(iter/s)": 0.029816 | |
| }, | |
| { | |
| "epoch": 0.187192512299508, | |
| "grad_norm": 0.05337219773586585, | |
| "learning_rate": 1.87580588563561e-05, | |
| "loss": 0.46318631172180175, | |
| "memory(GiB)": 73.24, | |
| "step": 585, | |
| "token_acc": 0.8725881039706586, | |
| "train_speed(iter/s)": 0.029851 | |
| }, | |
| { | |
| "epoch": 0.18879244830206793, | |
| "grad_norm": 0.05886716832883729, | |
| "learning_rate": 1.873287755143563e-05, | |
| "loss": 0.4604507923126221, | |
| "memory(GiB)": 73.24, | |
| "step": 590, | |
| "token_acc": 0.9041755130927105, | |
| "train_speed(iter/s)": 0.029822 | |
| }, | |
| { | |
| "epoch": 0.19039238430462782, | |
| "grad_norm": 0.053483810048332456, | |
| "learning_rate": 1.8707460771197773e-05, | |
| "loss": 0.46618080139160156, | |
| "memory(GiB)": 73.24, | |
| "step": 595, | |
| "token_acc": 0.8785046728971962, | |
| "train_speed(iter/s)": 0.029819 | |
| }, | |
| { | |
| "epoch": 0.1919923203071877, | |
| "grad_norm": 0.0518592001281956, | |
| "learning_rate": 1.868180920098644e-05, | |
| "loss": 0.4680916786193848, | |
| "memory(GiB)": 73.24, | |
| "step": 600, | |
| "token_acc": 0.8467063770147162, | |
| "train_speed(iter/s)": 0.029843 | |
| }, | |
| { | |
| "epoch": 0.1935922563097476, | |
| "grad_norm": 0.07018232236413237, | |
| "learning_rate": 1.8655923532476463e-05, | |
| "loss": 0.46170759201049805, | |
| "memory(GiB)": 73.24, | |
| "step": 605, | |
| "token_acc": 0.889030612244898, | |
| "train_speed(iter/s)": 0.02981 | |
| }, | |
| { | |
| "epoch": 0.1951921923123075, | |
| "grad_norm": 0.06030421269833889, | |
| "learning_rate": 1.8629804463654956e-05, | |
| "loss": 0.46511187553405764, | |
| "memory(GiB)": 73.24, | |
| "step": 610, | |
| "token_acc": 0.8554680664916885, | |
| "train_speed(iter/s)": 0.029852 | |
| }, | |
| { | |
| "epoch": 0.1967921283148674, | |
| "grad_norm": 0.056137765321266526, | |
| "learning_rate": 1.8603452698802498e-05, | |
| "loss": 0.47327299118041993, | |
| "memory(GiB)": 76.61, | |
| "step": 615, | |
| "token_acc": 0.8645191852202747, | |
| "train_speed(iter/s)": 0.029831 | |
| }, | |
| { | |
| "epoch": 0.1983920643174273, | |
| "grad_norm": 0.05458475201274465, | |
| "learning_rate": 1.857686894847413e-05, | |
| "loss": 0.45963249206542967, | |
| "memory(GiB)": 76.61, | |
| "step": 620, | |
| "token_acc": 0.8517509197438343, | |
| "train_speed(iter/s)": 0.029791 | |
| }, | |
| { | |
| "epoch": 0.1999920003199872, | |
| "grad_norm": 0.059902578480064236, | |
| "learning_rate": 1.8550053929480202e-05, | |
| "loss": 0.4687147617340088, | |
| "memory(GiB)": 76.61, | |
| "step": 625, | |
| "token_acc": 0.8958185683912119, | |
| "train_speed(iter/s)": 0.029833 | |
| }, | |
| { | |
| "epoch": 0.2015919363225471, | |
| "grad_norm": 0.0539478773118384, | |
| "learning_rate": 1.8523008364867056e-05, | |
| "loss": 0.4696544647216797, | |
| "memory(GiB)": 76.61, | |
| "step": 630, | |
| "token_acc": 0.8439355385920272, | |
| "train_speed(iter/s)": 0.029796 | |
| }, | |
| { | |
| "epoch": 0.203191872325107, | |
| "grad_norm": 0.05688926646164217, | |
| "learning_rate": 1.8495732983897504e-05, | |
| "loss": 0.4628334045410156, | |
| "memory(GiB)": 76.61, | |
| "step": 635, | |
| "token_acc": 0.8406656465187452, | |
| "train_speed(iter/s)": 0.029775 | |
| }, | |
| { | |
| "epoch": 0.20479180832766689, | |
| "grad_norm": 0.055104479428209605, | |
| "learning_rate": 1.8468228522031197e-05, | |
| "loss": 0.4559271812438965, | |
| "memory(GiB)": 76.61, | |
| "step": 640, | |
| "token_acc": 0.8823529411764706, | |
| "train_speed(iter/s)": 0.029794 | |
| }, | |
| { | |
| "epoch": 0.20639174433022678, | |
| "grad_norm": 0.058080447436547736, | |
| "learning_rate": 1.8440495720904758e-05, | |
| "loss": 0.4649765968322754, | |
| "memory(GiB)": 76.61, | |
| "step": 645, | |
| "token_acc": 0.8708735027753433, | |
| "train_speed(iter/s)": 0.029752 | |
| }, | |
| { | |
| "epoch": 0.2079916803327867, | |
| "grad_norm": 0.06300003986546152, | |
| "learning_rate": 1.8412535328311813e-05, | |
| "loss": 0.47095327377319335, | |
| "memory(GiB)": 76.61, | |
| "step": 650, | |
| "token_acc": 0.8504976200778883, | |
| "train_speed(iter/s)": 0.029755 | |
| }, | |
| { | |
| "epoch": 0.2095916163353466, | |
| "grad_norm": 0.06584526718748161, | |
| "learning_rate": 1.8384348098182815e-05, | |
| "loss": 0.46697392463684084, | |
| "memory(GiB)": 76.61, | |
| "step": 655, | |
| "token_acc": 0.8224407171775593, | |
| "train_speed(iter/s)": 0.029762 | |
| }, | |
| { | |
| "epoch": 0.21119155233790649, | |
| "grad_norm": 0.07147957728971413, | |
| "learning_rate": 1.8355934790564718e-05, | |
| "loss": 0.4684570789337158, | |
| "memory(GiB)": 76.61, | |
| "step": 660, | |
| "token_acc": 0.8842165898617511, | |
| "train_speed(iter/s)": 0.029723 | |
| }, | |
| { | |
| "epoch": 0.21279148834046638, | |
| "grad_norm": 0.06592046292925295, | |
| "learning_rate": 1.832729617160047e-05, | |
| "loss": 0.461454439163208, | |
| "memory(GiB)": 76.61, | |
| "step": 665, | |
| "token_acc": 0.9114801444043321, | |
| "train_speed(iter/s)": 0.02976 | |
| }, | |
| { | |
| "epoch": 0.21439142434302627, | |
| "grad_norm": 0.0656829490109071, | |
| "learning_rate": 1.8298433013508384e-05, | |
| "loss": 0.46404447555541994, | |
| "memory(GiB)": 76.61, | |
| "step": 670, | |
| "token_acc": 0.8516549891278087, | |
| "train_speed(iter/s)": 0.029736 | |
| }, | |
| { | |
| "epoch": 0.21599136034558616, | |
| "grad_norm": 0.05417998837874903, | |
| "learning_rate": 1.826934609456129e-05, | |
| "loss": 0.47208566665649415, | |
| "memory(GiB)": 76.61, | |
| "step": 675, | |
| "token_acc": 0.8798815733822078, | |
| "train_speed(iter/s)": 0.029718 | |
| }, | |
| { | |
| "epoch": 0.21759129634814608, | |
| "grad_norm": 0.06917195844649823, | |
| "learning_rate": 1.8240036199065546e-05, | |
| "loss": 0.4724391460418701, | |
| "memory(GiB)": 76.61, | |
| "step": 680, | |
| "token_acc": 0.875845675626257, | |
| "train_speed(iter/s)": 0.029745 | |
| }, | |
| { | |
| "epoch": 0.21919123235070598, | |
| "grad_norm": 0.055849189404917746, | |
| "learning_rate": 1.8210504117339917e-05, | |
| "loss": 0.463816499710083, | |
| "memory(GiB)": 76.61, | |
| "step": 685, | |
| "token_acc": 0.8841904379268782, | |
| "train_speed(iter/s)": 0.029711 | |
| }, | |
| { | |
| "epoch": 0.22079116835326587, | |
| "grad_norm": 0.059563786969142496, | |
| "learning_rate": 1.8180750645694236e-05, | |
| "loss": 0.4678086757659912, | |
| "memory(GiB)": 76.61, | |
| "step": 690, | |
| "token_acc": 0.8675231977159172, | |
| "train_speed(iter/s)": 0.029714 | |
| }, | |
| { | |
| "epoch": 0.22239110435582576, | |
| "grad_norm": 0.05908606421708839, | |
| "learning_rate": 1.8150776586407957e-05, | |
| "loss": 0.46315860748291016, | |
| "memory(GiB)": 76.61, | |
| "step": 695, | |
| "token_acc": 0.8914956011730205, | |
| "train_speed(iter/s)": 0.029731 | |
| }, | |
| { | |
| "epoch": 0.22399104035838566, | |
| "grad_norm": 0.05617530731492468, | |
| "learning_rate": 1.8120582747708503e-05, | |
| "loss": 0.46682062149047854, | |
| "memory(GiB)": 76.61, | |
| "step": 700, | |
| "token_acc": 0.8805088596092685, | |
| "train_speed(iter/s)": 0.029689 | |
| }, | |
| { | |
| "epoch": 0.22559097636094555, | |
| "grad_norm": 0.06138477303861948, | |
| "learning_rate": 1.8090169943749477e-05, | |
| "loss": 0.47155141830444336, | |
| "memory(GiB)": 76.61, | |
| "step": 705, | |
| "token_acc": 0.8753766681015928, | |
| "train_speed(iter/s)": 0.029703 | |
| }, | |
| { | |
| "epoch": 0.22719091236350547, | |
| "grad_norm": 0.07073141016351848, | |
| "learning_rate": 1.8059538994588715e-05, | |
| "loss": 0.45953845977783203, | |
| "memory(GiB)": 76.61, | |
| "step": 710, | |
| "token_acc": 0.8449233877757198, | |
| "train_speed(iter/s)": 0.0297 | |
| }, | |
| { | |
| "epoch": 0.22879084836606536, | |
| "grad_norm": 0.06266619359839708, | |
| "learning_rate": 1.8028690726166172e-05, | |
| "loss": 0.4604049205780029, | |
| "memory(GiB)": 76.61, | |
| "step": 715, | |
| "token_acc": 0.8688032048072108, | |
| "train_speed(iter/s)": 0.02966 | |
| }, | |
| { | |
| "epoch": 0.23039078436862526, | |
| "grad_norm": 0.0563660774004587, | |
| "learning_rate": 1.7997625970281652e-05, | |
| "loss": 0.4622708797454834, | |
| "memory(GiB)": 76.61, | |
| "step": 720, | |
| "token_acc": 0.8698216735253772, | |
| "train_speed(iter/s)": 0.029685 | |
| }, | |
| { | |
| "epoch": 0.23199072037118515, | |
| "grad_norm": 0.06596213612143108, | |
| "learning_rate": 1.796634556457236e-05, | |
| "loss": 0.4681892395019531, | |
| "memory(GiB)": 76.61, | |
| "step": 725, | |
| "token_acc": 0.8842619184376795, | |
| "train_speed(iter/s)": 0.029661 | |
| }, | |
| { | |
| "epoch": 0.23359065637374504, | |
| "grad_norm": 0.05364579438678848, | |
| "learning_rate": 1.793485035249036e-05, | |
| "loss": 0.46258745193481443, | |
| "memory(GiB)": 76.61, | |
| "step": 730, | |
| "token_acc": 0.8599531615925059, | |
| "train_speed(iter/s)": 0.02965 | |
| }, | |
| { | |
| "epoch": 0.23519059237630494, | |
| "grad_norm": 0.07509450433159735, | |
| "learning_rate": 1.7903141183279776e-05, | |
| "loss": 0.47242441177368166, | |
| "memory(GiB)": 76.61, | |
| "step": 735, | |
| "token_acc": 0.8404958677685951, | |
| "train_speed(iter/s)": 0.029665 | |
| }, | |
| { | |
| "epoch": 0.23679052837886486, | |
| "grad_norm": 0.06478313540282635, | |
| "learning_rate": 1.7871218911953942e-05, | |
| "loss": 0.4565444469451904, | |
| "memory(GiB)": 76.61, | |
| "step": 740, | |
| "token_acc": 0.8338650865998177, | |
| "train_speed(iter/s)": 0.029634 | |
| }, | |
| { | |
| "epoch": 0.23839046438142475, | |
| "grad_norm": 0.06348939893307848, | |
| "learning_rate": 1.7839084399272317e-05, | |
| "loss": 0.4670473575592041, | |
| "memory(GiB)": 76.61, | |
| "step": 745, | |
| "token_acc": 0.8652410477034038, | |
| "train_speed(iter/s)": 0.029638 | |
| }, | |
| { | |
| "epoch": 0.23999040038398464, | |
| "grad_norm": 0.07434587030241245, | |
| "learning_rate": 1.780673851171728e-05, | |
| "loss": 0.47047910690307615, | |
| "memory(GiB)": 76.61, | |
| "step": 750, | |
| "token_acc": 0.88801504530689, | |
| "train_speed(iter/s)": 0.029638 | |
| }, | |
| { | |
| "epoch": 0.23999040038398464, | |
| "eval_loss": 0.6746003031730652, | |
| "eval_runtime": 113.2223, | |
| "eval_samples_per_second": 177.421, | |
| "eval_steps_per_second": 0.892, | |
| "eval_token_acc": 0.8668385651547512, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.24159033638654454, | |
| "grad_norm": 0.06732795706859432, | |
| "learning_rate": 1.777418212147079e-05, | |
| "loss": 0.46190509796142576, | |
| "memory(GiB)": 76.61, | |
| "step": 755, | |
| "token_acc": 0.8881346728210697, | |
| "train_speed(iter/s)": 0.029543 | |
| }, | |
| { | |
| "epoch": 0.24319027238910443, | |
| "grad_norm": 0.06836940989947664, | |
| "learning_rate": 1.7741416106390828e-05, | |
| "loss": 0.46283302307128904, | |
| "memory(GiB)": 76.61, | |
| "step": 760, | |
| "token_acc": 0.8831443688586545, | |
| "train_speed(iter/s)": 0.029566 | |
| }, | |
| { | |
| "epoch": 0.24479020839166432, | |
| "grad_norm": 0.07072489516219096, | |
| "learning_rate": 1.7708441349987753e-05, | |
| "loss": 0.4619740962982178, | |
| "memory(GiB)": 76.61, | |
| "step": 765, | |
| "token_acc": 0.8610668789808917, | |
| "train_speed(iter/s)": 0.0296 | |
| }, | |
| { | |
| "epoch": 0.24639014439422424, | |
| "grad_norm": 0.07152232857362027, | |
| "learning_rate": 1.767525874140048e-05, | |
| "loss": 0.46694121360778806, | |
| "memory(GiB)": 76.61, | |
| "step": 770, | |
| "token_acc": 0.8397869022869023, | |
| "train_speed(iter/s)": 0.029606 | |
| }, | |
| { | |
| "epoch": 0.24799008039678413, | |
| "grad_norm": 0.059354056163304685, | |
| "learning_rate": 1.7641869175372493e-05, | |
| "loss": 0.4596868991851807, | |
| "memory(GiB)": 76.61, | |
| "step": 775, | |
| "token_acc": 0.8582827406764961, | |
| "train_speed(iter/s)": 0.029599 | |
| }, | |
| { | |
| "epoch": 0.24959001639934403, | |
| "grad_norm": 0.0629690289705531, | |
| "learning_rate": 1.7608273552227723e-05, | |
| "loss": 0.4583168029785156, | |
| "memory(GiB)": 76.61, | |
| "step": 780, | |
| "token_acc": 0.8841532106646639, | |
| "train_speed(iter/s)": 0.029639 | |
| }, | |
| { | |
| "epoch": 0.25118995240190395, | |
| "grad_norm": 0.05810355160479093, | |
| "learning_rate": 1.7574472777846276e-05, | |
| "loss": 0.47337069511413576, | |
| "memory(GiB)": 76.61, | |
| "step": 785, | |
| "token_acc": 0.8676557863501484, | |
| "train_speed(iter/s)": 0.029632 | |
| }, | |
| { | |
| "epoch": 0.2527898884044638, | |
| "grad_norm": 0.05365185572887828, | |
| "learning_rate": 1.7540467763639994e-05, | |
| "loss": 0.46567063331604003, | |
| "memory(GiB)": 76.61, | |
| "step": 790, | |
| "token_acc": 0.8745288099084545, | |
| "train_speed(iter/s)": 0.029629 | |
| }, | |
| { | |
| "epoch": 0.25438982440702373, | |
| "grad_norm": 0.054672322658953366, | |
| "learning_rate": 1.7506259426527903e-05, | |
| "loss": 0.47023472785949705, | |
| "memory(GiB)": 76.61, | |
| "step": 795, | |
| "token_acc": 0.874407844001322, | |
| "train_speed(iter/s)": 0.02965 | |
| }, | |
| { | |
| "epoch": 0.2559897604095836, | |
| "grad_norm": 0.057060955079149434, | |
| "learning_rate": 1.7471848688911465e-05, | |
| "loss": 0.4684537410736084, | |
| "memory(GiB)": 76.61, | |
| "step": 800, | |
| "token_acc": 0.8839382448537378, | |
| "train_speed(iter/s)": 0.029634 | |
| }, | |
| { | |
| "epoch": 0.2575896964121435, | |
| "grad_norm": 0.06051290772323595, | |
| "learning_rate": 1.7437236478649718e-05, | |
| "loss": 0.46199979782104494, | |
| "memory(GiB)": 76.61, | |
| "step": 805, | |
| "token_acc": 0.8673650919153983, | |
| "train_speed(iter/s)": 0.02966 | |
| }, | |
| { | |
| "epoch": 0.2591896324147034, | |
| "grad_norm": 0.0643397562387603, | |
| "learning_rate": 1.7402423729034252e-05, | |
| "loss": 0.4548381805419922, | |
| "memory(GiB)": 76.61, | |
| "step": 810, | |
| "token_acc": 0.83125, | |
| "train_speed(iter/s)": 0.029652 | |
| }, | |
| { | |
| "epoch": 0.2607895684172633, | |
| "grad_norm": 0.065624934571794, | |
| "learning_rate": 1.736741137876405e-05, | |
| "loss": 0.46353764533996583, | |
| "memory(GiB)": 76.61, | |
| "step": 815, | |
| "token_acc": 0.8907202528787537, | |
| "train_speed(iter/s)": 0.029628 | |
| }, | |
| { | |
| "epoch": 0.2623895044198232, | |
| "grad_norm": 0.053961693017135055, | |
| "learning_rate": 1.7332200371920173e-05, | |
| "loss": 0.46685361862182617, | |
| "memory(GiB)": 76.61, | |
| "step": 820, | |
| "token_acc": 0.8522188711762172, | |
| "train_speed(iter/s)": 0.029672 | |
| }, | |
| { | |
| "epoch": 0.2639894404223831, | |
| "grad_norm": 0.054388550053431586, | |
| "learning_rate": 1.72967916579403e-05, | |
| "loss": 0.46024084091186523, | |
| "memory(GiB)": 76.61, | |
| "step": 825, | |
| "token_acc": 0.8684630384683567, | |
| "train_speed(iter/s)": 0.02966 | |
| }, | |
| { | |
| "epoch": 0.265589376424943, | |
| "grad_norm": 0.0583019332597641, | |
| "learning_rate": 1.7261186191593135e-05, | |
| "loss": 0.47214059829711913, | |
| "memory(GiB)": 76.61, | |
| "step": 830, | |
| "token_acc": 0.8717123935666982, | |
| "train_speed(iter/s)": 0.029645 | |
| }, | |
| { | |
| "epoch": 0.2671893124275029, | |
| "grad_norm": 0.06004272220759217, | |
| "learning_rate": 1.7225384932952655e-05, | |
| "loss": 0.4626835823059082, | |
| "memory(GiB)": 76.61, | |
| "step": 835, | |
| "token_acc": 0.8737211788059246, | |
| "train_speed(iter/s)": 0.02967 | |
| }, | |
| { | |
| "epoch": 0.2687892484300628, | |
| "grad_norm": 0.05611993161069816, | |
| "learning_rate": 1.7189388847372227e-05, | |
| "loss": 0.46799750328063966, | |
| "memory(GiB)": 76.61, | |
| "step": 840, | |
| "token_acc": 0.8781684382665577, | |
| "train_speed(iter/s)": 0.029642 | |
| }, | |
| { | |
| "epoch": 0.2703891844326227, | |
| "grad_norm": 0.06345947319153013, | |
| "learning_rate": 1.715319890545857e-05, | |
| "loss": 0.4568619728088379, | |
| "memory(GiB)": 76.61, | |
| "step": 845, | |
| "token_acc": 0.860916969527537, | |
| "train_speed(iter/s)": 0.029655 | |
| }, | |
| { | |
| "epoch": 0.2719891204351826, | |
| "grad_norm": 0.0592531603954309, | |
| "learning_rate": 1.7116816083045603e-05, | |
| "loss": 0.46942729949951173, | |
| "memory(GiB)": 76.61, | |
| "step": 850, | |
| "token_acc": 0.8726317245194303, | |
| "train_speed(iter/s)": 0.029655 | |
| }, | |
| { | |
| "epoch": 0.2735890564377425, | |
| "grad_norm": 0.05711267065318382, | |
| "learning_rate": 1.7080241361168108e-05, | |
| "loss": 0.45801239013671874, | |
| "memory(GiB)": 76.61, | |
| "step": 855, | |
| "token_acc": 0.8834167608590344, | |
| "train_speed(iter/s)": 0.02963 | |
| }, | |
| { | |
| "epoch": 0.27518899244030237, | |
| "grad_norm": 0.05715792257951623, | |
| "learning_rate": 1.704347572603529e-05, | |
| "loss": 0.4675910472869873, | |
| "memory(GiB)": 76.61, | |
| "step": 860, | |
| "token_acc": 0.8361073624231519, | |
| "train_speed(iter/s)": 0.029659 | |
| }, | |
| { | |
| "epoch": 0.2767889284428623, | |
| "grad_norm": 0.056617536923221766, | |
| "learning_rate": 1.700652016900419e-05, | |
| "loss": 0.467483377456665, | |
| "memory(GiB)": 76.61, | |
| "step": 865, | |
| "token_acc": 0.8753590807532716, | |
| "train_speed(iter/s)": 0.029639 | |
| }, | |
| { | |
| "epoch": 0.27838886444542216, | |
| "grad_norm": 0.060433939578350394, | |
| "learning_rate": 1.696937568655294e-05, | |
| "loss": 0.46129570007324217, | |
| "memory(GiB)": 76.61, | |
| "step": 870, | |
| "token_acc": 0.8700755748512623, | |
| "train_speed(iter/s)": 0.029622 | |
| }, | |
| { | |
| "epoch": 0.2799888004479821, | |
| "grad_norm": 0.06826391103956585, | |
| "learning_rate": 1.6932043280253892e-05, | |
| "loss": 0.47449960708618166, | |
| "memory(GiB)": 76.61, | |
| "step": 875, | |
| "token_acc": 0.8767408356010885, | |
| "train_speed(iter/s)": 0.02965 | |
| }, | |
| { | |
| "epoch": 0.281588736450542, | |
| "grad_norm": 0.060978189753072065, | |
| "learning_rate": 1.689452395674664e-05, | |
| "loss": 0.464243745803833, | |
| "memory(GiB)": 76.61, | |
| "step": 880, | |
| "token_acc": 0.8622170179547228, | |
| "train_speed(iter/s)": 0.029624 | |
| }, | |
| { | |
| "epoch": 0.28318867245310186, | |
| "grad_norm": 0.0760276206328267, | |
| "learning_rate": 1.6856818727710847e-05, | |
| "loss": 0.4566212177276611, | |
| "memory(GiB)": 76.61, | |
| "step": 885, | |
| "token_acc": 0.8465499485066942, | |
| "train_speed(iter/s)": 0.029618 | |
| }, | |
| { | |
| "epoch": 0.2847886084556618, | |
| "grad_norm": 0.05693121191664627, | |
| "learning_rate": 1.6818928609838967e-05, | |
| "loss": 0.46042599678039553, | |
| "memory(GiB)": 76.61, | |
| "step": 890, | |
| "token_acc": 0.8798391728891441, | |
| "train_speed(iter/s)": 0.029627 | |
| }, | |
| { | |
| "epoch": 0.28638854445822165, | |
| "grad_norm": 0.05744826995499506, | |
| "learning_rate": 1.678085462480885e-05, | |
| "loss": 0.4604465961456299, | |
| "memory(GiB)": 76.61, | |
| "step": 895, | |
| "token_acc": 0.8780676542118063, | |
| "train_speed(iter/s)": 0.029599 | |
| }, | |
| { | |
| "epoch": 0.28798848046078157, | |
| "grad_norm": 0.06271464886952488, | |
| "learning_rate": 1.6742597799256182e-05, | |
| "loss": 0.46231966018676757, | |
| "memory(GiB)": 76.61, | |
| "step": 900, | |
| "token_acc": 0.8866765515780555, | |
| "train_speed(iter/s)": 0.029611 | |
| }, | |
| { | |
| "epoch": 0.2895884164633415, | |
| "grad_norm": 0.06044356676681803, | |
| "learning_rate": 1.6704159164746797e-05, | |
| "loss": 0.47655544281005857, | |
| "memory(GiB)": 76.61, | |
| "step": 905, | |
| "token_acc": 0.8872944211544663, | |
| "train_speed(iter/s)": 0.029601 | |
| }, | |
| { | |
| "epoch": 0.29118835246590136, | |
| "grad_norm": 0.05103569816400521, | |
| "learning_rate": 1.6665539757748866e-05, | |
| "loss": 0.4603917121887207, | |
| "memory(GiB)": 76.61, | |
| "step": 910, | |
| "token_acc": 0.8611705475141599, | |
| "train_speed(iter/s)": 0.029574 | |
| }, | |
| { | |
| "epoch": 0.2927882884684613, | |
| "grad_norm": 0.055811472748585486, | |
| "learning_rate": 1.6626740619604967e-05, | |
| "loss": 0.46213679313659667, | |
| "memory(GiB)": 76.61, | |
| "step": 915, | |
| "token_acc": 0.8148507643775783, | |
| "train_speed(iter/s)": 0.029594 | |
| }, | |
| { | |
| "epoch": 0.29438822447102114, | |
| "grad_norm": 0.05463929857953068, | |
| "learning_rate": 1.658776279650397e-05, | |
| "loss": 0.4658839702606201, | |
| "memory(GiB)": 76.61, | |
| "step": 920, | |
| "token_acc": 0.8766637089618456, | |
| "train_speed(iter/s)": 0.029577 | |
| }, | |
| { | |
| "epoch": 0.29598816047358106, | |
| "grad_norm": 0.06343067949686905, | |
| "learning_rate": 1.6548607339452853e-05, | |
| "loss": 0.46423888206481934, | |
| "memory(GiB)": 76.61, | |
| "step": 925, | |
| "token_acc": 0.8785782119115453, | |
| "train_speed(iter/s)": 0.029564 | |
| }, | |
| { | |
| "epoch": 0.29758809647614093, | |
| "grad_norm": 0.052431934937864355, | |
| "learning_rate": 1.6509275304248366e-05, | |
| "loss": 0.46324734687805175, | |
| "memory(GiB)": 76.61, | |
| "step": 930, | |
| "token_acc": 0.8571011956838729, | |
| "train_speed(iter/s)": 0.02958 | |
| }, | |
| { | |
| "epoch": 0.29918803247870085, | |
| "grad_norm": 0.059009943510604755, | |
| "learning_rate": 1.6469767751448538e-05, | |
| "loss": 0.46290836334228513, | |
| "memory(GiB)": 76.61, | |
| "step": 935, | |
| "token_acc": 0.8388616290480864, | |
| "train_speed(iter/s)": 0.029556 | |
| }, | |
| { | |
| "epoch": 0.30078796848126077, | |
| "grad_norm": 0.05160057372757322, | |
| "learning_rate": 1.6430085746344107e-05, | |
| "loss": 0.45898871421813964, | |
| "memory(GiB)": 76.61, | |
| "step": 940, | |
| "token_acc": 0.8690580344123651, | |
| "train_speed(iter/s)": 0.029556 | |
| }, | |
| { | |
| "epoch": 0.30238790448382064, | |
| "grad_norm": 0.05612231994140208, | |
| "learning_rate": 1.639023035892978e-05, | |
| "loss": 0.4546724796295166, | |
| "memory(GiB)": 76.61, | |
| "step": 945, | |
| "token_acc": 0.876509544215037, | |
| "train_speed(iter/s)": 0.02956 | |
| }, | |
| { | |
| "epoch": 0.30398784048638056, | |
| "grad_norm": 0.06733149115024578, | |
| "learning_rate": 1.6350202663875385e-05, | |
| "loss": 0.4598522663116455, | |
| "memory(GiB)": 76.61, | |
| "step": 950, | |
| "token_acc": 0.8623452294246177, | |
| "train_speed(iter/s)": 0.029531 | |
| }, | |
| { | |
| "epoch": 0.3055877764889404, | |
| "grad_norm": 0.05450569676621943, | |
| "learning_rate": 1.6310003740496887e-05, | |
| "loss": 0.4602477550506592, | |
| "memory(GiB)": 76.61, | |
| "step": 955, | |
| "token_acc": 0.8647700701480904, | |
| "train_speed(iter/s)": 0.029548 | |
| }, | |
| { | |
| "epoch": 0.30718771249150034, | |
| "grad_norm": 0.06736921151917717, | |
| "learning_rate": 1.6269634672727296e-05, | |
| "loss": 0.4589672565460205, | |
| "memory(GiB)": 76.61, | |
| "step": 960, | |
| "token_acc": 0.877502001601281, | |
| "train_speed(iter/s)": 0.029536 | |
| }, | |
| { | |
| "epoch": 0.30878764849406026, | |
| "grad_norm": 0.06166660436042404, | |
| "learning_rate": 1.6229096549087434e-05, | |
| "loss": 0.4601268291473389, | |
| "memory(GiB)": 76.61, | |
| "step": 965, | |
| "token_acc": 0.8723534201954397, | |
| "train_speed(iter/s)": 0.029518 | |
| }, | |
| { | |
| "epoch": 0.31038758449662013, | |
| "grad_norm": 0.055128746386822226, | |
| "learning_rate": 1.618839046265658e-05, | |
| "loss": 0.4666788101196289, | |
| "memory(GiB)": 76.61, | |
| "step": 970, | |
| "token_acc": 0.8550563360689943, | |
| "train_speed(iter/s)": 0.029541 | |
| }, | |
| { | |
| "epoch": 0.31198752049918005, | |
| "grad_norm": 0.056867326711030626, | |
| "learning_rate": 1.614751751104301e-05, | |
| "loss": 0.4646125793457031, | |
| "memory(GiB)": 76.61, | |
| "step": 975, | |
| "token_acc": 0.8651571964234208, | |
| "train_speed(iter/s)": 0.029524 | |
| }, | |
| { | |
| "epoch": 0.3135874565017399, | |
| "grad_norm": 0.05501107287069041, | |
| "learning_rate": 1.6106478796354382e-05, | |
| "loss": 0.4588280200958252, | |
| "memory(GiB)": 76.61, | |
| "step": 980, | |
| "token_acc": 0.8767766331985918, | |
| "train_speed(iter/s)": 0.029517 | |
| }, | |
| { | |
| "epoch": 0.31518739250429983, | |
| "grad_norm": 0.08099201898186387, | |
| "learning_rate": 1.6065275425168034e-05, | |
| "loss": 0.4589373111724854, | |
| "memory(GiB)": 76.61, | |
| "step": 985, | |
| "token_acc": 0.8917890157694399, | |
| "train_speed(iter/s)": 0.029526 | |
| }, | |
| { | |
| "epoch": 0.3167873285068597, | |
| "grad_norm": 0.0522899382710734, | |
| "learning_rate": 1.602390850850113e-05, | |
| "loss": 0.46761279106140136, | |
| "memory(GiB)": 76.61, | |
| "step": 990, | |
| "token_acc": 0.8461229409401366, | |
| "train_speed(iter/s)": 0.029505 | |
| }, | |
| { | |
| "epoch": 0.3183872645094196, | |
| "grad_norm": 0.05838858698011934, | |
| "learning_rate": 1.5982379161780722e-05, | |
| "loss": 0.44941887855529783, | |
| "memory(GiB)": 76.61, | |
| "step": 995, | |
| "token_acc": 0.8547228871294421, | |
| "train_speed(iter/s)": 0.029511 | |
| }, | |
| { | |
| "epoch": 0.31998720051197954, | |
| "grad_norm": 0.054930484370324516, | |
| "learning_rate": 1.5940688504813664e-05, | |
| "loss": 0.4591392517089844, | |
| "memory(GiB)": 76.61, | |
| "step": 1000, | |
| "token_acc": 0.8995555555555556, | |
| "train_speed(iter/s)": 0.029505 | |
| }, | |
| { | |
| "epoch": 0.31998720051197954, | |
| "eval_loss": 0.671963095664978, | |
| "eval_runtime": 110.8694, | |
| "eval_samples_per_second": 181.186, | |
| "eval_steps_per_second": 0.911, | |
| "eval_token_acc": 0.8676077802864524, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.3215871365145394, | |
| "grad_norm": 0.0578985798516978, | |
| "learning_rate": 1.5898837661756405e-05, | |
| "loss": 0.46222972869873047, | |
| "memory(GiB)": 76.61, | |
| "step": 1005, | |
| "token_acc": 0.8840002569208042, | |
| "train_speed(iter/s)": 0.029425 | |
| }, | |
| { | |
| "epoch": 0.3231870725170993, | |
| "grad_norm": 0.05872050053297838, | |
| "learning_rate": 1.5856827761084698e-05, | |
| "loss": 0.45543718338012695, | |
| "memory(GiB)": 76.61, | |
| "step": 1010, | |
| "token_acc": 0.8753668220265838, | |
| "train_speed(iter/s)": 0.02945 | |
| }, | |
| { | |
| "epoch": 0.3247870085196592, | |
| "grad_norm": 0.05268695066428434, | |
| "learning_rate": 1.5814659935563165e-05, | |
| "loss": 0.46614727973937986, | |
| "memory(GiB)": 76.61, | |
| "step": 1015, | |
| "token_acc": 0.8792250035355678, | |
| "train_speed(iter/s)": 0.029474 | |
| }, | |
| { | |
| "epoch": 0.3263869445222191, | |
| "grad_norm": 0.059454673806441594, | |
| "learning_rate": 1.577233532221474e-05, | |
| "loss": 0.45902605056762696, | |
| "memory(GiB)": 76.61, | |
| "step": 1020, | |
| "token_acc": 0.86709886547812, | |
| "train_speed(iter/s)": 0.029475 | |
| }, | |
| { | |
| "epoch": 0.32798688052477903, | |
| "grad_norm": 0.053728974295076275, | |
| "learning_rate": 1.5729855062290024e-05, | |
| "loss": 0.46491541862487795, | |
| "memory(GiB)": 76.61, | |
| "step": 1025, | |
| "token_acc": 0.8708870261478794, | |
| "train_speed(iter/s)": 0.029469 | |
| }, | |
| { | |
| "epoch": 0.3295868165273389, | |
| "grad_norm": 0.07030309576814114, | |
| "learning_rate": 1.568722030123651e-05, | |
| "loss": 0.453840970993042, | |
| "memory(GiB)": 76.61, | |
| "step": 1030, | |
| "token_acc": 0.8568111455108359, | |
| "train_speed(iter/s)": 0.029496 | |
| }, | |
| { | |
| "epoch": 0.3311867525298988, | |
| "grad_norm": 0.07385415365022158, | |
| "learning_rate": 1.5644432188667695e-05, | |
| "loss": 0.45582828521728513, | |
| "memory(GiB)": 76.61, | |
| "step": 1035, | |
| "token_acc": 0.8800162140251317, | |
| "train_speed(iter/s)": 0.029488 | |
| }, | |
| { | |
| "epoch": 0.3327866885324587, | |
| "grad_norm": 0.05407863995123405, | |
| "learning_rate": 1.5601491878332077e-05, | |
| "loss": 0.4665637969970703, | |
| "memory(GiB)": 76.61, | |
| "step": 1040, | |
| "token_acc": 0.8628481345244351, | |
| "train_speed(iter/s)": 0.029487 | |
| }, | |
| { | |
| "epoch": 0.3343866245350186, | |
| "grad_norm": 0.05879461372080454, | |
| "learning_rate": 1.5558400528082057e-05, | |
| "loss": 0.4657593250274658, | |
| "memory(GiB)": 76.61, | |
| "step": 1045, | |
| "token_acc": 0.879185119574845, | |
| "train_speed(iter/s)": 0.02951 | |
| }, | |
| { | |
| "epoch": 0.33598656053757847, | |
| "grad_norm": 0.06618244368029796, | |
| "learning_rate": 1.551515929984271e-05, | |
| "loss": 0.45760574340820315, | |
| "memory(GiB)": 76.61, | |
| "step": 1050, | |
| "token_acc": 0.8899380348185305, | |
| "train_speed(iter/s)": 0.029502 | |
| }, | |
| { | |
| "epoch": 0.3375864965401384, | |
| "grad_norm": 0.06388796415692906, | |
| "learning_rate": 1.547176935958044e-05, | |
| "loss": 0.46065597534179686, | |
| "memory(GiB)": 76.61, | |
| "step": 1055, | |
| "token_acc": 0.8536853685368537, | |
| "train_speed(iter/s)": 0.029524 | |
| }, | |
| { | |
| "epoch": 0.3391864325426983, | |
| "grad_norm": 0.05811152365312673, | |
| "learning_rate": 1.5428231877271584e-05, | |
| "loss": 0.46312780380249025, | |
| "memory(GiB)": 76.61, | |
| "step": 1060, | |
| "token_acc": 0.8520375161707633, | |
| "train_speed(iter/s)": 0.029515 | |
| }, | |
| { | |
| "epoch": 0.3407863685452582, | |
| "grad_norm": 0.05545936328508829, | |
| "learning_rate": 1.538454802687081e-05, | |
| "loss": 0.4615220546722412, | |
| "memory(GiB)": 76.61, | |
| "step": 1065, | |
| "token_acc": 0.8744265080713679, | |
| "train_speed(iter/s)": 0.029504 | |
| }, | |
| { | |
| "epoch": 0.3423863045478181, | |
| "grad_norm": 0.05964362984731802, | |
| "learning_rate": 1.5340718986279505e-05, | |
| "loss": 0.46706466674804686, | |
| "memory(GiB)": 76.61, | |
| "step": 1070, | |
| "token_acc": 0.8592233009708737, | |
| "train_speed(iter/s)": 0.029536 | |
| }, | |
| { | |
| "epoch": 0.34398624055037796, | |
| "grad_norm": 0.05356886450328198, | |
| "learning_rate": 1.529674593731399e-05, | |
| "loss": 0.45301499366760256, | |
| "memory(GiB)": 76.61, | |
| "step": 1075, | |
| "token_acc": 0.8575192096597146, | |
| "train_speed(iter/s)": 0.029526 | |
| }, | |
| { | |
| "epoch": 0.3455861765529379, | |
| "grad_norm": 0.05995962073425321, | |
| "learning_rate": 1.5252630065673662e-05, | |
| "loss": 0.46819314956665037, | |
| "memory(GiB)": 76.61, | |
| "step": 1080, | |
| "token_acc": 0.8875031814711123, | |
| "train_speed(iter/s)": 0.029518 | |
| }, | |
| { | |
| "epoch": 0.3471861125554978, | |
| "grad_norm": 0.05389432634852101, | |
| "learning_rate": 1.5208372560909031e-05, | |
| "loss": 0.46298394203186033, | |
| "memory(GiB)": 76.61, | |
| "step": 1085, | |
| "token_acc": 0.8872426699937617, | |
| "train_speed(iter/s)": 0.029543 | |
| }, | |
| { | |
| "epoch": 0.34878604855805767, | |
| "grad_norm": 0.06642390255342462, | |
| "learning_rate": 1.5163974616389621e-05, | |
| "loss": 0.45978522300720215, | |
| "memory(GiB)": 76.61, | |
| "step": 1090, | |
| "token_acc": 0.8246258860593332, | |
| "train_speed(iter/s)": 0.029525 | |
| }, | |
| { | |
| "epoch": 0.3503859845606176, | |
| "grad_norm": 0.06115184110491886, | |
| "learning_rate": 1.5119437429271813e-05, | |
| "loss": 0.4637304782867432, | |
| "memory(GiB)": 76.61, | |
| "step": 1095, | |
| "token_acc": 0.8666082895504962, | |
| "train_speed(iter/s)": 0.029534 | |
| }, | |
| { | |
| "epoch": 0.35198592056317746, | |
| "grad_norm": 0.060865150660591956, | |
| "learning_rate": 1.5074762200466557e-05, | |
| "loss": 0.4542848587036133, | |
| "memory(GiB)": 76.61, | |
| "step": 1100, | |
| "token_acc": 0.8913602663035255, | |
| "train_speed(iter/s)": 0.029544 | |
| }, | |
| { | |
| "epoch": 0.3535858565657374, | |
| "grad_norm": 0.057666943430007674, | |
| "learning_rate": 1.5029950134606991e-05, | |
| "loss": 0.4574248790740967, | |
| "memory(GiB)": 76.61, | |
| "step": 1105, | |
| "token_acc": 0.8634470336597996, | |
| "train_speed(iter/s)": 0.029524 | |
| }, | |
| { | |
| "epoch": 0.35518579256829724, | |
| "grad_norm": 0.054034554153381265, | |
| "learning_rate": 1.4985002440015959e-05, | |
| "loss": 0.4520272254943848, | |
| "memory(GiB)": 76.61, | |
| "step": 1110, | |
| "token_acc": 0.8674898358680921, | |
| "train_speed(iter/s)": 0.029551 | |
| }, | |
| { | |
| "epoch": 0.35678572857085716, | |
| "grad_norm": 0.06416854479766453, | |
| "learning_rate": 1.4939920328673422e-05, | |
| "loss": 0.4668846130371094, | |
| "memory(GiB)": 76.61, | |
| "step": 1115, | |
| "token_acc": 0.9170854271356784, | |
| "train_speed(iter/s)": 0.029541 | |
| }, | |
| { | |
| "epoch": 0.3583856645734171, | |
| "grad_norm": 0.05775941336987237, | |
| "learning_rate": 1.4894705016183803e-05, | |
| "loss": 0.4518620491027832, | |
| "memory(GiB)": 76.61, | |
| "step": 1120, | |
| "token_acc": 0.8672782874617737, | |
| "train_speed(iter/s)": 0.029531 | |
| }, | |
| { | |
| "epoch": 0.35998560057597695, | |
| "grad_norm": 0.0625175589581686, | |
| "learning_rate": 1.4849357721743169e-05, | |
| "loss": 0.4566941738128662, | |
| "memory(GiB)": 76.61, | |
| "step": 1125, | |
| "token_acc": 0.8505491793163026, | |
| "train_speed(iter/s)": 0.029557 | |
| }, | |
| { | |
| "epoch": 0.36158553657853687, | |
| "grad_norm": 0.05911529293553411, | |
| "learning_rate": 1.4803879668106393e-05, | |
| "loss": 0.4640664577484131, | |
| "memory(GiB)": 76.61, | |
| "step": 1130, | |
| "token_acc": 0.8772325625117503, | |
| "train_speed(iter/s)": 0.029544 | |
| }, | |
| { | |
| "epoch": 0.36318547258109674, | |
| "grad_norm": 0.06483783687935218, | |
| "learning_rate": 1.4758272081554168e-05, | |
| "loss": 0.45419878959655763, | |
| "memory(GiB)": 76.61, | |
| "step": 1135, | |
| "token_acc": 0.8594914930223667, | |
| "train_speed(iter/s)": 0.029539 | |
| }, | |
| { | |
| "epoch": 0.36478540858365666, | |
| "grad_norm": 0.06032730304497941, | |
| "learning_rate": 1.4712536191859934e-05, | |
| "loss": 0.45779004096984866, | |
| "memory(GiB)": 76.61, | |
| "step": 1140, | |
| "token_acc": 0.8938053097345132, | |
| "train_speed(iter/s)": 0.029564 | |
| }, | |
| { | |
| "epoch": 0.3663853445862166, | |
| "grad_norm": 0.0637380940226065, | |
| "learning_rate": 1.4666673232256738e-05, | |
| "loss": 0.46385722160339354, | |
| "memory(GiB)": 76.61, | |
| "step": 1145, | |
| "token_acc": 0.8621830209481808, | |
| "train_speed(iter/s)": 0.029544 | |
| }, | |
| { | |
| "epoch": 0.36798528058877644, | |
| "grad_norm": 0.057006770373085346, | |
| "learning_rate": 1.4620684439403962e-05, | |
| "loss": 0.4613553524017334, | |
| "memory(GiB)": 76.61, | |
| "step": 1150, | |
| "token_acc": 0.8831837819873712, | |
| "train_speed(iter/s)": 0.029558 | |
| }, | |
| { | |
| "epoch": 0.36958521659133636, | |
| "grad_norm": 0.057569299635009126, | |
| "learning_rate": 1.4574571053353987e-05, | |
| "loss": 0.4598341464996338, | |
| "memory(GiB)": 76.61, | |
| "step": 1155, | |
| "token_acc": 0.8825154371140721, | |
| "train_speed(iter/s)": 0.029557 | |
| }, | |
| { | |
| "epoch": 0.37118515259389623, | |
| "grad_norm": 0.06747695219063263, | |
| "learning_rate": 1.452833431751875e-05, | |
| "loss": 0.4570640563964844, | |
| "memory(GiB)": 76.61, | |
| "step": 1160, | |
| "token_acc": 0.8726823238566132, | |
| "train_speed(iter/s)": 0.029543 | |
| }, | |
| { | |
| "epoch": 0.37278508859645615, | |
| "grad_norm": 0.05405367649749466, | |
| "learning_rate": 1.448197547863622e-05, | |
| "loss": 0.4516812801361084, | |
| "memory(GiB)": 76.61, | |
| "step": 1165, | |
| "token_acc": 0.8704696273608984, | |
| "train_speed(iter/s)": 0.029568 | |
| }, | |
| { | |
| "epoch": 0.374385024599016, | |
| "grad_norm": 0.06041157710672601, | |
| "learning_rate": 1.4435495786736796e-05, | |
| "loss": 0.465837287902832, | |
| "memory(GiB)": 76.61, | |
| "step": 1170, | |
| "token_acc": 0.8673412029229904, | |
| "train_speed(iter/s)": 0.029554 | |
| }, | |
| { | |
| "epoch": 0.37598496060157593, | |
| "grad_norm": 0.05229585247228306, | |
| "learning_rate": 1.438889649510956e-05, | |
| "loss": 0.4427653789520264, | |
| "memory(GiB)": 76.61, | |
| "step": 1175, | |
| "token_acc": 0.8558139534883721, | |
| "train_speed(iter/s)": 0.02954 | |
| }, | |
| { | |
| "epoch": 0.37758489660413586, | |
| "grad_norm": 0.0547875272797444, | |
| "learning_rate": 1.4342178860268523e-05, | |
| "loss": 0.45673260688781736, | |
| "memory(GiB)": 76.61, | |
| "step": 1180, | |
| "token_acc": 0.880563238622077, | |
| "train_speed(iter/s)": 0.029563 | |
| }, | |
| { | |
| "epoch": 0.3791848326066957, | |
| "grad_norm": 0.0565328006493161, | |
| "learning_rate": 1.4295344141918734e-05, | |
| "loss": 0.46208748817443845, | |
| "memory(GiB)": 76.61, | |
| "step": 1185, | |
| "token_acc": 0.8671328671328671, | |
| "train_speed(iter/s)": 0.029544 | |
| }, | |
| { | |
| "epoch": 0.38078476860925564, | |
| "grad_norm": 0.062473905403265834, | |
| "learning_rate": 1.4248393602922299e-05, | |
| "loss": 0.46883163452148435, | |
| "memory(GiB)": 76.61, | |
| "step": 1190, | |
| "token_acc": 0.8412252145605209, | |
| "train_speed(iter/s)": 0.029548 | |
| }, | |
| { | |
| "epoch": 0.3823847046118155, | |
| "grad_norm": 0.05646151042315891, | |
| "learning_rate": 1.420132850926434e-05, | |
| "loss": 0.45732822418212893, | |
| "memory(GiB)": 76.61, | |
| "step": 1195, | |
| "token_acc": 0.8820655966503839, | |
| "train_speed(iter/s)": 0.02956 | |
| }, | |
| { | |
| "epoch": 0.3839846406143754, | |
| "grad_norm": 0.052981558367052706, | |
| "learning_rate": 1.4154150130018867e-05, | |
| "loss": 0.45579113960266116, | |
| "memory(GiB)": 76.61, | |
| "step": 1200, | |
| "token_acc": 0.8677085226240233, | |
| "train_speed(iter/s)": 0.029546 | |
| }, | |
| { | |
| "epoch": 0.38558457661693535, | |
| "grad_norm": 0.052315204322432474, | |
| "learning_rate": 1.4106859737314532e-05, | |
| "loss": 0.45348801612854006, | |
| "memory(GiB)": 76.61, | |
| "step": 1205, | |
| "token_acc": 0.8616187989556136, | |
| "train_speed(iter/s)": 0.029561 | |
| }, | |
| { | |
| "epoch": 0.3871845126194952, | |
| "grad_norm": 0.05319888084520812, | |
| "learning_rate": 1.4059458606300358e-05, | |
| "loss": 0.45279593467712403, | |
| "memory(GiB)": 76.61, | |
| "step": 1210, | |
| "token_acc": 0.86090645233311, | |
| "train_speed(iter/s)": 0.029565 | |
| }, | |
| { | |
| "epoch": 0.38878444862205513, | |
| "grad_norm": 0.054475973938428034, | |
| "learning_rate": 1.4011948015111334e-05, | |
| "loss": 0.4616706848144531, | |
| "memory(GiB)": 76.61, | |
| "step": 1215, | |
| "token_acc": 0.8390133684805121, | |
| "train_speed(iter/s)": 0.029549 | |
| }, | |
| { | |
| "epoch": 0.390384384624615, | |
| "grad_norm": 0.054891067059900926, | |
| "learning_rate": 1.396432924483396e-05, | |
| "loss": 0.4553243637084961, | |
| "memory(GiB)": 76.61, | |
| "step": 1220, | |
| "token_acc": 0.8715350793347353, | |
| "train_speed(iter/s)": 0.029571 | |
| }, | |
| { | |
| "epoch": 0.3919843206271749, | |
| "grad_norm": 0.06058246643434403, | |
| "learning_rate": 1.3916603579471705e-05, | |
| "loss": 0.47067904472351074, | |
| "memory(GiB)": 76.61, | |
| "step": 1225, | |
| "token_acc": 0.8662144337667232, | |
| "train_speed(iter/s)": 0.029556 | |
| }, | |
| { | |
| "epoch": 0.3935842566297348, | |
| "grad_norm": 0.05715510214651738, | |
| "learning_rate": 1.3868772305910376e-05, | |
| "loss": 0.46147928237915037, | |
| "memory(GiB)": 76.61, | |
| "step": 1230, | |
| "token_acc": 0.868918375552875, | |
| "train_speed(iter/s)": 0.029548 | |
| }, | |
| { | |
| "epoch": 0.3951841926322947, | |
| "grad_norm": 0.06593047910666934, | |
| "learning_rate": 1.3820836713883424e-05, | |
| "loss": 0.45935769081115724, | |
| "memory(GiB)": 76.61, | |
| "step": 1235, | |
| "token_acc": 0.8596291476903057, | |
| "train_speed(iter/s)": 0.02957 | |
| }, | |
| { | |
| "epoch": 0.3967841286348546, | |
| "grad_norm": 0.056071042953882384, | |
| "learning_rate": 1.3772798095937172e-05, | |
| "loss": 0.4495890140533447, | |
| "memory(GiB)": 76.61, | |
| "step": 1240, | |
| "token_acc": 0.8471917163476623, | |
| "train_speed(iter/s)": 0.029553 | |
| }, | |
| { | |
| "epoch": 0.3983840646374145, | |
| "grad_norm": 0.05810589720196263, | |
| "learning_rate": 1.3724657747395957e-05, | |
| "loss": 0.4619898319244385, | |
| "memory(GiB)": 76.61, | |
| "step": 1245, | |
| "token_acc": 0.8691186216037111, | |
| "train_speed(iter/s)": 0.029561 | |
| }, | |
| { | |
| "epoch": 0.3999840006399744, | |
| "grad_norm": 0.055604926632171425, | |
| "learning_rate": 1.3676416966327201e-05, | |
| "loss": 0.4587514400482178, | |
| "memory(GiB)": 76.61, | |
| "step": 1250, | |
| "token_acc": 0.8369355461211887, | |
| "train_speed(iter/s)": 0.029564 | |
| }, | |
| { | |
| "epoch": 0.3999840006399744, | |
| "eval_loss": 0.6690404415130615, | |
| "eval_runtime": 106.3444, | |
| "eval_samples_per_second": 188.896, | |
| "eval_steps_per_second": 0.95, | |
| "eval_token_acc": 0.8683678146748934, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.4015839366425343, | |
| "grad_norm": 0.04782987834900457, | |
| "learning_rate": 1.362807705350641e-05, | |
| "loss": 0.46315851211547854, | |
| "memory(GiB)": 76.61, | |
| "step": 1255, | |
| "token_acc": 0.8767961498796838, | |
| "train_speed(iter/s)": 0.029512 | |
| }, | |
| { | |
| "epoch": 0.4031838726450942, | |
| "grad_norm": 0.05995996443795485, | |
| "learning_rate": 1.3579639312382105e-05, | |
| "loss": 0.46349530220031737, | |
| "memory(GiB)": 76.61, | |
| "step": 1260, | |
| "token_acc": 0.8588617886178862, | |
| "train_speed(iter/s)": 0.029524 | |
| }, | |
| { | |
| "epoch": 0.4047838086476541, | |
| "grad_norm": 0.06488882353036057, | |
| "learning_rate": 1.3531105049040667e-05, | |
| "loss": 0.45726447105407714, | |
| "memory(GiB)": 76.61, | |
| "step": 1265, | |
| "token_acc": 0.8802249582003344, | |
| "train_speed(iter/s)": 0.029543 | |
| }, | |
| { | |
| "epoch": 0.406383744650214, | |
| "grad_norm": 0.05350128050935312, | |
| "learning_rate": 1.3482475572171132e-05, | |
| "loss": 0.4516806125640869, | |
| "memory(GiB)": 76.61, | |
| "step": 1270, | |
| "token_acc": 0.8560765550239234, | |
| "train_speed(iter/s)": 0.029549 | |
| }, | |
| { | |
| "epoch": 0.4079836806527739, | |
| "grad_norm": 0.05672697687392494, | |
| "learning_rate": 1.3433752193029888e-05, | |
| "loss": 0.46581568717956545, | |
| "memory(GiB)": 76.61, | |
| "step": 1275, | |
| "token_acc": 0.8881742738589211, | |
| "train_speed(iter/s)": 0.029547 | |
| }, | |
| { | |
| "epoch": 0.40958361665533377, | |
| "grad_norm": 0.0598115330947421, | |
| "learning_rate": 1.3384936225405326e-05, | |
| "loss": 0.46333680152893064, | |
| "memory(GiB)": 76.61, | |
| "step": 1280, | |
| "token_acc": 0.8608710985716804, | |
| "train_speed(iter/s)": 0.029573 | |
| }, | |
| { | |
| "epoch": 0.4111835526578937, | |
| "grad_norm": 0.05384417907735887, | |
| "learning_rate": 1.333602898558242e-05, | |
| "loss": 0.4611030578613281, | |
| "memory(GiB)": 76.61, | |
| "step": 1285, | |
| "token_acc": 0.8845689770746749, | |
| "train_speed(iter/s)": 0.029567 | |
| }, | |
| { | |
| "epoch": 0.41278348866045356, | |
| "grad_norm": 0.06043637267465684, | |
| "learning_rate": 1.3287031792307226e-05, | |
| "loss": 0.46013875007629396, | |
| "memory(GiB)": 76.61, | |
| "step": 1290, | |
| "token_acc": 0.870195210303884, | |
| "train_speed(iter/s)": 0.029565 | |
| }, | |
| { | |
| "epoch": 0.4143834246630135, | |
| "grad_norm": 0.06140603532631629, | |
| "learning_rate": 1.323794596675132e-05, | |
| "loss": 0.45681238174438477, | |
| "memory(GiB)": 76.61, | |
| "step": 1295, | |
| "token_acc": 0.8450012281994596, | |
| "train_speed(iter/s)": 0.029583 | |
| }, | |
| { | |
| "epoch": 0.4159833606655734, | |
| "grad_norm": 0.062077229851937275, | |
| "learning_rate": 1.318877283247619e-05, | |
| "loss": 0.4490199565887451, | |
| "memory(GiB)": 76.61, | |
| "step": 1300, | |
| "token_acc": 0.89259877573734, | |
| "train_speed(iter/s)": 0.029573 | |
| }, | |
| { | |
| "epoch": 0.41758329666813326, | |
| "grad_norm": 0.05216177276902916, | |
| "learning_rate": 1.3139513715397521e-05, | |
| "loss": 0.45108351707458494, | |
| "memory(GiB)": 76.61, | |
| "step": 1305, | |
| "token_acc": 0.8547701815372731, | |
| "train_speed(iter/s)": 0.029594 | |
| }, | |
| { | |
| "epoch": 0.4191832326706932, | |
| "grad_norm": 0.05738628087610287, | |
| "learning_rate": 1.3090169943749475e-05, | |
| "loss": 0.4558550834655762, | |
| "memory(GiB)": 76.61, | |
| "step": 1310, | |
| "token_acc": 0.840696686491079, | |
| "train_speed(iter/s)": 0.029587 | |
| }, | |
| { | |
| "epoch": 0.42078316867325305, | |
| "grad_norm": 0.05518036740697275, | |
| "learning_rate": 1.304074284804885e-05, | |
| "loss": 0.4631648063659668, | |
| "memory(GiB)": 76.61, | |
| "step": 1315, | |
| "token_acc": 0.8788111708941839, | |
| "train_speed(iter/s)": 0.029578 | |
| }, | |
| { | |
| "epoch": 0.42238310467581297, | |
| "grad_norm": 0.05902492258138098, | |
| "learning_rate": 1.2991233761059214e-05, | |
| "loss": 0.45921921730041504, | |
| "memory(GiB)": 76.61, | |
| "step": 1320, | |
| "token_acc": 0.866059646344682, | |
| "train_speed(iter/s)": 0.029604 | |
| }, | |
| { | |
| "epoch": 0.4239830406783729, | |
| "grad_norm": 0.059535437419073044, | |
| "learning_rate": 1.2941644017754964e-05, | |
| "loss": 0.46445517539978026, | |
| "memory(GiB)": 76.61, | |
| "step": 1325, | |
| "token_acc": 0.8831345826235094, | |
| "train_speed(iter/s)": 0.029591 | |
| }, | |
| { | |
| "epoch": 0.42558297668093276, | |
| "grad_norm": 0.04863893443696892, | |
| "learning_rate": 1.289197495528534e-05, | |
| "loss": 0.45836362838745115, | |
| "memory(GiB)": 76.61, | |
| "step": 1330, | |
| "token_acc": 0.8986429177268872, | |
| "train_speed(iter/s)": 0.029582 | |
| }, | |
| { | |
| "epoch": 0.4271829126834927, | |
| "grad_norm": 0.05945822860509985, | |
| "learning_rate": 1.284222791293836e-05, | |
| "loss": 0.45783252716064454, | |
| "memory(GiB)": 76.61, | |
| "step": 1335, | |
| "token_acc": 0.8500874125874126, | |
| "train_speed(iter/s)": 0.029596 | |
| }, | |
| { | |
| "epoch": 0.42878284868605254, | |
| "grad_norm": 0.05989616737178823, | |
| "learning_rate": 1.2792404232104699e-05, | |
| "loss": 0.45293269157409666, | |
| "memory(GiB)": 76.61, | |
| "step": 1340, | |
| "token_acc": 0.8637480798771121, | |
| "train_speed(iter/s)": 0.029584 | |
| }, | |
| { | |
| "epoch": 0.43038278468861246, | |
| "grad_norm": 0.0586629819404024, | |
| "learning_rate": 1.2742505256241543e-05, | |
| "loss": 0.45876450538635255, | |
| "memory(GiB)": 76.61, | |
| "step": 1345, | |
| "token_acc": 0.8296499119890475, | |
| "train_speed(iter/s)": 0.029588 | |
| }, | |
| { | |
| "epoch": 0.43198272069117233, | |
| "grad_norm": 0.052924904785980484, | |
| "learning_rate": 1.2692532330836346e-05, | |
| "loss": 0.45821080207824705, | |
| "memory(GiB)": 76.61, | |
| "step": 1350, | |
| "token_acc": 0.8636084374360025, | |
| "train_speed(iter/s)": 0.029594 | |
| }, | |
| { | |
| "epoch": 0.43358265669373225, | |
| "grad_norm": 0.059304249814977644, | |
| "learning_rate": 1.2642486803370553e-05, | |
| "loss": 0.45485148429870603, | |
| "memory(GiB)": 76.61, | |
| "step": 1355, | |
| "token_acc": 0.8686690223792697, | |
| "train_speed(iter/s)": 0.029579 | |
| }, | |
| { | |
| "epoch": 0.43518259269629217, | |
| "grad_norm": 0.06253442360689314, | |
| "learning_rate": 1.2592370023283268e-05, | |
| "loss": 0.45198469161987304, | |
| "memory(GiB)": 76.61, | |
| "step": 1360, | |
| "token_acc": 0.8737075332348597, | |
| "train_speed(iter/s)": 0.0296 | |
| }, | |
| { | |
| "epoch": 0.43678252869885204, | |
| "grad_norm": 0.05314091037792793, | |
| "learning_rate": 1.2542183341934873e-05, | |
| "loss": 0.4516898155212402, | |
| "memory(GiB)": 76.61, | |
| "step": 1365, | |
| "token_acc": 0.8714476021314387, | |
| "train_speed(iter/s)": 0.029596 | |
| }, | |
| { | |
| "epoch": 0.43838246470141196, | |
| "grad_norm": 0.06014404788689081, | |
| "learning_rate": 1.2491928112570568e-05, | |
| "loss": 0.45399184226989747, | |
| "memory(GiB)": 76.61, | |
| "step": 1370, | |
| "token_acc": 0.8657097288676237, | |
| "train_speed(iter/s)": 0.029583 | |
| }, | |
| { | |
| "epoch": 0.4399824007039718, | |
| "grad_norm": 0.05910144328100835, | |
| "learning_rate": 1.2441605690283915e-05, | |
| "loss": 0.4607128143310547, | |
| "memory(GiB)": 76.61, | |
| "step": 1375, | |
| "token_acc": 0.8990952307928232, | |
| "train_speed(iter/s)": 0.029603 | |
| }, | |
| { | |
| "epoch": 0.44158233670653174, | |
| "grad_norm": 0.059073628736854025, | |
| "learning_rate": 1.2391217431980273e-05, | |
| "loss": 0.4515543937683105, | |
| "memory(GiB)": 76.61, | |
| "step": 1380, | |
| "token_acc": 0.9016349860428021, | |
| "train_speed(iter/s)": 0.029591 | |
| }, | |
| { | |
| "epoch": 0.44318227270909166, | |
| "grad_norm": 0.058358968679540275, | |
| "learning_rate": 1.234076469634022e-05, | |
| "loss": 0.45762925148010253, | |
| "memory(GiB)": 76.61, | |
| "step": 1385, | |
| "token_acc": 0.8919261822376009, | |
| "train_speed(iter/s)": 0.029584 | |
| }, | |
| { | |
| "epoch": 0.4447822087116515, | |
| "grad_norm": 0.0672513399669503, | |
| "learning_rate": 1.2290248843782915e-05, | |
| "loss": 0.44803729057312014, | |
| "memory(GiB)": 76.61, | |
| "step": 1390, | |
| "token_acc": 0.8975998070196599, | |
| "train_speed(iter/s)": 0.029597 | |
| }, | |
| { | |
| "epoch": 0.44638214471421145, | |
| "grad_norm": 0.05793114375836921, | |
| "learning_rate": 1.2239671236429413e-05, | |
| "loss": 0.4537235736846924, | |
| "memory(GiB)": 76.61, | |
| "step": 1395, | |
| "token_acc": 0.8839514422541486, | |
| "train_speed(iter/s)": 0.02958 | |
| }, | |
| { | |
| "epoch": 0.4479820807167713, | |
| "grad_norm": 0.05955306099185102, | |
| "learning_rate": 1.218903323806595e-05, | |
| "loss": 0.4573692798614502, | |
| "memory(GiB)": 76.61, | |
| "step": 1400, | |
| "token_acc": 0.8418099547511312, | |
| "train_speed(iter/s)": 0.029594 | |
| }, | |
| { | |
| "epoch": 0.44958201671933123, | |
| "grad_norm": 0.058484796569864064, | |
| "learning_rate": 1.2138336214107148e-05, | |
| "loss": 0.44894704818725584, | |
| "memory(GiB)": 76.61, | |
| "step": 1405, | |
| "token_acc": 0.8525200458190149, | |
| "train_speed(iter/s)": 0.029594 | |
| }, | |
| { | |
| "epoch": 0.4511819527218911, | |
| "grad_norm": 0.05092836798588581, | |
| "learning_rate": 1.2087581531559208e-05, | |
| "loss": 0.45393967628479004, | |
| "memory(GiB)": 76.61, | |
| "step": 1410, | |
| "token_acc": 0.8791390728476821, | |
| "train_speed(iter/s)": 0.02958 | |
| }, | |
| { | |
| "epoch": 0.452781888724451, | |
| "grad_norm": 0.07033477253264378, | |
| "learning_rate": 1.2036770558983067e-05, | |
| "loss": 0.45307221412658694, | |
| "memory(GiB)": 76.61, | |
| "step": 1415, | |
| "token_acc": 0.8387482900136799, | |
| "train_speed(iter/s)": 0.029599 | |
| }, | |
| { | |
| "epoch": 0.45438182472701094, | |
| "grad_norm": 0.05966547548288182, | |
| "learning_rate": 1.1985904666457455e-05, | |
| "loss": 0.455959415435791, | |
| "memory(GiB)": 76.61, | |
| "step": 1420, | |
| "token_acc": 0.9042096902303416, | |
| "train_speed(iter/s)": 0.029583 | |
| }, | |
| { | |
| "epoch": 0.4559817607295708, | |
| "grad_norm": 0.08159145764722696, | |
| "learning_rate": 1.1934985225541998e-05, | |
| "loss": 0.462065601348877, | |
| "memory(GiB)": 76.61, | |
| "step": 1425, | |
| "token_acc": 0.885252444621832, | |
| "train_speed(iter/s)": 0.029573 | |
| }, | |
| { | |
| "epoch": 0.4575816967321307, | |
| "grad_norm": 0.05540814227664117, | |
| "learning_rate": 1.18840136092402e-05, | |
| "loss": 0.4551572322845459, | |
| "memory(GiB)": 76.61, | |
| "step": 1430, | |
| "token_acc": 0.8559651934966797, | |
| "train_speed(iter/s)": 0.029592 | |
| }, | |
| { | |
| "epoch": 0.4591816327346906, | |
| "grad_norm": 0.05534004007067895, | |
| "learning_rate": 1.1832991191962435e-05, | |
| "loss": 0.4455368995666504, | |
| "memory(GiB)": 76.61, | |
| "step": 1435, | |
| "token_acc": 0.875560538116592, | |
| "train_speed(iter/s)": 0.029576 | |
| }, | |
| { | |
| "epoch": 0.4607815687372505, | |
| "grad_norm": 0.058276771895487044, | |
| "learning_rate": 1.1781919349488894e-05, | |
| "loss": 0.4590908527374268, | |
| "memory(GiB)": 76.61, | |
| "step": 1440, | |
| "token_acc": 0.8510418460478733, | |
| "train_speed(iter/s)": 0.029576 | |
| }, | |
| { | |
| "epoch": 0.46238150473981043, | |
| "grad_norm": 0.05839975543902795, | |
| "learning_rate": 1.1730799458932473e-05, | |
| "loss": 0.462816858291626, | |
| "memory(GiB)": 76.61, | |
| "step": 1445, | |
| "token_acc": 0.9052378085490669, | |
| "train_speed(iter/s)": 0.029586 | |
| }, | |
| { | |
| "epoch": 0.4639814407423703, | |
| "grad_norm": 0.07084434546926481, | |
| "learning_rate": 1.1679632898701649e-05, | |
| "loss": 0.4550295829772949, | |
| "memory(GiB)": 76.61, | |
| "step": 1450, | |
| "token_acc": 0.8805626598465474, | |
| "train_speed(iter/s)": 0.029572 | |
| }, | |
| { | |
| "epoch": 0.4655813767449302, | |
| "grad_norm": 0.06519996046237972, | |
| "learning_rate": 1.1628421048463315e-05, | |
| "loss": 0.46291208267211914, | |
| "memory(GiB)": 76.61, | |
| "step": 1455, | |
| "token_acc": 0.8565744600227359, | |
| "train_speed(iter/s)": 0.029581 | |
| }, | |
| { | |
| "epoch": 0.4671813127474901, | |
| "grad_norm": 0.05799269979733804, | |
| "learning_rate": 1.1577165289105565e-05, | |
| "loss": 0.4474311351776123, | |
| "memory(GiB)": 76.61, | |
| "step": 1460, | |
| "token_acc": 0.8579789309403043, | |
| "train_speed(iter/s)": 0.029568 | |
| }, | |
| { | |
| "epoch": 0.46878124875005, | |
| "grad_norm": 0.057120675003187855, | |
| "learning_rate": 1.1525867002700484e-05, | |
| "loss": 0.46109714508056643, | |
| "memory(GiB)": 76.61, | |
| "step": 1465, | |
| "token_acc": 0.8752182516587126, | |
| "train_speed(iter/s)": 0.029548 | |
| }, | |
| { | |
| "epoch": 0.47038118475260987, | |
| "grad_norm": 0.05696370798749074, | |
| "learning_rate": 1.1474527572466847e-05, | |
| "loss": 0.4501948833465576, | |
| "memory(GiB)": 76.61, | |
| "step": 1470, | |
| "token_acc": 0.8529032258064516, | |
| "train_speed(iter/s)": 0.029562 | |
| }, | |
| { | |
| "epoch": 0.4719811207551698, | |
| "grad_norm": 0.05518112754329221, | |
| "learning_rate": 1.1423148382732854e-05, | |
| "loss": 0.45941987037658694, | |
| "memory(GiB)": 76.61, | |
| "step": 1475, | |
| "token_acc": 0.9009282399143164, | |
| "train_speed(iter/s)": 0.029545 | |
| }, | |
| { | |
| "epoch": 0.4735810567577297, | |
| "grad_norm": 0.051496444525703684, | |
| "learning_rate": 1.1371730818898785e-05, | |
| "loss": 0.45296878814697267, | |
| "memory(GiB)": 76.61, | |
| "step": 1480, | |
| "token_acc": 0.8814303638644918, | |
| "train_speed(iter/s)": 0.029538 | |
| }, | |
| { | |
| "epoch": 0.4751809927602896, | |
| "grad_norm": 0.0677105428949175, | |
| "learning_rate": 1.132027626739965e-05, | |
| "loss": 0.45635080337524414, | |
| "memory(GiB)": 76.61, | |
| "step": 1485, | |
| "token_acc": 0.880248833592535, | |
| "train_speed(iter/s)": 0.029546 | |
| }, | |
| { | |
| "epoch": 0.4767809287628495, | |
| "grad_norm": 0.0673509631098402, | |
| "learning_rate": 1.1268786115667798e-05, | |
| "loss": 0.4614115715026855, | |
| "memory(GiB)": 76.61, | |
| "step": 1490, | |
| "token_acc": 0.8609592251210748, | |
| "train_speed(iter/s)": 0.029525 | |
| }, | |
| { | |
| "epoch": 0.47838086476540936, | |
| "grad_norm": 0.053337771378298794, | |
| "learning_rate": 1.1217261752095518e-05, | |
| "loss": 0.45500664710998534, | |
| "memory(GiB)": 76.61, | |
| "step": 1495, | |
| "token_acc": 0.8794466403162056, | |
| "train_speed(iter/s)": 0.029522 | |
| }, | |
| { | |
| "epoch": 0.4799808007679693, | |
| "grad_norm": 0.05429302474155136, | |
| "learning_rate": 1.1165704565997593e-05, | |
| "loss": 0.44763407707214353, | |
| "memory(GiB)": 76.61, | |
| "step": 1500, | |
| "token_acc": 0.8700440528634361, | |
| "train_speed(iter/s)": 0.02952 | |
| }, | |
| { | |
| "epoch": 0.4799808007679693, | |
| "eval_loss": 0.6668144464492798, | |
| "eval_runtime": 124.2589, | |
| "eval_samples_per_second": 161.662, | |
| "eval_steps_per_second": 0.813, | |
| "eval_token_acc": 0.8694583558206896, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.4815807367705292, | |
| "grad_norm": 0.06146640524587408, | |
| "learning_rate": 1.1114115947573834e-05, | |
| "loss": 0.45711498260498046, | |
| "memory(GiB)": 76.61, | |
| "step": 1505, | |
| "token_acc": 0.8695166967121641, | |
| "train_speed(iter/s)": 0.029461 | |
| }, | |
| { | |
| "epoch": 0.48318067277308907, | |
| "grad_norm": 0.06183782693437151, | |
| "learning_rate": 1.1062497287871606e-05, | |
| "loss": 0.4499336242675781, | |
| "memory(GiB)": 76.61, | |
| "step": 1510, | |
| "token_acc": 0.8487557381009906, | |
| "train_speed(iter/s)": 0.029468 | |
| }, | |
| { | |
| "epoch": 0.484780608775649, | |
| "grad_norm": 0.056753269624682155, | |
| "learning_rate": 1.1010849978748314e-05, | |
| "loss": 0.4551094055175781, | |
| "memory(GiB)": 76.61, | |
| "step": 1515, | |
| "token_acc": 0.8579035448045033, | |
| "train_speed(iter/s)": 0.02948 | |
| }, | |
| { | |
| "epoch": 0.48638054477820886, | |
| "grad_norm": 0.05188962595699218, | |
| "learning_rate": 1.0959175412833869e-05, | |
| "loss": 0.4483503818511963, | |
| "memory(GiB)": 76.61, | |
| "step": 1520, | |
| "token_acc": 0.8649127992905705, | |
| "train_speed(iter/s)": 0.029477 | |
| }, | |
| { | |
| "epoch": 0.4879804807807688, | |
| "grad_norm": 0.0526414480661873, | |
| "learning_rate": 1.0907474983493144e-05, | |
| "loss": 0.45140752792358396, | |
| "memory(GiB)": 76.61, | |
| "step": 1525, | |
| "token_acc": 0.8700204290091931, | |
| "train_speed(iter/s)": 0.029464 | |
| }, | |
| { | |
| "epoch": 0.48958041678332864, | |
| "grad_norm": 0.06682159988119828, | |
| "learning_rate": 1.08557500847884e-05, | |
| "loss": 0.4480952262878418, | |
| "memory(GiB)": 76.61, | |
| "step": 1530, | |
| "token_acc": 0.8906385187748745, | |
| "train_speed(iter/s)": 0.029478 | |
| }, | |
| { | |
| "epoch": 0.49118035278588856, | |
| "grad_norm": 0.06117568492897364, | |
| "learning_rate": 1.080400211144169e-05, | |
| "loss": 0.453688907623291, | |
| "memory(GiB)": 76.61, | |
| "step": 1535, | |
| "token_acc": 0.8130096719135217, | |
| "train_speed(iter/s)": 0.029469 | |
| }, | |
| { | |
| "epoch": 0.4927802887884485, | |
| "grad_norm": 0.05143211947513191, | |
| "learning_rate": 1.0752232458797262e-05, | |
| "loss": 0.44568753242492676, | |
| "memory(GiB)": 76.61, | |
| "step": 1540, | |
| "token_acc": 0.847257743677181, | |
| "train_speed(iter/s)": 0.029464 | |
| }, | |
| { | |
| "epoch": 0.49438022479100835, | |
| "grad_norm": 0.05201971134010435, | |
| "learning_rate": 1.070044252278393e-05, | |
| "loss": 0.46500363349914553, | |
| "memory(GiB)": 76.61, | |
| "step": 1545, | |
| "token_acc": 0.84, | |
| "train_speed(iter/s)": 0.029474 | |
| }, | |
| { | |
| "epoch": 0.49598016079356827, | |
| "grad_norm": 0.05304880581645989, | |
| "learning_rate": 1.064863369987743e-05, | |
| "loss": 0.4501206398010254, | |
| "memory(GiB)": 76.61, | |
| "step": 1550, | |
| "token_acc": 0.8888641920426762, | |
| "train_speed(iter/s)": 0.029465 | |
| }, | |
| { | |
| "epoch": 0.49758009679612814, | |
| "grad_norm": 0.050584443072610216, | |
| "learning_rate": 1.0596807387062772e-05, | |
| "loss": 0.456621789932251, | |
| "memory(GiB)": 76.61, | |
| "step": 1555, | |
| "token_acc": 0.8793768317137128, | |
| "train_speed(iter/s)": 0.02948 | |
| }, | |
| { | |
| "epoch": 0.49918003279868806, | |
| "grad_norm": 0.05907676168100355, | |
| "learning_rate": 1.0544964981796563e-05, | |
| "loss": 0.4567122936248779, | |
| "memory(GiB)": 76.61, | |
| "step": 1560, | |
| "token_acc": 0.8505747126436781, | |
| "train_speed(iter/s)": 0.029477 | |
| }, | |
| { | |
| "epoch": 0.500779968801248, | |
| "grad_norm": 0.055037989511506104, | |
| "learning_rate": 1.0493107881969335e-05, | |
| "loss": 0.44720020294189455, | |
| "memory(GiB)": 76.61, | |
| "step": 1565, | |
| "token_acc": 0.8853304383227032, | |
| "train_speed(iter/s)": 0.029466 | |
| }, | |
| { | |
| "epoch": 0.5023799048038079, | |
| "grad_norm": 0.0597376748229471, | |
| "learning_rate": 1.0441237485867845e-05, | |
| "loss": 0.4492997169494629, | |
| "memory(GiB)": 76.61, | |
| "step": 1570, | |
| "token_acc": 0.8809347181008902, | |
| "train_speed(iter/s)": 0.029489 | |
| }, | |
| { | |
| "epoch": 0.5039798408063677, | |
| "grad_norm": 0.060265741182571844, | |
| "learning_rate": 1.0389355192137379e-05, | |
| "loss": 0.4525942325592041, | |
| "memory(GiB)": 76.61, | |
| "step": 1575, | |
| "token_acc": 0.8839541547277937, | |
| "train_speed(iter/s)": 0.029481 | |
| }, | |
| { | |
| "epoch": 0.5055797768089276, | |
| "grad_norm": 0.06015007204584338, | |
| "learning_rate": 1.0337462399744025e-05, | |
| "loss": 0.4606604099273682, | |
| "memory(GiB)": 76.61, | |
| "step": 1580, | |
| "token_acc": 0.8439696373348328, | |
| "train_speed(iter/s)": 0.029471 | |
| }, | |
| { | |
| "epoch": 0.5071797128114875, | |
| "grad_norm": 0.0539606724017438, | |
| "learning_rate": 1.0285560507936962e-05, | |
| "loss": 0.46471481323242186, | |
| "memory(GiB)": 76.61, | |
| "step": 1585, | |
| "token_acc": 0.8212732305258995, | |
| "train_speed(iter/s)": 0.029486 | |
| }, | |
| { | |
| "epoch": 0.5087796488140475, | |
| "grad_norm": 0.0588254805138369, | |
| "learning_rate": 1.0233650916210736e-05, | |
| "loss": 0.45154604911804197, | |
| "memory(GiB)": 76.61, | |
| "step": 1590, | |
| "token_acc": 0.883357041251778, | |
| "train_speed(iter/s)": 0.029474 | |
| }, | |
| { | |
| "epoch": 0.5103795848166074, | |
| "grad_norm": 0.06409304780777438, | |
| "learning_rate": 1.0181735024267504e-05, | |
| "loss": 0.45000271797180175, | |
| "memory(GiB)": 76.61, | |
| "step": 1595, | |
| "token_acc": 0.8340197693574959, | |
| "train_speed(iter/s)": 0.029485 | |
| }, | |
| { | |
| "epoch": 0.5119795208191672, | |
| "grad_norm": 0.058200459243944895, | |
| "learning_rate": 1.012981423197931e-05, | |
| "loss": 0.4608008861541748, | |
| "memory(GiB)": 76.61, | |
| "step": 1600, | |
| "token_acc": 0.8793913904007917, | |
| "train_speed(iter/s)": 0.029484 | |
| }, | |
| { | |
| "epoch": 0.5135794568217271, | |
| "grad_norm": 0.052541653818392466, | |
| "learning_rate": 1.007788993935033e-05, | |
| "loss": 0.45448942184448243, | |
| "memory(GiB)": 76.61, | |
| "step": 1605, | |
| "token_acc": 0.8615229110512129, | |
| "train_speed(iter/s)": 0.029472 | |
| }, | |
| { | |
| "epoch": 0.515179392824287, | |
| "grad_norm": 0.06526426191856917, | |
| "learning_rate": 1.002596354647912e-05, | |
| "loss": 0.45614914894104003, | |
| "memory(GiB)": 76.61, | |
| "step": 1610, | |
| "token_acc": 0.871312462372065, | |
| "train_speed(iter/s)": 0.029489 | |
| }, | |
| { | |
| "epoch": 0.516779328826847, | |
| "grad_norm": 0.05743481751682084, | |
| "learning_rate": 9.974036453520881e-06, | |
| "loss": 0.447450590133667, | |
| "memory(GiB)": 76.61, | |
| "step": 1615, | |
| "token_acc": 0.8760352658295485, | |
| "train_speed(iter/s)": 0.02948 | |
| }, | |
| { | |
| "epoch": 0.5183792648294068, | |
| "grad_norm": 0.06577378911981274, | |
| "learning_rate": 9.922110060649672e-06, | |
| "loss": 0.45809640884399416, | |
| "memory(GiB)": 76.61, | |
| "step": 1620, | |
| "token_acc": 0.9048299514146899, | |
| "train_speed(iter/s)": 0.029468 | |
| }, | |
| { | |
| "epoch": 0.5199792008319667, | |
| "grad_norm": 0.05180626448607971, | |
| "learning_rate": 9.870185768020694e-06, | |
| "loss": 0.4360641002655029, | |
| "memory(GiB)": 76.61, | |
| "step": 1625, | |
| "token_acc": 0.890797148412184, | |
| "train_speed(iter/s)": 0.029481 | |
| }, | |
| { | |
| "epoch": 0.5215791368345266, | |
| "grad_norm": 0.048795347699578454, | |
| "learning_rate": 9.818264975732497e-06, | |
| "loss": 0.4505919933319092, | |
| "memory(GiB)": 76.61, | |
| "step": 1630, | |
| "token_acc": 0.8830155979202773, | |
| "train_speed(iter/s)": 0.029462 | |
| }, | |
| { | |
| "epoch": 0.5231790728370865, | |
| "grad_norm": 0.054325754690003863, | |
| "learning_rate": 9.766349083789266e-06, | |
| "loss": 0.4518167495727539, | |
| "memory(GiB)": 76.61, | |
| "step": 1635, | |
| "token_acc": 0.8740636704119851, | |
| "train_speed(iter/s)": 0.029458 | |
| }, | |
| { | |
| "epoch": 0.5247790088396465, | |
| "grad_norm": 0.05473270131153146, | |
| "learning_rate": 9.71443949206304e-06, | |
| "loss": 0.4629377841949463, | |
| "memory(GiB)": 76.61, | |
| "step": 1640, | |
| "token_acc": 0.8927648578811369, | |
| "train_speed(iter/s)": 0.029469 | |
| }, | |
| { | |
| "epoch": 0.5263789448422063, | |
| "grad_norm": 0.05962553624327487, | |
| "learning_rate": 9.662537600255979e-06, | |
| "loss": 0.4535552501678467, | |
| "memory(GiB)": 76.61, | |
| "step": 1645, | |
| "token_acc": 0.8980960623106881, | |
| "train_speed(iter/s)": 0.029457 | |
| }, | |
| { | |
| "epoch": 0.5279788808447662, | |
| "grad_norm": 0.06367541172972058, | |
| "learning_rate": 9.610644807862625e-06, | |
| "loss": 0.44418978691101074, | |
| "memory(GiB)": 76.61, | |
| "step": 1650, | |
| "token_acc": 0.8769371011850501, | |
| "train_speed(iter/s)": 0.029468 | |
| }, | |
| { | |
| "epoch": 0.5295788168473261, | |
| "grad_norm": 0.05288367644088033, | |
| "learning_rate": 9.558762514132157e-06, | |
| "loss": 0.4513704299926758, | |
| "memory(GiB)": 76.61, | |
| "step": 1655, | |
| "token_acc": 0.8576478906434126, | |
| "train_speed(iter/s)": 0.029464 | |
| }, | |
| { | |
| "epoch": 0.531178752849886, | |
| "grad_norm": 0.054919719691940275, | |
| "learning_rate": 9.506892118030668e-06, | |
| "loss": 0.4454075336456299, | |
| "memory(GiB)": 76.61, | |
| "step": 1660, | |
| "token_acc": 0.8535078688042359, | |
| "train_speed(iter/s)": 0.029456 | |
| }, | |
| { | |
| "epoch": 0.532778688852446, | |
| "grad_norm": 0.0561391497524031, | |
| "learning_rate": 9.455035018203439e-06, | |
| "loss": 0.4459484100341797, | |
| "memory(GiB)": 76.61, | |
| "step": 1665, | |
| "token_acc": 0.8793124922157181, | |
| "train_speed(iter/s)": 0.029471 | |
| }, | |
| { | |
| "epoch": 0.5343786248550058, | |
| "grad_norm": 0.051526152917333715, | |
| "learning_rate": 9.40319261293723e-06, | |
| "loss": 0.4593966484069824, | |
| "memory(GiB)": 76.61, | |
| "step": 1670, | |
| "token_acc": 0.8957568638966378, | |
| "train_speed(iter/s)": 0.029466 | |
| }, | |
| { | |
| "epoch": 0.5359785608575657, | |
| "grad_norm": 0.05336577516465571, | |
| "learning_rate": 9.351366300122569e-06, | |
| "loss": 0.45195541381835935, | |
| "memory(GiB)": 76.61, | |
| "step": 1675, | |
| "token_acc": 0.8254120659305488, | |
| "train_speed(iter/s)": 0.029459 | |
| }, | |
| { | |
| "epoch": 0.5375784968601256, | |
| "grad_norm": 0.05605931671991975, | |
| "learning_rate": 9.299557477216073e-06, | |
| "loss": 0.4473400115966797, | |
| "memory(GiB)": 76.61, | |
| "step": 1680, | |
| "token_acc": 0.8684433164128595, | |
| "train_speed(iter/s)": 0.029474 | |
| }, | |
| { | |
| "epoch": 0.5391784328626855, | |
| "grad_norm": 0.05284398220938548, | |
| "learning_rate": 9.247767541202738e-06, | |
| "loss": 0.4539934158325195, | |
| "memory(GiB)": 76.61, | |
| "step": 1685, | |
| "token_acc": 0.8787515006002401, | |
| "train_speed(iter/s)": 0.029458 | |
| }, | |
| { | |
| "epoch": 0.5407783688652454, | |
| "grad_norm": 0.06074778175058899, | |
| "learning_rate": 9.195997888558312e-06, | |
| "loss": 0.4540121078491211, | |
| "memory(GiB)": 76.61, | |
| "step": 1690, | |
| "token_acc": 0.882076702321941, | |
| "train_speed(iter/s)": 0.029458 | |
| }, | |
| { | |
| "epoch": 0.5423783048678052, | |
| "grad_norm": 0.06072504661929311, | |
| "learning_rate": 9.144249915211605e-06, | |
| "loss": 0.45176243782043457, | |
| "memory(GiB)": 76.61, | |
| "step": 1695, | |
| "token_acc": 0.8652057386094908, | |
| "train_speed(iter/s)": 0.029465 | |
| }, | |
| { | |
| "epoch": 0.5439782408703652, | |
| "grad_norm": 0.058552695609385315, | |
| "learning_rate": 9.092525016506858e-06, | |
| "loss": 0.4491862773895264, | |
| "memory(GiB)": 76.61, | |
| "step": 1700, | |
| "token_acc": 0.8822588020118884, | |
| "train_speed(iter/s)": 0.02945 | |
| }, | |
| { | |
| "epoch": 0.5455781768729251, | |
| "grad_norm": 0.056892490634495974, | |
| "learning_rate": 9.040824587166136e-06, | |
| "loss": 0.45043745040893557, | |
| "memory(GiB)": 76.61, | |
| "step": 1705, | |
| "token_acc": 0.8825789923142613, | |
| "train_speed(iter/s)": 0.029461 | |
| }, | |
| { | |
| "epoch": 0.547178112875485, | |
| "grad_norm": 0.05885692671807609, | |
| "learning_rate": 8.98915002125169e-06, | |
| "loss": 0.4475353240966797, | |
| "memory(GiB)": 76.61, | |
| "step": 1710, | |
| "token_acc": 0.8721031538595574, | |
| "train_speed(iter/s)": 0.029454 | |
| }, | |
| { | |
| "epoch": 0.5487780488780449, | |
| "grad_norm": 0.060276094585736115, | |
| "learning_rate": 8.9375027121284e-06, | |
| "loss": 0.4502556800842285, | |
| "memory(GiB)": 76.61, | |
| "step": 1715, | |
| "token_acc": 0.8562842259917189, | |
| "train_speed(iter/s)": 0.029449 | |
| }, | |
| { | |
| "epoch": 0.5503779848806047, | |
| "grad_norm": 0.06782068962590707, | |
| "learning_rate": 8.885884052426168e-06, | |
| "loss": 0.4532322883605957, | |
| "memory(GiB)": 76.61, | |
| "step": 1720, | |
| "token_acc": 0.8593545573484518, | |
| "train_speed(iter/s)": 0.029466 | |
| }, | |
| { | |
| "epoch": 0.5519779208831647, | |
| "grad_norm": 0.06070839045848377, | |
| "learning_rate": 8.83429543400241e-06, | |
| "loss": 0.45258092880249023, | |
| "memory(GiB)": 76.61, | |
| "step": 1725, | |
| "token_acc": 0.8751242791807516, | |
| "train_speed(iter/s)": 0.029452 | |
| }, | |
| { | |
| "epoch": 0.5535778568857246, | |
| "grad_norm": 0.049979952181739715, | |
| "learning_rate": 8.78273824790448e-06, | |
| "loss": 0.4340657234191895, | |
| "memory(GiB)": 76.61, | |
| "step": 1730, | |
| "token_acc": 0.8650843222985634, | |
| "train_speed(iter/s)": 0.029451 | |
| }, | |
| { | |
| "epoch": 0.5551777928882845, | |
| "grad_norm": 0.059124658124222323, | |
| "learning_rate": 8.731213884332205e-06, | |
| "loss": 0.43556828498840333, | |
| "memory(GiB)": 76.61, | |
| "step": 1735, | |
| "token_acc": 0.8524390243902439, | |
| "train_speed(iter/s)": 0.029459 | |
| }, | |
| { | |
| "epoch": 0.5567777288908443, | |
| "grad_norm": 0.05228309031135195, | |
| "learning_rate": 8.679723732600355e-06, | |
| "loss": 0.4483633041381836, | |
| "memory(GiB)": 76.61, | |
| "step": 1740, | |
| "token_acc": 0.9039166284928997, | |
| "train_speed(iter/s)": 0.029445 | |
| }, | |
| { | |
| "epoch": 0.5583776648934042, | |
| "grad_norm": 0.05659321921396489, | |
| "learning_rate": 8.628269181101216e-06, | |
| "loss": 0.45377864837646487, | |
| "memory(GiB)": 76.61, | |
| "step": 1745, | |
| "token_acc": 0.8812897628687102, | |
| "train_speed(iter/s)": 0.029449 | |
| }, | |
| { | |
| "epoch": 0.5599776008959642, | |
| "grad_norm": 0.0610469666222746, | |
| "learning_rate": 8.576851617267151e-06, | |
| "loss": 0.4495216369628906, | |
| "memory(GiB)": 76.61, | |
| "step": 1750, | |
| "token_acc": 0.8734145104008117, | |
| "train_speed(iter/s)": 0.029452 | |
| }, | |
| { | |
| "epoch": 0.5599776008959642, | |
| "eval_loss": 0.6640093922615051, | |
| "eval_runtime": 114.9985, | |
| "eval_samples_per_second": 174.681, | |
| "eval_steps_per_second": 0.878, | |
| "eval_token_acc": 0.8701128116616424, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.5615775368985241, | |
| "grad_norm": 0.061306770256929585, | |
| "learning_rate": 8.525472427533156e-06, | |
| "loss": 0.44908857345581055, | |
| "memory(GiB)": 77.63, | |
| "step": 1755, | |
| "token_acc": 0.8715457946180765, | |
| "train_speed(iter/s)": 0.029409 | |
| }, | |
| { | |
| "epoch": 0.563177472901084, | |
| "grad_norm": 0.05235639827008535, | |
| "learning_rate": 8.474132997299521e-06, | |
| "loss": 0.4579316139221191, | |
| "memory(GiB)": 77.63, | |
| "step": 1760, | |
| "token_acc": 0.8922923256201098, | |
| "train_speed(iter/s)": 0.029422 | |
| }, | |
| { | |
| "epoch": 0.5647774089036438, | |
| "grad_norm": 0.051281007426132216, | |
| "learning_rate": 8.422834710894434e-06, | |
| "loss": 0.45467004776000974, | |
| "memory(GiB)": 77.63, | |
| "step": 1765, | |
| "token_acc": 0.903878366189924, | |
| "train_speed(iter/s)": 0.029438 | |
| }, | |
| { | |
| "epoch": 0.5663773449062037, | |
| "grad_norm": 0.05049109520782513, | |
| "learning_rate": 8.371578951536689e-06, | |
| "loss": 0.45294957160949706, | |
| "memory(GiB)": 77.63, | |
| "step": 1770, | |
| "token_acc": 0.8928110202324581, | |
| "train_speed(iter/s)": 0.029439 | |
| }, | |
| { | |
| "epoch": 0.5679772809087636, | |
| "grad_norm": 0.04946427707777728, | |
| "learning_rate": 8.320367101298351e-06, | |
| "loss": 0.4473431587219238, | |
| "memory(GiB)": 77.63, | |
| "step": 1775, | |
| "token_acc": 0.8723599632690542, | |
| "train_speed(iter/s)": 0.029439 | |
| }, | |
| { | |
| "epoch": 0.5695772169113236, | |
| "grad_norm": 0.053606352244487274, | |
| "learning_rate": 8.26920054106753e-06, | |
| "loss": 0.4495864391326904, | |
| "memory(GiB)": 77.63, | |
| "step": 1780, | |
| "token_acc": 0.8844315111203492, | |
| "train_speed(iter/s)": 0.029459 | |
| }, | |
| { | |
| "epoch": 0.5711771529138835, | |
| "grad_norm": 0.05525614374940963, | |
| "learning_rate": 8.218080650511107e-06, | |
| "loss": 0.44890499114990234, | |
| "memory(GiB)": 77.63, | |
| "step": 1785, | |
| "token_acc": 0.8749736453721273, | |
| "train_speed(iter/s)": 0.02946 | |
| }, | |
| { | |
| "epoch": 0.5727770889164433, | |
| "grad_norm": 0.05882148265537131, | |
| "learning_rate": 8.167008808037568e-06, | |
| "loss": 0.44676194190979, | |
| "memory(GiB)": 77.63, | |
| "step": 1790, | |
| "token_acc": 0.8807511737089202, | |
| "train_speed(iter/s)": 0.029457 | |
| }, | |
| { | |
| "epoch": 0.5743770249190032, | |
| "grad_norm": 0.048821121641334515, | |
| "learning_rate": 8.115986390759805e-06, | |
| "loss": 0.4417415142059326, | |
| "memory(GiB)": 77.63, | |
| "step": 1795, | |
| "token_acc": 0.8531673379714391, | |
| "train_speed(iter/s)": 0.029469 | |
| }, | |
| { | |
| "epoch": 0.5759769609215631, | |
| "grad_norm": 0.054949264031140505, | |
| "learning_rate": 8.065014774458004e-06, | |
| "loss": 0.46439437866210936, | |
| "memory(GiB)": 77.63, | |
| "step": 1800, | |
| "token_acc": 0.8333022213925705, | |
| "train_speed(iter/s)": 0.029467 | |
| }, | |
| { | |
| "epoch": 0.5775768969241231, | |
| "grad_norm": 0.059507220518762304, | |
| "learning_rate": 8.014095333542548e-06, | |
| "loss": 0.4539642333984375, | |
| "memory(GiB)": 77.63, | |
| "step": 1805, | |
| "token_acc": 0.8577178858942948, | |
| "train_speed(iter/s)": 0.029483 | |
| }, | |
| { | |
| "epoch": 0.579176832926683, | |
| "grad_norm": 0.05302143027350534, | |
| "learning_rate": 7.963229441016938e-06, | |
| "loss": 0.4606470108032227, | |
| "memory(GiB)": 77.63, | |
| "step": 1810, | |
| "token_acc": 0.8760775862068966, | |
| "train_speed(iter/s)": 0.029484 | |
| }, | |
| { | |
| "epoch": 0.5807767689292428, | |
| "grad_norm": 0.0699581228289572, | |
| "learning_rate": 7.912418468440794e-06, | |
| "loss": 0.4488551139831543, | |
| "memory(GiB)": 77.63, | |
| "step": 1815, | |
| "token_acc": 0.8892276422764228, | |
| "train_speed(iter/s)": 0.029481 | |
| }, | |
| { | |
| "epoch": 0.5823767049318027, | |
| "grad_norm": 0.053456667148895104, | |
| "learning_rate": 7.861663785892857e-06, | |
| "loss": 0.45035881996154786, | |
| "memory(GiB)": 77.63, | |
| "step": 1820, | |
| "token_acc": 0.8806643202815662, | |
| "train_speed(iter/s)": 0.029498 | |
| }, | |
| { | |
| "epoch": 0.5839766409343626, | |
| "grad_norm": 0.05451209338463787, | |
| "learning_rate": 7.810966761934053e-06, | |
| "loss": 0.44800753593444825, | |
| "memory(GiB)": 77.63, | |
| "step": 1825, | |
| "token_acc": 0.8771571298819255, | |
| "train_speed(iter/s)": 0.029496 | |
| }, | |
| { | |
| "epoch": 0.5855765769369226, | |
| "grad_norm": 0.05912934985203241, | |
| "learning_rate": 7.760328763570589e-06, | |
| "loss": 0.4499057769775391, | |
| "memory(GiB)": 77.63, | |
| "step": 1830, | |
| "token_acc": 0.868710326675956, | |
| "train_speed(iter/s)": 0.029487 | |
| }, | |
| { | |
| "epoch": 0.5871765129394825, | |
| "grad_norm": 0.052841905445767515, | |
| "learning_rate": 7.709751156217088e-06, | |
| "loss": 0.4497323989868164, | |
| "memory(GiB)": 77.63, | |
| "step": 1835, | |
| "token_acc": 0.8117094325984822, | |
| "train_speed(iter/s)": 0.029501 | |
| }, | |
| { | |
| "epoch": 0.5887764489420423, | |
| "grad_norm": 0.060076953422732254, | |
| "learning_rate": 7.659235303659784e-06, | |
| "loss": 0.4582187652587891, | |
| "memory(GiB)": 77.63, | |
| "step": 1840, | |
| "token_acc": 0.8795674258561363, | |
| "train_speed(iter/s)": 0.029492 | |
| }, | |
| { | |
| "epoch": 0.5903763849446022, | |
| "grad_norm": 0.06307528499562465, | |
| "learning_rate": 7.608782568019729e-06, | |
| "loss": 0.4430552005767822, | |
| "memory(GiB)": 77.63, | |
| "step": 1845, | |
| "token_acc": 0.8452444922084901, | |
| "train_speed(iter/s)": 0.029498 | |
| }, | |
| { | |
| "epoch": 0.5919763209471621, | |
| "grad_norm": 0.05378691938628143, | |
| "learning_rate": 7.558394309716088e-06, | |
| "loss": 0.459810209274292, | |
| "memory(GiB)": 77.63, | |
| "step": 1850, | |
| "token_acc": 0.8506092736192435, | |
| "train_speed(iter/s)": 0.029503 | |
| }, | |
| { | |
| "epoch": 0.593576256949722, | |
| "grad_norm": 0.0586506530143339, | |
| "learning_rate": 7.508071887429433e-06, | |
| "loss": 0.46239190101623534, | |
| "memory(GiB)": 77.63, | |
| "step": 1855, | |
| "token_acc": 0.9115304709141274, | |
| "train_speed(iter/s)": 0.029495 | |
| }, | |
| { | |
| "epoch": 0.5951761929522819, | |
| "grad_norm": 0.053290473896441634, | |
| "learning_rate": 7.4578166580651335e-06, | |
| "loss": 0.4524221897125244, | |
| "memory(GiB)": 77.63, | |
| "step": 1860, | |
| "token_acc": 0.8817879571481345, | |
| "train_speed(iter/s)": 0.029508 | |
| }, | |
| { | |
| "epoch": 0.5967761289548418, | |
| "grad_norm": 0.051901913358510056, | |
| "learning_rate": 7.4076299767167325e-06, | |
| "loss": 0.4579325675964355, | |
| "memory(GiB)": 77.63, | |
| "step": 1865, | |
| "token_acc": 0.8617401668653158, | |
| "train_speed(iter/s)": 0.029506 | |
| }, | |
| { | |
| "epoch": 0.5983760649574017, | |
| "grad_norm": 0.05256077511072294, | |
| "learning_rate": 7.35751319662945e-06, | |
| "loss": 0.45406513214111327, | |
| "memory(GiB)": 77.63, | |
| "step": 1870, | |
| "token_acc": 0.8924402944873406, | |
| "train_speed(iter/s)": 0.029497 | |
| }, | |
| { | |
| "epoch": 0.5999760009599616, | |
| "grad_norm": 0.056121622843709036, | |
| "learning_rate": 7.307467669163655e-06, | |
| "loss": 0.450104284286499, | |
| "memory(GiB)": 77.63, | |
| "step": 1875, | |
| "token_acc": 0.8646184340931615, | |
| "train_speed(iter/s)": 0.02951 | |
| }, | |
| { | |
| "epoch": 0.6015759369625215, | |
| "grad_norm": 0.051068951060234354, | |
| "learning_rate": 7.25749474375846e-06, | |
| "loss": 0.45695791244506834, | |
| "memory(GiB)": 77.63, | |
| "step": 1880, | |
| "token_acc": 0.9112655568126717, | |
| "train_speed(iter/s)": 0.029503 | |
| }, | |
| { | |
| "epoch": 0.6031758729650813, | |
| "grad_norm": 0.05120698584703106, | |
| "learning_rate": 7.207595767895303e-06, | |
| "loss": 0.4460740089416504, | |
| "memory(GiB)": 77.63, | |
| "step": 1885, | |
| "token_acc": 0.8637192342752963, | |
| "train_speed(iter/s)": 0.029499 | |
| }, | |
| { | |
| "epoch": 0.6047758089676413, | |
| "grad_norm": 0.05826366701259215, | |
| "learning_rate": 7.157772087061645e-06, | |
| "loss": 0.4498391628265381, | |
| "memory(GiB)": 77.63, | |
| "step": 1890, | |
| "token_acc": 0.8602477214302408, | |
| "train_speed(iter/s)": 0.029509 | |
| }, | |
| { | |
| "epoch": 0.6063757449702012, | |
| "grad_norm": 0.05454678875604061, | |
| "learning_rate": 7.108025044714661e-06, | |
| "loss": 0.44768247604370115, | |
| "memory(GiB)": 77.63, | |
| "step": 1895, | |
| "token_acc": 0.8998014357721094, | |
| "train_speed(iter/s)": 0.029496 | |
| }, | |
| { | |
| "epoch": 0.6079756809727611, | |
| "grad_norm": 0.04862560763785379, | |
| "learning_rate": 7.058355982245038e-06, | |
| "loss": 0.44283151626586914, | |
| "memory(GiB)": 77.63, | |
| "step": 1900, | |
| "token_acc": 0.8749580958766343, | |
| "train_speed(iter/s)": 0.029501 | |
| }, | |
| { | |
| "epoch": 0.609575616975321, | |
| "grad_norm": 0.05390239428952395, | |
| "learning_rate": 7.00876623894079e-06, | |
| "loss": 0.4445077419281006, | |
| "memory(GiB)": 77.63, | |
| "step": 1905, | |
| "token_acc": 0.8588156123822341, | |
| "train_speed(iter/s)": 0.029504 | |
| }, | |
| { | |
| "epoch": 0.6111755529778808, | |
| "grad_norm": 0.052917745372876655, | |
| "learning_rate": 6.959257151951153e-06, | |
| "loss": 0.45001955032348634, | |
| "memory(GiB)": 77.63, | |
| "step": 1910, | |
| "token_acc": 0.8768155911013054, | |
| "train_speed(iter/s)": 0.029494 | |
| }, | |
| { | |
| "epoch": 0.6127754889804408, | |
| "grad_norm": 0.05432256049056495, | |
| "learning_rate": 6.909830056250527e-06, | |
| "loss": 0.44944238662719727, | |
| "memory(GiB)": 77.63, | |
| "step": 1915, | |
| "token_acc": 0.8941244909831297, | |
| "train_speed(iter/s)": 0.029507 | |
| }, | |
| { | |
| "epoch": 0.6143754249830007, | |
| "grad_norm": 0.05852297407331436, | |
| "learning_rate": 6.860486284602479e-06, | |
| "loss": 0.4477729797363281, | |
| "memory(GiB)": 77.63, | |
| "step": 1920, | |
| "token_acc": 0.8854845719252499, | |
| "train_speed(iter/s)": 0.029501 | |
| }, | |
| { | |
| "epoch": 0.6159753609855606, | |
| "grad_norm": 0.05474007805899836, | |
| "learning_rate": 6.8112271675238154e-06, | |
| "loss": 0.4501204013824463, | |
| "memory(GiB)": 77.63, | |
| "step": 1925, | |
| "token_acc": 0.8803290949887809, | |
| "train_speed(iter/s)": 0.029496 | |
| }, | |
| { | |
| "epoch": 0.6175752969881205, | |
| "grad_norm": 0.05545012433641634, | |
| "learning_rate": 6.762054033248681e-06, | |
| "loss": 0.44565958976745607, | |
| "memory(GiB)": 77.63, | |
| "step": 1930, | |
| "token_acc": 0.8480542195015304, | |
| "train_speed(iter/s)": 0.029507 | |
| }, | |
| { | |
| "epoch": 0.6191752329906803, | |
| "grad_norm": 0.05495247298953925, | |
| "learning_rate": 6.712968207692778e-06, | |
| "loss": 0.44170804023742677, | |
| "memory(GiB)": 77.63, | |
| "step": 1935, | |
| "token_acc": 0.8709073900841908, | |
| "train_speed(iter/s)": 0.029498 | |
| }, | |
| { | |
| "epoch": 0.6207751689932403, | |
| "grad_norm": 0.05792014047889592, | |
| "learning_rate": 6.663971014417585e-06, | |
| "loss": 0.4454016208648682, | |
| "memory(GiB)": 77.63, | |
| "step": 1940, | |
| "token_acc": 0.8606651376146789, | |
| "train_speed(iter/s)": 0.0295 | |
| }, | |
| { | |
| "epoch": 0.6223751049958002, | |
| "grad_norm": 0.04853659131630362, | |
| "learning_rate": 6.615063774594677e-06, | |
| "loss": 0.4387532711029053, | |
| "memory(GiB)": 77.63, | |
| "step": 1945, | |
| "token_acc": 0.8920454545454546, | |
| "train_speed(iter/s)": 0.029507 | |
| }, | |
| { | |
| "epoch": 0.6239750409983601, | |
| "grad_norm": 0.05266495974136303, | |
| "learning_rate": 6.566247806970119e-06, | |
| "loss": 0.4472493171691895, | |
| "memory(GiB)": 77.63, | |
| "step": 1950, | |
| "token_acc": 0.848505251817937, | |
| "train_speed(iter/s)": 0.029497 | |
| }, | |
| { | |
| "epoch": 0.62557497700092, | |
| "grad_norm": 0.054994759694813, | |
| "learning_rate": 6.5175244278288705e-06, | |
| "loss": 0.44487895965576174, | |
| "memory(GiB)": 77.63, | |
| "step": 1955, | |
| "token_acc": 0.8689320388349514, | |
| "train_speed(iter/s)": 0.029506 | |
| }, | |
| { | |
| "epoch": 0.6271749130034798, | |
| "grad_norm": 0.057067387083368365, | |
| "learning_rate": 6.468894950959336e-06, | |
| "loss": 0.4466127395629883, | |
| "memory(GiB)": 77.63, | |
| "step": 1960, | |
| "token_acc": 0.846737755286463, | |
| "train_speed(iter/s)": 0.0295 | |
| }, | |
| { | |
| "epoch": 0.6287748490060397, | |
| "grad_norm": 0.05417940634204734, | |
| "learning_rate": 6.420360687617897e-06, | |
| "loss": 0.44883151054382325, | |
| "memory(GiB)": 77.63, | |
| "step": 1965, | |
| "token_acc": 0.8795967892477132, | |
| "train_speed(iter/s)": 0.02949 | |
| }, | |
| { | |
| "epoch": 0.6303747850085997, | |
| "grad_norm": 0.05176488752108695, | |
| "learning_rate": 6.3719229464935915e-06, | |
| "loss": 0.4542849063873291, | |
| "memory(GiB)": 77.63, | |
| "step": 1970, | |
| "token_acc": 0.8568893191352049, | |
| "train_speed(iter/s)": 0.029503 | |
| }, | |
| { | |
| "epoch": 0.6319747210111596, | |
| "grad_norm": 0.04867135924369273, | |
| "learning_rate": 6.323583033672799e-06, | |
| "loss": 0.44331774711608884, | |
| "memory(GiB)": 77.63, | |
| "step": 1975, | |
| "token_acc": 0.8647865559204172, | |
| "train_speed(iter/s)": 0.029491 | |
| }, | |
| { | |
| "epoch": 0.6335746570137194, | |
| "grad_norm": 0.06076783884358601, | |
| "learning_rate": 6.275342252604044e-06, | |
| "loss": 0.44751858711242676, | |
| "memory(GiB)": 77.63, | |
| "step": 1980, | |
| "token_acc": 0.871765773944621, | |
| "train_speed(iter/s)": 0.029486 | |
| }, | |
| { | |
| "epoch": 0.6351745930162793, | |
| "grad_norm": 0.0520886449098567, | |
| "learning_rate": 6.22720190406283e-06, | |
| "loss": 0.46150927543640136, | |
| "memory(GiB)": 77.63, | |
| "step": 1985, | |
| "token_acc": 0.8921661480178595, | |
| "train_speed(iter/s)": 0.029497 | |
| }, | |
| { | |
| "epoch": 0.6367745290188392, | |
| "grad_norm": 0.058090405193780774, | |
| "learning_rate": 6.179163286116581e-06, | |
| "loss": 0.44019436836242676, | |
| "memory(GiB)": 77.63, | |
| "step": 1990, | |
| "token_acc": 0.9157033805888768, | |
| "train_speed(iter/s)": 0.029488 | |
| }, | |
| { | |
| "epoch": 0.6383744650213992, | |
| "grad_norm": 0.057472120727550105, | |
| "learning_rate": 6.13122769408963e-06, | |
| "loss": 0.4466409683227539, | |
| "memory(GiB)": 77.63, | |
| "step": 1995, | |
| "token_acc": 0.8608313968499871, | |
| "train_speed(iter/s)": 0.029492 | |
| }, | |
| { | |
| "epoch": 0.6399744010239591, | |
| "grad_norm": 0.05665485079826101, | |
| "learning_rate": 6.083396420528298e-06, | |
| "loss": 0.451153039932251, | |
| "memory(GiB)": 77.63, | |
| "step": 2000, | |
| "token_acc": 0.8910367046369808, | |
| "train_speed(iter/s)": 0.029496 | |
| }, | |
| { | |
| "epoch": 0.6399744010239591, | |
| "eval_loss": 0.6622327566146851, | |
| "eval_runtime": 115.9166, | |
| "eval_samples_per_second": 173.297, | |
| "eval_steps_per_second": 0.871, | |
| "eval_token_acc": 0.8709580958089251, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.6415743370265189, | |
| "grad_norm": 0.05247711719910443, | |
| "learning_rate": 6.0356707551660434e-06, | |
| "loss": 0.45055346488952636, | |
| "memory(GiB)": 77.63, | |
| "step": 2005, | |
| "token_acc": 0.8961562482257424, | |
| "train_speed(iter/s)": 0.029453 | |
| }, | |
| { | |
| "epoch": 0.6431742730290788, | |
| "grad_norm": 0.0533769803289562, | |
| "learning_rate": 5.988051984888668e-06, | |
| "loss": 0.4436792373657227, | |
| "memory(GiB)": 77.63, | |
| "step": 2010, | |
| "token_acc": 0.8894836272040302, | |
| "train_speed(iter/s)": 0.029463 | |
| }, | |
| { | |
| "epoch": 0.6447742090316387, | |
| "grad_norm": 0.05428299581856707, | |
| "learning_rate": 5.940541393699646e-06, | |
| "loss": 0.44562363624572754, | |
| "memory(GiB)": 77.63, | |
| "step": 2015, | |
| "token_acc": 0.8804031789106416, | |
| "train_speed(iter/s)": 0.029477 | |
| }, | |
| { | |
| "epoch": 0.6463741450341987, | |
| "grad_norm": 0.055930584071511934, | |
| "learning_rate": 5.893140262685469e-06, | |
| "loss": 0.4412201404571533, | |
| "memory(GiB)": 77.63, | |
| "step": 2020, | |
| "token_acc": 0.8791348600508906, | |
| "train_speed(iter/s)": 0.029482 | |
| }, | |
| { | |
| "epoch": 0.6479740810367586, | |
| "grad_norm": 0.06077731970466293, | |
| "learning_rate": 5.845849869981137e-06, | |
| "loss": 0.44964237213134767, | |
| "memory(GiB)": 77.63, | |
| "step": 2025, | |
| "token_acc": 0.8710450018908358, | |
| "train_speed(iter/s)": 0.029484 | |
| }, | |
| { | |
| "epoch": 0.6495740170393184, | |
| "grad_norm": 0.05824848510516177, | |
| "learning_rate": 5.7986714907356614e-06, | |
| "loss": 0.4586543083190918, | |
| "memory(GiB)": 77.63, | |
| "step": 2030, | |
| "token_acc": 0.8852591792656588, | |
| "train_speed(iter/s)": 0.029498 | |
| }, | |
| { | |
| "epoch": 0.6511739530418783, | |
| "grad_norm": 0.06066761562869553, | |
| "learning_rate": 5.751606397077703e-06, | |
| "loss": 0.44632205963134763, | |
| "memory(GiB)": 77.63, | |
| "step": 2035, | |
| "token_acc": 0.8871352785145888, | |
| "train_speed(iter/s)": 0.029494 | |
| }, | |
| { | |
| "epoch": 0.6527738890444382, | |
| "grad_norm": 0.055201144436432543, | |
| "learning_rate": 5.704655858081268e-06, | |
| "loss": 0.43164916038513185, | |
| "memory(GiB)": 77.63, | |
| "step": 2040, | |
| "token_acc": 0.8937977909940527, | |
| "train_speed(iter/s)": 0.029496 | |
| }, | |
| { | |
| "epoch": 0.6543738250469981, | |
| "grad_norm": 0.05987076771844116, | |
| "learning_rate": 5.6578211397314765e-06, | |
| "loss": 0.4560856819152832, | |
| "memory(GiB)": 77.63, | |
| "step": 2045, | |
| "token_acc": 0.8462914545204349, | |
| "train_speed(iter/s)": 0.029507 | |
| }, | |
| { | |
| "epoch": 0.6559737610495581, | |
| "grad_norm": 0.05549604363093839, | |
| "learning_rate": 5.611103504890444e-06, | |
| "loss": 0.44809746742248535, | |
| "memory(GiB)": 77.63, | |
| "step": 2050, | |
| "token_acc": 0.8783254279232832, | |
| "train_speed(iter/s)": 0.029503 | |
| }, | |
| { | |
| "epoch": 0.6575736970521179, | |
| "grad_norm": 0.060605872447174955, | |
| "learning_rate": 5.564504213263205e-06, | |
| "loss": 0.43492536544799804, | |
| "memory(GiB)": 77.63, | |
| "step": 2055, | |
| "token_acc": 0.8383036405886909, | |
| "train_speed(iter/s)": 0.029515 | |
| }, | |
| { | |
| "epoch": 0.6591736330546778, | |
| "grad_norm": 0.05003885513998493, | |
| "learning_rate": 5.5180245213637785e-06, | |
| "loss": 0.44741315841674806, | |
| "memory(GiB)": 77.63, | |
| "step": 2060, | |
| "token_acc": 0.8784857874174862, | |
| "train_speed(iter/s)": 0.029515 | |
| }, | |
| { | |
| "epoch": 0.6607735690572377, | |
| "grad_norm": 0.053248832099036005, | |
| "learning_rate": 5.4716656824812505e-06, | |
| "loss": 0.4469279766082764, | |
| "memory(GiB)": 77.63, | |
| "step": 2065, | |
| "token_acc": 0.8853107344632768, | |
| "train_speed(iter/s)": 0.029507 | |
| }, | |
| { | |
| "epoch": 0.6623735050597976, | |
| "grad_norm": 0.050937098304756526, | |
| "learning_rate": 5.425428946646016e-06, | |
| "loss": 0.44948582649230956, | |
| "memory(GiB)": 77.63, | |
| "step": 2070, | |
| "token_acc": 0.8934210526315789, | |
| "train_speed(iter/s)": 0.029526 | |
| }, | |
| { | |
| "epoch": 0.6639734410623576, | |
| "grad_norm": 0.050018563981163396, | |
| "learning_rate": 5.379315560596038e-06, | |
| "loss": 0.4475410461425781, | |
| "memory(GiB)": 77.63, | |
| "step": 2075, | |
| "token_acc": 0.8478816857555876, | |
| "train_speed(iter/s)": 0.029518 | |
| }, | |
| { | |
| "epoch": 0.6655733770649174, | |
| "grad_norm": 0.05745892430696422, | |
| "learning_rate": 5.333326767743263e-06, | |
| "loss": 0.45008225440979005, | |
| "memory(GiB)": 77.63, | |
| "step": 2080, | |
| "token_acc": 0.8264099454214675, | |
| "train_speed(iter/s)": 0.029515 | |
| }, | |
| { | |
| "epoch": 0.6671733130674773, | |
| "grad_norm": 0.05536507134607956, | |
| "learning_rate": 5.287463808140069e-06, | |
| "loss": 0.4393789291381836, | |
| "memory(GiB)": 77.63, | |
| "step": 2085, | |
| "token_acc": 0.8450008816787162, | |
| "train_speed(iter/s)": 0.029527 | |
| }, | |
| { | |
| "epoch": 0.6687732490700372, | |
| "grad_norm": 0.06142641017026178, | |
| "learning_rate": 5.241727918445836e-06, | |
| "loss": 0.4437687873840332, | |
| "memory(GiB)": 77.63, | |
| "step": 2090, | |
| "token_acc": 0.8837277242185217, | |
| "train_speed(iter/s)": 0.02952 | |
| }, | |
| { | |
| "epoch": 0.6703731850725971, | |
| "grad_norm": 0.05426196603270913, | |
| "learning_rate": 5.1961203318936116e-06, | |
| "loss": 0.4427367687225342, | |
| "memory(GiB)": 77.63, | |
| "step": 2095, | |
| "token_acc": 0.856048805815161, | |
| "train_speed(iter/s)": 0.029524 | |
| }, | |
| { | |
| "epoch": 0.6719731210751569, | |
| "grad_norm": 0.054169398190345976, | |
| "learning_rate": 5.1506422782568345e-06, | |
| "loss": 0.4520686626434326, | |
| "memory(GiB)": 77.63, | |
| "step": 2100, | |
| "token_acc": 0.8747133027522935, | |
| "train_speed(iter/s)": 0.029529 | |
| }, | |
| { | |
| "epoch": 0.6735730570777169, | |
| "grad_norm": 0.054436537230257924, | |
| "learning_rate": 5.105294983816203e-06, | |
| "loss": 0.44482645988464353, | |
| "memory(GiB)": 77.63, | |
| "step": 2105, | |
| "token_acc": 0.8637377049180328, | |
| "train_speed(iter/s)": 0.029517 | |
| }, | |
| { | |
| "epoch": 0.6751729930802768, | |
| "grad_norm": 0.05860088154390529, | |
| "learning_rate": 5.060079671326577e-06, | |
| "loss": 0.44719686508178713, | |
| "memory(GiB)": 77.63, | |
| "step": 2110, | |
| "token_acc": 0.8593150866058442, | |
| "train_speed(iter/s)": 0.029529 | |
| }, | |
| { | |
| "epoch": 0.6767729290828367, | |
| "grad_norm": 0.05264024149284518, | |
| "learning_rate": 5.014997559984045e-06, | |
| "loss": 0.43972039222717285, | |
| "memory(GiB)": 77.63, | |
| "step": 2115, | |
| "token_acc": 0.8533221194280909, | |
| "train_speed(iter/s)": 0.029526 | |
| }, | |
| { | |
| "epoch": 0.6783728650853966, | |
| "grad_norm": 0.0534652970629265, | |
| "learning_rate": 4.970049865393009e-06, | |
| "loss": 0.4468375205993652, | |
| "memory(GiB)": 77.63, | |
| "step": 2120, | |
| "token_acc": 0.8628782287822878, | |
| "train_speed(iter/s)": 0.029518 | |
| }, | |
| { | |
| "epoch": 0.6799728010879564, | |
| "grad_norm": 0.05246927821047006, | |
| "learning_rate": 4.925237799533445e-06, | |
| "loss": 0.4498266696929932, | |
| "memory(GiB)": 77.63, | |
| "step": 2125, | |
| "token_acc": 0.9048205760049284, | |
| "train_speed(iter/s)": 0.029533 | |
| }, | |
| { | |
| "epoch": 0.6815727370905164, | |
| "grad_norm": 0.05104237083350841, | |
| "learning_rate": 4.880562570728188e-06, | |
| "loss": 0.4389338970184326, | |
| "memory(GiB)": 77.63, | |
| "step": 2130, | |
| "token_acc": 0.8844444444444445, | |
| "train_speed(iter/s)": 0.029525 | |
| }, | |
| { | |
| "epoch": 0.6831726730930763, | |
| "grad_norm": 0.05297787940328326, | |
| "learning_rate": 4.836025383610382e-06, | |
| "loss": 0.4495584487915039, | |
| "memory(GiB)": 77.63, | |
| "step": 2135, | |
| "token_acc": 0.8647426233038984, | |
| "train_speed(iter/s)": 0.029524 | |
| }, | |
| { | |
| "epoch": 0.6847726090956362, | |
| "grad_norm": 0.05092547333787791, | |
| "learning_rate": 4.791627439090975e-06, | |
| "loss": 0.4421692848205566, | |
| "memory(GiB)": 77.63, | |
| "step": 2140, | |
| "token_acc": 0.8828041384231181, | |
| "train_speed(iter/s)": 0.029534 | |
| }, | |
| { | |
| "epoch": 0.6863725450981961, | |
| "grad_norm": 0.053418572817851825, | |
| "learning_rate": 4.74736993432634e-06, | |
| "loss": 0.44208922386169436, | |
| "memory(GiB)": 77.63, | |
| "step": 2145, | |
| "token_acc": 0.8888520238885202, | |
| "train_speed(iter/s)": 0.029522 | |
| }, | |
| { | |
| "epoch": 0.6879724811007559, | |
| "grad_norm": 0.053760421496406786, | |
| "learning_rate": 4.703254062686017e-06, | |
| "loss": 0.4469425201416016, | |
| "memory(GiB)": 77.63, | |
| "step": 2150, | |
| "token_acc": 0.8764145324597975, | |
| "train_speed(iter/s)": 0.02953 | |
| }, | |
| { | |
| "epoch": 0.6895724171033158, | |
| "grad_norm": 0.05280613332203591, | |
| "learning_rate": 4.6592810137205e-06, | |
| "loss": 0.45023741722106936, | |
| "memory(GiB)": 77.63, | |
| "step": 2155, | |
| "token_acc": 0.8968010517090271, | |
| "train_speed(iter/s)": 0.029531 | |
| }, | |
| { | |
| "epoch": 0.6911723531058758, | |
| "grad_norm": 0.05438679052798784, | |
| "learning_rate": 4.615451973129196e-06, | |
| "loss": 0.4470167636871338, | |
| "memory(GiB)": 77.63, | |
| "step": 2160, | |
| "token_acc": 0.8761696818465378, | |
| "train_speed(iter/s)": 0.02952 | |
| }, | |
| { | |
| "epoch": 0.6927722891084357, | |
| "grad_norm": 0.056426544102266905, | |
| "learning_rate": 4.571768122728421e-06, | |
| "loss": 0.4486443042755127, | |
| "memory(GiB)": 77.63, | |
| "step": 2165, | |
| "token_acc": 0.8781996587030717, | |
| "train_speed(iter/s)": 0.029533 | |
| }, | |
| { | |
| "epoch": 0.6943722251109956, | |
| "grad_norm": 0.05337656902490804, | |
| "learning_rate": 4.528230640419562e-06, | |
| "loss": 0.4497722625732422, | |
| "memory(GiB)": 77.63, | |
| "step": 2170, | |
| "token_acc": 0.8571428571428571, | |
| "train_speed(iter/s)": 0.029522 | |
| }, | |
| { | |
| "epoch": 0.6959721611135554, | |
| "grad_norm": 0.054129658100736014, | |
| "learning_rate": 4.4848407001572945e-06, | |
| "loss": 0.44121665954589845, | |
| "memory(GiB)": 77.63, | |
| "step": 2175, | |
| "token_acc": 0.8674278464954012, | |
| "train_speed(iter/s)": 0.029518 | |
| }, | |
| { | |
| "epoch": 0.6975720971161153, | |
| "grad_norm": 0.05332136363084243, | |
| "learning_rate": 4.441599471917946e-06, | |
| "loss": 0.43872866630554197, | |
| "memory(GiB)": 77.63, | |
| "step": 2180, | |
| "token_acc": 0.8575067664384652, | |
| "train_speed(iter/s)": 0.029529 | |
| }, | |
| { | |
| "epoch": 0.6991720331186753, | |
| "grad_norm": 0.06093731322456081, | |
| "learning_rate": 4.398508121667925e-06, | |
| "loss": 0.42902402877807616, | |
| "memory(GiB)": 77.63, | |
| "step": 2185, | |
| "token_acc": 0.8526187576126675, | |
| "train_speed(iter/s)": 0.029519 | |
| }, | |
| { | |
| "epoch": 0.7007719691212352, | |
| "grad_norm": 0.05959591977220614, | |
| "learning_rate": 4.355567811332311e-06, | |
| "loss": 0.44504075050354003, | |
| "memory(GiB)": 77.63, | |
| "step": 2190, | |
| "token_acc": 0.8604511878618487, | |
| "train_speed(iter/s)": 0.029519 | |
| }, | |
| { | |
| "epoch": 0.702371905123795, | |
| "grad_norm": 0.052994813328955795, | |
| "learning_rate": 4.312779698763493e-06, | |
| "loss": 0.4408130168914795, | |
| "memory(GiB)": 77.63, | |
| "step": 2195, | |
| "token_acc": 0.8787728847105394, | |
| "train_speed(iter/s)": 0.029526 | |
| }, | |
| { | |
| "epoch": 0.7039718411263549, | |
| "grad_norm": 0.05402320661185779, | |
| "learning_rate": 4.270144937709981e-06, | |
| "loss": 0.4396658897399902, | |
| "memory(GiB)": 77.63, | |
| "step": 2200, | |
| "token_acc": 0.8677248677248677, | |
| "train_speed(iter/s)": 0.029515 | |
| }, | |
| { | |
| "epoch": 0.7055717771289148, | |
| "grad_norm": 0.053270864469091045, | |
| "learning_rate": 4.227664677785264e-06, | |
| "loss": 0.4493250846862793, | |
| "memory(GiB)": 77.63, | |
| "step": 2205, | |
| "token_acc": 0.8801270542742715, | |
| "train_speed(iter/s)": 0.029523 | |
| }, | |
| { | |
| "epoch": 0.7071717131314748, | |
| "grad_norm": 0.06248819511930574, | |
| "learning_rate": 4.1853400644368395e-06, | |
| "loss": 0.44740095138549807, | |
| "memory(GiB)": 77.63, | |
| "step": 2210, | |
| "token_acc": 0.8988747408942849, | |
| "train_speed(iter/s)": 0.029514 | |
| }, | |
| { | |
| "epoch": 0.7087716491340347, | |
| "grad_norm": 0.05807018681919018, | |
| "learning_rate": 4.143172238915302e-06, | |
| "loss": 0.4508991241455078, | |
| "memory(GiB)": 77.63, | |
| "step": 2215, | |
| "token_acc": 0.8739803562510404, | |
| "train_speed(iter/s)": 0.029506 | |
| }, | |
| { | |
| "epoch": 0.7103715851365945, | |
| "grad_norm": 0.05376168037907147, | |
| "learning_rate": 4.101162338243595e-06, | |
| "loss": 0.4486696243286133, | |
| "memory(GiB)": 77.63, | |
| "step": 2220, | |
| "token_acc": 0.8750778169744761, | |
| "train_speed(iter/s)": 0.029514 | |
| }, | |
| { | |
| "epoch": 0.7119715211391544, | |
| "grad_norm": 0.05460944158847028, | |
| "learning_rate": 4.059311495186338e-06, | |
| "loss": 0.4484865188598633, | |
| "memory(GiB)": 77.63, | |
| "step": 2225, | |
| "token_acc": 0.8524350054924936, | |
| "train_speed(iter/s)": 0.029506 | |
| }, | |
| { | |
| "epoch": 0.7135714571417143, | |
| "grad_norm": 0.05716955035585288, | |
| "learning_rate": 4.017620838219276e-06, | |
| "loss": 0.44258599281311034, | |
| "memory(GiB)": 77.63, | |
| "step": 2230, | |
| "token_acc": 0.8597191629955947, | |
| "train_speed(iter/s)": 0.029504 | |
| }, | |
| { | |
| "epoch": 0.7151713931442742, | |
| "grad_norm": 0.05984913995816041, | |
| "learning_rate": 3.9760914914988716e-06, | |
| "loss": 0.4547589778900146, | |
| "memory(GiB)": 77.63, | |
| "step": 2235, | |
| "token_acc": 0.8679617117117117, | |
| "train_speed(iter/s)": 0.029511 | |
| }, | |
| { | |
| "epoch": 0.7167713291468342, | |
| "grad_norm": 0.05686589162715874, | |
| "learning_rate": 3.93472457483197e-06, | |
| "loss": 0.4416301727294922, | |
| "memory(GiB)": 77.63, | |
| "step": 2240, | |
| "token_acc": 0.826577064816822, | |
| "train_speed(iter/s)": 0.029498 | |
| }, | |
| { | |
| "epoch": 0.718371265149394, | |
| "grad_norm": 0.05780707586931182, | |
| "learning_rate": 3.893521203645618e-06, | |
| "loss": 0.45052361488342285, | |
| "memory(GiB)": 77.63, | |
| "step": 2245, | |
| "token_acc": 0.8836182062608028, | |
| "train_speed(iter/s)": 0.0295 | |
| }, | |
| { | |
| "epoch": 0.7199712011519539, | |
| "grad_norm": 0.049110615928360885, | |
| "learning_rate": 3.852482488956992e-06, | |
| "loss": 0.4427218437194824, | |
| "memory(GiB)": 77.63, | |
| "step": 2250, | |
| "token_acc": 0.8621255642183012, | |
| "train_speed(iter/s)": 0.029501 | |
| }, | |
| { | |
| "epoch": 0.7199712011519539, | |
| "eval_loss": 0.659950852394104, | |
| "eval_runtime": 108.3142, | |
| "eval_samples_per_second": 185.461, | |
| "eval_steps_per_second": 0.932, | |
| "eval_token_acc": 0.8716289458342705, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.7215711371545138, | |
| "grad_norm": 0.04899882607919235, | |
| "learning_rate": 3.8116095373434204e-06, | |
| "loss": 0.4487879753112793, | |
| "memory(GiB)": 77.63, | |
| "step": 2255, | |
| "token_acc": 0.8912671818368324, | |
| "train_speed(iter/s)": 0.029473 | |
| }, | |
| { | |
| "epoch": 0.7231710731570737, | |
| "grad_norm": 0.05080548488435112, | |
| "learning_rate": 3.7709034509125706e-06, | |
| "loss": 0.44452829360961915, | |
| "memory(GiB)": 77.63, | |
| "step": 2260, | |
| "token_acc": 0.8442477876106195, | |
| "train_speed(iter/s)": 0.029482 | |
| }, | |
| { | |
| "epoch": 0.7247710091596337, | |
| "grad_norm": 0.048986009146357284, | |
| "learning_rate": 3.7303653272727057e-06, | |
| "loss": 0.4472095012664795, | |
| "memory(GiB)": 77.63, | |
| "step": 2265, | |
| "token_acc": 0.870567815521944, | |
| "train_speed(iter/s)": 0.029495 | |
| }, | |
| { | |
| "epoch": 0.7263709451621935, | |
| "grad_norm": 0.05152412916361422, | |
| "learning_rate": 3.689996259503116e-06, | |
| "loss": 0.440493106842041, | |
| "memory(GiB)": 77.63, | |
| "step": 2270, | |
| "token_acc": 0.8795436455293181, | |
| "train_speed(iter/s)": 0.029499 | |
| }, | |
| { | |
| "epoch": 0.7279708811647534, | |
| "grad_norm": 0.055480142184644934, | |
| "learning_rate": 3.6497973361246153e-06, | |
| "loss": 0.4417555809020996, | |
| "memory(GiB)": 77.63, | |
| "step": 2275, | |
| "token_acc": 0.8660460713158725, | |
| "train_speed(iter/s)": 0.029499 | |
| }, | |
| { | |
| "epoch": 0.7295708171673133, | |
| "grad_norm": 0.05625540509082736, | |
| "learning_rate": 3.609769641070221e-06, | |
| "loss": 0.4407214164733887, | |
| "memory(GiB)": 77.63, | |
| "step": 2280, | |
| "token_acc": 0.8890608875128999, | |
| "train_speed(iter/s)": 0.029513 | |
| }, | |
| { | |
| "epoch": 0.7311707531698732, | |
| "grad_norm": 0.05002588428206622, | |
| "learning_rate": 3.569914253655896e-06, | |
| "loss": 0.4413386344909668, | |
| "memory(GiB)": 77.63, | |
| "step": 2285, | |
| "token_acc": 0.8921049390319005, | |
| "train_speed(iter/s)": 0.029513 | |
| }, | |
| { | |
| "epoch": 0.7327706891724332, | |
| "grad_norm": 0.05088814815973685, | |
| "learning_rate": 3.530232248551466e-06, | |
| "loss": 0.4507819652557373, | |
| "memory(GiB)": 77.63, | |
| "step": 2290, | |
| "token_acc": 0.8278411830895355, | |
| "train_speed(iter/s)": 0.029513 | |
| }, | |
| { | |
| "epoch": 0.734370625174993, | |
| "grad_norm": 0.05399937134620822, | |
| "learning_rate": 3.4907246957516416e-06, | |
| "loss": 0.4447961330413818, | |
| "memory(GiB)": 77.63, | |
| "step": 2295, | |
| "token_acc": 0.8888263967004124, | |
| "train_speed(iter/s)": 0.029522 | |
| }, | |
| { | |
| "epoch": 0.7359705611775529, | |
| "grad_norm": 0.06200035405309708, | |
| "learning_rate": 3.4513926605471504e-06, | |
| "loss": 0.45868444442749023, | |
| "memory(GiB)": 77.63, | |
| "step": 2300, | |
| "token_acc": 0.8513141426783479, | |
| "train_speed(iter/s)": 0.02952 | |
| }, | |
| { | |
| "epoch": 0.7375704971801128, | |
| "grad_norm": 0.055876255912378235, | |
| "learning_rate": 3.412237203496036e-06, | |
| "loss": 0.4431456089019775, | |
| "memory(GiB)": 77.63, | |
| "step": 2305, | |
| "token_acc": 0.8651997041420119, | |
| "train_speed(iter/s)": 0.029532 | |
| }, | |
| { | |
| "epoch": 0.7391704331826727, | |
| "grad_norm": 0.06032844036632358, | |
| "learning_rate": 3.3732593803950354e-06, | |
| "loss": 0.4452229976654053, | |
| "memory(GiB)": 77.63, | |
| "step": 2310, | |
| "token_acc": 0.8915232899706252, | |
| "train_speed(iter/s)": 0.02953 | |
| }, | |
| { | |
| "epoch": 0.7407703691852325, | |
| "grad_norm": 0.05255216039270682, | |
| "learning_rate": 3.3344602422511343e-06, | |
| "loss": 0.4414207458496094, | |
| "memory(GiB)": 77.63, | |
| "step": 2315, | |
| "token_acc": 0.8901802257032171, | |
| "train_speed(iter/s)": 0.029522 | |
| }, | |
| { | |
| "epoch": 0.7423703051877925, | |
| "grad_norm": 0.05535966142690852, | |
| "learning_rate": 3.2958408352532055e-06, | |
| "loss": 0.43938393592834474, | |
| "memory(GiB)": 77.63, | |
| "step": 2320, | |
| "token_acc": 0.8354077253218885, | |
| "train_speed(iter/s)": 0.029536 | |
| }, | |
| { | |
| "epoch": 0.7439702411903524, | |
| "grad_norm": 0.0505418855319798, | |
| "learning_rate": 3.257402200743821e-06, | |
| "loss": 0.44445362091064455, | |
| "memory(GiB)": 77.63, | |
| "step": 2325, | |
| "token_acc": 0.864262790258637, | |
| "train_speed(iter/s)": 0.02953 | |
| }, | |
| { | |
| "epoch": 0.7455701771929123, | |
| "grad_norm": 0.0563107101835597, | |
| "learning_rate": 3.2191453751911505e-06, | |
| "loss": 0.45569453239440916, | |
| "memory(GiB)": 77.63, | |
| "step": 2330, | |
| "token_acc": 0.8784313725490196, | |
| "train_speed(iter/s)": 0.029523 | |
| }, | |
| { | |
| "epoch": 0.7471701131954722, | |
| "grad_norm": 0.05000152007266613, | |
| "learning_rate": 3.1810713901610367e-06, | |
| "loss": 0.4395348072052002, | |
| "memory(GiB)": 77.63, | |
| "step": 2335, | |
| "token_acc": 0.8867111781175964, | |
| "train_speed(iter/s)": 0.029536 | |
| }, | |
| { | |
| "epoch": 0.748770049198032, | |
| "grad_norm": 0.057169590375126145, | |
| "learning_rate": 3.1431812722891598e-06, | |
| "loss": 0.4397278785705566, | |
| "memory(GiB)": 77.63, | |
| "step": 2340, | |
| "token_acc": 0.8577532891037895, | |
| "train_speed(iter/s)": 0.029529 | |
| }, | |
| { | |
| "epoch": 0.750369985200592, | |
| "grad_norm": 0.05730745865195846, | |
| "learning_rate": 3.1054760432533626e-06, | |
| "loss": 0.45998029708862304, | |
| "memory(GiB)": 77.63, | |
| "step": 2345, | |
| "token_acc": 0.8845755097339016, | |
| "train_speed(iter/s)": 0.029534 | |
| }, | |
| { | |
| "epoch": 0.7519699212031519, | |
| "grad_norm": 0.05180470840824953, | |
| "learning_rate": 3.0679567197461135e-06, | |
| "loss": 0.45008273124694825, | |
| "memory(GiB)": 77.63, | |
| "step": 2350, | |
| "token_acc": 0.8394425931535898, | |
| "train_speed(iter/s)": 0.029537 | |
| }, | |
| { | |
| "epoch": 0.7535698572057118, | |
| "grad_norm": 0.06025883780673481, | |
| "learning_rate": 3.0306243134470668e-06, | |
| "loss": 0.4444745540618896, | |
| "memory(GiB)": 77.63, | |
| "step": 2355, | |
| "token_acc": 0.889631386074585, | |
| "train_speed(iter/s)": 0.02953 | |
| }, | |
| { | |
| "epoch": 0.7551697932082717, | |
| "grad_norm": 0.05199872680450009, | |
| "learning_rate": 2.993479830995815e-06, | |
| "loss": 0.451768159866333, | |
| "memory(GiB)": 77.63, | |
| "step": 2360, | |
| "token_acc": 0.8736520199581522, | |
| "train_speed(iter/s)": 0.029541 | |
| }, | |
| { | |
| "epoch": 0.7567697292108315, | |
| "grad_norm": 0.05489027404588469, | |
| "learning_rate": 2.9565242739647115e-06, | |
| "loss": 0.4442115306854248, | |
| "memory(GiB)": 77.63, | |
| "step": 2365, | |
| "token_acc": 0.8865552903739061, | |
| "train_speed(iter/s)": 0.029538 | |
| }, | |
| { | |
| "epoch": 0.7583696652133914, | |
| "grad_norm": 0.06334021131539457, | |
| "learning_rate": 2.919758638831893e-06, | |
| "loss": 0.4570741653442383, | |
| "memory(GiB)": 77.63, | |
| "step": 2370, | |
| "token_acc": 0.8652606912712361, | |
| "train_speed(iter/s)": 0.029531 | |
| }, | |
| { | |
| "epoch": 0.7599696012159514, | |
| "grad_norm": 0.053831666314624105, | |
| "learning_rate": 2.8831839169543998e-06, | |
| "loss": 0.44495415687561035, | |
| "memory(GiB)": 77.63, | |
| "step": 2375, | |
| "token_acc": 0.8756476683937824, | |
| "train_speed(iter/s)": 0.029541 | |
| }, | |
| { | |
| "epoch": 0.7615695372185113, | |
| "grad_norm": 0.0527583973457582, | |
| "learning_rate": 2.84680109454143e-06, | |
| "loss": 0.4472104549407959, | |
| "memory(GiB)": 77.63, | |
| "step": 2380, | |
| "token_acc": 0.8725328947368421, | |
| "train_speed(iter/s)": 0.029536 | |
| }, | |
| { | |
| "epoch": 0.7631694732210712, | |
| "grad_norm": 0.058941021305098804, | |
| "learning_rate": 2.810611152627777e-06, | |
| "loss": 0.4499720573425293, | |
| "memory(GiB)": 77.63, | |
| "step": 2385, | |
| "token_acc": 0.8632213889794588, | |
| "train_speed(iter/s)": 0.029535 | |
| }, | |
| { | |
| "epoch": 0.764769409223631, | |
| "grad_norm": 0.05393008170855123, | |
| "learning_rate": 2.774615067047346e-06, | |
| "loss": 0.43872222900390623, | |
| "memory(GiB)": 77.63, | |
| "step": 2390, | |
| "token_acc": 0.8742202234150588, | |
| "train_speed(iter/s)": 0.02954 | |
| }, | |
| { | |
| "epoch": 0.7663693452261909, | |
| "grad_norm": 0.0556417645809335, | |
| "learning_rate": 2.738813808406866e-06, | |
| "loss": 0.4399220943450928, | |
| "memory(GiB)": 77.63, | |
| "step": 2395, | |
| "token_acc": 0.8997599039615847, | |
| "train_speed(iter/s)": 0.029531 | |
| }, | |
| { | |
| "epoch": 0.7679692812287509, | |
| "grad_norm": 0.05379432234404155, | |
| "learning_rate": 2.7032083420597e-06, | |
| "loss": 0.4382453441619873, | |
| "memory(GiB)": 77.63, | |
| "step": 2400, | |
| "token_acc": 0.8875784668061633, | |
| "train_speed(iter/s)": 0.029541 | |
| }, | |
| { | |
| "epoch": 0.7695692172313108, | |
| "grad_norm": 0.05806842163630925, | |
| "learning_rate": 2.667799628079829e-06, | |
| "loss": 0.44454326629638674, | |
| "memory(GiB)": 77.63, | |
| "step": 2405, | |
| "token_acc": 0.8880662020905923, | |
| "train_speed(iter/s)": 0.029535 | |
| }, | |
| { | |
| "epoch": 0.7711691532338707, | |
| "grad_norm": 0.06148704112133217, | |
| "learning_rate": 2.6325886212359496e-06, | |
| "loss": 0.43945813179016113, | |
| "memory(GiB)": 77.63, | |
| "step": 2410, | |
| "token_acc": 0.8767772511848341, | |
| "train_speed(iter/s)": 0.029527 | |
| }, | |
| { | |
| "epoch": 0.7727690892364305, | |
| "grad_norm": 0.056530065759685846, | |
| "learning_rate": 2.5975762709657506e-06, | |
| "loss": 0.4438450813293457, | |
| "memory(GiB)": 77.63, | |
| "step": 2415, | |
| "token_acc": 0.8570975416336241, | |
| "train_speed(iter/s)": 0.029538 | |
| }, | |
| { | |
| "epoch": 0.7743690252389904, | |
| "grad_norm": 0.056107845444701834, | |
| "learning_rate": 2.5627635213502832e-06, | |
| "loss": 0.43836054801940916, | |
| "memory(GiB)": 77.63, | |
| "step": 2420, | |
| "token_acc": 0.8966822253059165, | |
| "train_speed(iter/s)": 0.029532 | |
| }, | |
| { | |
| "epoch": 0.7759689612415503, | |
| "grad_norm": 0.05796065696017405, | |
| "learning_rate": 2.528151311088537e-06, | |
| "loss": 0.4400279998779297, | |
| "memory(GiB)": 77.63, | |
| "step": 2425, | |
| "token_acc": 0.8552805280528053, | |
| "train_speed(iter/s)": 0.029526 | |
| }, | |
| { | |
| "epoch": 0.7775688972441103, | |
| "grad_norm": 0.05418546630146028, | |
| "learning_rate": 2.4937405734720964e-06, | |
| "loss": 0.44541444778442385, | |
| "memory(GiB)": 77.63, | |
| "step": 2430, | |
| "token_acc": 0.8620764552562988, | |
| "train_speed(iter/s)": 0.029532 | |
| }, | |
| { | |
| "epoch": 0.7791688332466701, | |
| "grad_norm": 0.05654250831277805, | |
| "learning_rate": 2.459532236360007e-06, | |
| "loss": 0.43491110801696775, | |
| "memory(GiB)": 77.63, | |
| "step": 2435, | |
| "token_acc": 0.8570184983677911, | |
| "train_speed(iter/s)": 0.029522 | |
| }, | |
| { | |
| "epoch": 0.78076876924923, | |
| "grad_norm": 0.05526372242621089, | |
| "learning_rate": 2.4255272221537295e-06, | |
| "loss": 0.4378859043121338, | |
| "memory(GiB)": 77.63, | |
| "step": 2440, | |
| "token_acc": 0.8631236857197476, | |
| "train_speed(iter/s)": 0.029526 | |
| }, | |
| { | |
| "epoch": 0.7823687052517899, | |
| "grad_norm": 0.05404315424969483, | |
| "learning_rate": 2.391726447772279e-06, | |
| "loss": 0.45857391357421873, | |
| "memory(GiB)": 77.63, | |
| "step": 2445, | |
| "token_acc": 0.8634816932081122, | |
| "train_speed(iter/s)": 0.029527 | |
| }, | |
| { | |
| "epoch": 0.7839686412543498, | |
| "grad_norm": 0.05765113554061621, | |
| "learning_rate": 2.3581308246275103e-06, | |
| "loss": 0.4473139762878418, | |
| "memory(GiB)": 77.63, | |
| "step": 2450, | |
| "token_acc": 0.8979846898922044, | |
| "train_speed(iter/s)": 0.029518 | |
| }, | |
| { | |
| "epoch": 0.7855685772569098, | |
| "grad_norm": 0.058353221842389495, | |
| "learning_rate": 2.324741258599521e-06, | |
| "loss": 0.44444866180419923, | |
| "memory(GiB)": 77.63, | |
| "step": 2455, | |
| "token_acc": 0.8648913576213038, | |
| "train_speed(iter/s)": 0.029527 | |
| }, | |
| { | |
| "epoch": 0.7871685132594696, | |
| "grad_norm": 0.05309289512529438, | |
| "learning_rate": 2.29155865001225e-06, | |
| "loss": 0.43857607841491697, | |
| "memory(GiB)": 77.63, | |
| "step": 2460, | |
| "token_acc": 0.894580549368968, | |
| "train_speed(iter/s)": 0.029521 | |
| }, | |
| { | |
| "epoch": 0.7887684492620295, | |
| "grad_norm": 0.0540412312893473, | |
| "learning_rate": 2.2585838936091753e-06, | |
| "loss": 0.43953213691711424, | |
| "memory(GiB)": 77.63, | |
| "step": 2465, | |
| "token_acc": 0.8868672731513879, | |
| "train_speed(iter/s)": 0.029515 | |
| }, | |
| { | |
| "epoch": 0.7903683852645894, | |
| "grad_norm": 0.05880520800742855, | |
| "learning_rate": 2.225817878529214e-06, | |
| "loss": 0.4457580089569092, | |
| "memory(GiB)": 77.63, | |
| "step": 2470, | |
| "token_acc": 0.8630282437884901, | |
| "train_speed(iter/s)": 0.029525 | |
| }, | |
| { | |
| "epoch": 0.7919683212671493, | |
| "grad_norm": 0.058328885138964066, | |
| "learning_rate": 2.1932614882827196e-06, | |
| "loss": 0.4424918174743652, | |
| "memory(GiB)": 77.63, | |
| "step": 2475, | |
| "token_acc": 0.8814697747925722, | |
| "train_speed(iter/s)": 0.029517 | |
| }, | |
| { | |
| "epoch": 0.7935682572697093, | |
| "grad_norm": 0.05685263085809571, | |
| "learning_rate": 2.160915600727688e-06, | |
| "loss": 0.43921732902526855, | |
| "memory(GiB)": 77.63, | |
| "step": 2480, | |
| "token_acc": 0.913681738109219, | |
| "train_speed(iter/s)": 0.029516 | |
| }, | |
| { | |
| "epoch": 0.7951681932722691, | |
| "grad_norm": 0.056639845561812056, | |
| "learning_rate": 2.1287810880460636e-06, | |
| "loss": 0.44060502052307127, | |
| "memory(GiB)": 77.63, | |
| "step": 2485, | |
| "token_acc": 0.8829075425790754, | |
| "train_speed(iter/s)": 0.02952 | |
| }, | |
| { | |
| "epoch": 0.796768129274829, | |
| "grad_norm": 0.05230960490349676, | |
| "learning_rate": 2.0968588167202265e-06, | |
| "loss": 0.43935480117797854, | |
| "memory(GiB)": 77.63, | |
| "step": 2490, | |
| "token_acc": 0.8856997455470738, | |
| "train_speed(iter/s)": 0.029511 | |
| }, | |
| { | |
| "epoch": 0.7983680652773889, | |
| "grad_norm": 0.05305183045142263, | |
| "learning_rate": 2.0651496475096455e-06, | |
| "loss": 0.4360368728637695, | |
| "memory(GiB)": 77.63, | |
| "step": 2495, | |
| "token_acc": 0.8394655704008221, | |
| "train_speed(iter/s)": 0.029517 | |
| }, | |
| { | |
| "epoch": 0.7999680012799488, | |
| "grad_norm": 0.05620012484228566, | |
| "learning_rate": 2.03365443542764e-06, | |
| "loss": 0.44507203102111814, | |
| "memory(GiB)": 77.63, | |
| "step": 2500, | |
| "token_acc": 0.8857098429482195, | |
| "train_speed(iter/s)": 0.029515 | |
| }, | |
| { | |
| "epoch": 0.7999680012799488, | |
| "eval_loss": 0.6586322784423828, | |
| "eval_runtime": 105.1966, | |
| "eval_samples_per_second": 190.957, | |
| "eval_steps_per_second": 0.96, | |
| "eval_token_acc": 0.8721292963419328, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.8015679372825087, | |
| "grad_norm": 0.05541548293022976, | |
| "learning_rate": 2.0023740297183536e-06, | |
| "loss": 0.44654192924499514, | |
| "memory(GiB)": 77.63, | |
| "step": 2505, | |
| "token_acc": 0.8819252077562327, | |
| "train_speed(iter/s)": 0.029489 | |
| }, | |
| { | |
| "epoch": 0.8031678732850686, | |
| "grad_norm": 0.057374199419424524, | |
| "learning_rate": 1.971309273833828e-06, | |
| "loss": 0.44596128463745116, | |
| "memory(GiB)": 77.63, | |
| "step": 2510, | |
| "token_acc": 0.855553561815898, | |
| "train_speed(iter/s)": 0.029494 | |
| }, | |
| { | |
| "epoch": 0.8047678092876285, | |
| "grad_norm": 0.05297595418967434, | |
| "learning_rate": 1.940461005411288e-06, | |
| "loss": 0.45099148750305174, | |
| "memory(GiB)": 77.63, | |
| "step": 2515, | |
| "token_acc": 0.8958361962347121, | |
| "train_speed(iter/s)": 0.029503 | |
| }, | |
| { | |
| "epoch": 0.8063677452901884, | |
| "grad_norm": 0.056714681257095015, | |
| "learning_rate": 1.9098300562505266e-06, | |
| "loss": 0.4423669338226318, | |
| "memory(GiB)": 77.63, | |
| "step": 2520, | |
| "token_acc": 0.879045996592845, | |
| "train_speed(iter/s)": 0.029507 | |
| }, | |
| { | |
| "epoch": 0.8079676812927483, | |
| "grad_norm": 0.05868047347236466, | |
| "learning_rate": 1.8794172522915022e-06, | |
| "loss": 0.4462554931640625, | |
| "memory(GiB)": 77.63, | |
| "step": 2525, | |
| "token_acc": 0.8811320754716981, | |
| "train_speed(iter/s)": 0.029507 | |
| }, | |
| { | |
| "epoch": 0.8095676172953082, | |
| "grad_norm": 0.04860202133053876, | |
| "learning_rate": 1.849223413592046e-06, | |
| "loss": 0.4488513946533203, | |
| "memory(GiB)": 77.63, | |
| "step": 2530, | |
| "token_acc": 0.8654490616621984, | |
| "train_speed(iter/s)": 0.029517 | |
| }, | |
| { | |
| "epoch": 0.811167553297868, | |
| "grad_norm": 0.055005201314914515, | |
| "learning_rate": 1.8192493543057676e-06, | |
| "loss": 0.45094904899597166, | |
| "memory(GiB)": 77.63, | |
| "step": 2535, | |
| "token_acc": 0.8800874078120732, | |
| "train_speed(iter/s)": 0.029516 | |
| }, | |
| { | |
| "epoch": 0.812767489300428, | |
| "grad_norm": 0.058350813115584405, | |
| "learning_rate": 1.7894958826600884e-06, | |
| "loss": 0.4489152908325195, | |
| "memory(GiB)": 77.63, | |
| "step": 2540, | |
| "token_acc": 0.8784655623365301, | |
| "train_speed(iter/s)": 0.02952 | |
| }, | |
| { | |
| "epoch": 0.8143674253029879, | |
| "grad_norm": 0.05243037926696882, | |
| "learning_rate": 1.7599638009344566e-06, | |
| "loss": 0.4506648063659668, | |
| "memory(GiB)": 77.63, | |
| "step": 2545, | |
| "token_acc": 0.8711162255466053, | |
| "train_speed(iter/s)": 0.029531 | |
| }, | |
| { | |
| "epoch": 0.8159673613055478, | |
| "grad_norm": 0.055874525296942985, | |
| "learning_rate": 1.730653905438714e-06, | |
| "loss": 0.451121187210083, | |
| "memory(GiB)": 77.63, | |
| "step": 2550, | |
| "token_acc": 0.875531914893617, | |
| "train_speed(iter/s)": 0.029527 | |
| }, | |
| { | |
| "epoch": 0.8175672973081076, | |
| "grad_norm": 0.052382476873803714, | |
| "learning_rate": 1.701566986491614e-06, | |
| "loss": 0.43659415245056155, | |
| "memory(GiB)": 77.63, | |
| "step": 2555, | |
| "token_acc": 0.8824301518844928, | |
| "train_speed(iter/s)": 0.029539 | |
| }, | |
| { | |
| "epoch": 0.8191672333106675, | |
| "grad_norm": 0.05679998544269888, | |
| "learning_rate": 1.672703828399529e-06, | |
| "loss": 0.44143290519714357, | |
| "memory(GiB)": 77.63, | |
| "step": 2560, | |
| "token_acc": 0.9194786645241921, | |
| "train_speed(iter/s)": 0.029538 | |
| }, | |
| { | |
| "epoch": 0.8207671693132275, | |
| "grad_norm": 0.05177151194489443, | |
| "learning_rate": 1.6440652094352838e-06, | |
| "loss": 0.44036478996276857, | |
| "memory(GiB)": 77.63, | |
| "step": 2565, | |
| "token_acc": 0.86801315171442, | |
| "train_speed(iter/s)": 0.029534 | |
| }, | |
| { | |
| "epoch": 0.8223671053157874, | |
| "grad_norm": 0.047985617439266506, | |
| "learning_rate": 1.6156519018171856e-06, | |
| "loss": 0.44090909957885743, | |
| "memory(GiB)": 77.63, | |
| "step": 2570, | |
| "token_acc": 0.8926761055759482, | |
| "train_speed(iter/s)": 0.029546 | |
| }, | |
| { | |
| "epoch": 0.8239670413183473, | |
| "grad_norm": 0.06022869670658613, | |
| "learning_rate": 1.587464671688187e-06, | |
| "loss": 0.4480876922607422, | |
| "memory(GiB)": 77.63, | |
| "step": 2575, | |
| "token_acc": 0.8749468913751594, | |
| "train_speed(iter/s)": 0.029542 | |
| }, | |
| { | |
| "epoch": 0.8255669773209071, | |
| "grad_norm": 0.05591525813204934, | |
| "learning_rate": 1.5595042790952442e-06, | |
| "loss": 0.4516183853149414, | |
| "memory(GiB)": 77.63, | |
| "step": 2580, | |
| "token_acc": 0.8408594319009468, | |
| "train_speed(iter/s)": 0.029539 | |
| }, | |
| { | |
| "epoch": 0.827166913323467, | |
| "grad_norm": 0.05315795994218538, | |
| "learning_rate": 1.5317714779688076e-06, | |
| "loss": 0.44116387367248533, | |
| "memory(GiB)": 77.63, | |
| "step": 2585, | |
| "token_acc": 0.8697747394374089, | |
| "train_speed(iter/s)": 0.029549 | |
| }, | |
| { | |
| "epoch": 0.828766849326027, | |
| "grad_norm": 0.054322305095737905, | |
| "learning_rate": 1.5042670161024975e-06, | |
| "loss": 0.4457075119018555, | |
| "memory(GiB)": 77.63, | |
| "step": 2590, | |
| "token_acc": 0.8946940985381701, | |
| "train_speed(iter/s)": 0.029543 | |
| }, | |
| { | |
| "epoch": 0.8303667853285869, | |
| "grad_norm": 0.05625328099370699, | |
| "learning_rate": 1.4769916351329495e-06, | |
| "loss": 0.4413478851318359, | |
| "memory(GiB)": 77.63, | |
| "step": 2595, | |
| "token_acc": 0.8992218637312583, | |
| "train_speed(iter/s)": 0.029548 | |
| }, | |
| { | |
| "epoch": 0.8319667213311468, | |
| "grad_norm": 0.055362561635933255, | |
| "learning_rate": 1.4499460705198e-06, | |
| "loss": 0.4511932373046875, | |
| "memory(GiB)": 77.63, | |
| "step": 2600, | |
| "token_acc": 0.8438552188552189, | |
| "train_speed(iter/s)": 0.029552 | |
| }, | |
| { | |
| "epoch": 0.8335666573337066, | |
| "grad_norm": 0.052438473367666195, | |
| "learning_rate": 1.4231310515258745e-06, | |
| "loss": 0.441973352432251, | |
| "memory(GiB)": 77.63, | |
| "step": 2605, | |
| "token_acc": 0.8753952017853822, | |
| "train_speed(iter/s)": 0.029546 | |
| }, | |
| { | |
| "epoch": 0.8351665933362665, | |
| "grad_norm": 0.05274974730016364, | |
| "learning_rate": 1.396547301197504e-06, | |
| "loss": 0.4393311977386475, | |
| "memory(GiB)": 77.63, | |
| "step": 2610, | |
| "token_acc": 0.8518848700967906, | |
| "train_speed(iter/s)": 0.029557 | |
| }, | |
| { | |
| "epoch": 0.8367665293388264, | |
| "grad_norm": 0.05853294004614818, | |
| "learning_rate": 1.3701955363450447e-06, | |
| "loss": 0.4380232810974121, | |
| "memory(GiB)": 77.63, | |
| "step": 2615, | |
| "token_acc": 0.8570597362296354, | |
| "train_speed(iter/s)": 0.029554 | |
| }, | |
| { | |
| "epoch": 0.8383664653413864, | |
| "grad_norm": 0.05410978790127522, | |
| "learning_rate": 1.3440764675235384e-06, | |
| "loss": 0.4373164653778076, | |
| "memory(GiB)": 77.63, | |
| "step": 2620, | |
| "token_acc": 0.8798353909465021, | |
| "train_speed(iter/s)": 0.029552 | |
| }, | |
| { | |
| "epoch": 0.8399664013439463, | |
| "grad_norm": 0.048967955063799855, | |
| "learning_rate": 1.3181907990135624e-06, | |
| "loss": 0.4333020210266113, | |
| "memory(GiB)": 77.63, | |
| "step": 2625, | |
| "token_acc": 0.8836341008089608, | |
| "train_speed(iter/s)": 0.029564 | |
| }, | |
| { | |
| "epoch": 0.8415663373465061, | |
| "grad_norm": 0.048274089580157754, | |
| "learning_rate": 1.2925392288022299e-06, | |
| "loss": 0.4414947509765625, | |
| "memory(GiB)": 77.63, | |
| "step": 2630, | |
| "token_acc": 0.8760546404178385, | |
| "train_speed(iter/s)": 0.02956 | |
| }, | |
| { | |
| "epoch": 0.843166273349066, | |
| "grad_norm": 0.053684823491607934, | |
| "learning_rate": 1.267122448564374e-06, | |
| "loss": 0.44922800064086915, | |
| "memory(GiB)": 77.63, | |
| "step": 2635, | |
| "token_acc": 0.8554064052425748, | |
| "train_speed(iter/s)": 0.029558 | |
| }, | |
| { | |
| "epoch": 0.8447662093516259, | |
| "grad_norm": 0.05262528572569429, | |
| "learning_rate": 1.2419411436439021e-06, | |
| "loss": 0.4328805923461914, | |
| "memory(GiB)": 77.63, | |
| "step": 2640, | |
| "token_acc": 0.8400081317340923, | |
| "train_speed(iter/s)": 0.029565 | |
| }, | |
| { | |
| "epoch": 0.8463661453541859, | |
| "grad_norm": 0.05549273276976771, | |
| "learning_rate": 1.2169959930353049e-06, | |
| "loss": 0.4460554599761963, | |
| "memory(GiB)": 77.63, | |
| "step": 2645, | |
| "token_acc": 0.8804424157303371, | |
| "train_speed(iter/s)": 0.029559 | |
| }, | |
| { | |
| "epoch": 0.8479660813567458, | |
| "grad_norm": 0.05180537064402683, | |
| "learning_rate": 1.1922876693653584e-06, | |
| "loss": 0.4503427505493164, | |
| "memory(GiB)": 77.63, | |
| "step": 2650, | |
| "token_acc": 0.8934362934362934, | |
| "train_speed(iter/s)": 0.029564 | |
| }, | |
| { | |
| "epoch": 0.8495660173593056, | |
| "grad_norm": 0.05011013702559624, | |
| "learning_rate": 1.1678168388749788e-06, | |
| "loss": 0.4415099620819092, | |
| "memory(GiB)": 77.63, | |
| "step": 2655, | |
| "token_acc": 0.8995949690897463, | |
| "train_speed(iter/s)": 0.029566 | |
| }, | |
| { | |
| "epoch": 0.8511659533618655, | |
| "grad_norm": 0.057170743989864214, | |
| "learning_rate": 1.1435841614012666e-06, | |
| "loss": 0.44884433746337893, | |
| "memory(GiB)": 77.63, | |
| "step": 2660, | |
| "token_acc": 0.854816112084063, | |
| "train_speed(iter/s)": 0.02956 | |
| }, | |
| { | |
| "epoch": 0.8527658893644254, | |
| "grad_norm": 0.054388272142607975, | |
| "learning_rate": 1.1195902903597023e-06, | |
| "loss": 0.439667797088623, | |
| "memory(GiB)": 77.63, | |
| "step": 2665, | |
| "token_acc": 0.8846260387811634, | |
| "train_speed(iter/s)": 0.02957 | |
| }, | |
| { | |
| "epoch": 0.8543658253669854, | |
| "grad_norm": 0.05039207218424233, | |
| "learning_rate": 1.0958358727265438e-06, | |
| "loss": 0.4384475231170654, | |
| "memory(GiB)": 77.63, | |
| "step": 2670, | |
| "token_acc": 0.8525793222533995, | |
| "train_speed(iter/s)": 0.029565 | |
| }, | |
| { | |
| "epoch": 0.8559657613695452, | |
| "grad_norm": 0.0543645938545219, | |
| "learning_rate": 1.0723215490213635e-06, | |
| "loss": 0.4338691711425781, | |
| "memory(GiB)": 77.63, | |
| "step": 2675, | |
| "token_acc": 0.853824495541999, | |
| "train_speed(iter/s)": 0.029558 | |
| }, | |
| { | |
| "epoch": 0.8575656973721051, | |
| "grad_norm": 0.0579168704227633, | |
| "learning_rate": 1.0490479532897946e-06, | |
| "loss": 0.458463716506958, | |
| "memory(GiB)": 77.63, | |
| "step": 2680, | |
| "token_acc": 0.867092866756393, | |
| "train_speed(iter/s)": 0.029566 | |
| }, | |
| { | |
| "epoch": 0.859165633374665, | |
| "grad_norm": 0.04996147776053655, | |
| "learning_rate": 1.0260157130864178e-06, | |
| "loss": 0.43754091262817385, | |
| "memory(GiB)": 77.63, | |
| "step": 2685, | |
| "token_acc": 0.8611873713109128, | |
| "train_speed(iter/s)": 0.029558 | |
| }, | |
| { | |
| "epoch": 0.8607655693772249, | |
| "grad_norm": 0.053537353272037014, | |
| "learning_rate": 1.0032254494578519e-06, | |
| "loss": 0.44204487800598147, | |
| "memory(GiB)": 77.63, | |
| "step": 2690, | |
| "token_acc": 0.8575780654988576, | |
| "train_speed(iter/s)": 0.029558 | |
| }, | |
| { | |
| "epoch": 0.8623655053797848, | |
| "grad_norm": 0.055119127073411836, | |
| "learning_rate": 9.806777769260034e-07, | |
| "loss": 0.4500781536102295, | |
| "memory(GiB)": 77.63, | |
| "step": 2695, | |
| "token_acc": 0.8872294372294373, | |
| "train_speed(iter/s)": 0.029562 | |
| }, | |
| { | |
| "epoch": 0.8639654413823447, | |
| "grad_norm": 0.0558713442289911, | |
| "learning_rate": 9.583733034714982e-07, | |
| "loss": 0.43947248458862304, | |
| "memory(GiB)": 77.63, | |
| "step": 2700, | |
| "token_acc": 0.8926744522729466, | |
| "train_speed(iter/s)": 0.029555 | |
| }, | |
| { | |
| "epoch": 0.8655653773849046, | |
| "grad_norm": 0.0552196144062876, | |
| "learning_rate": 9.363126305172831e-07, | |
| "loss": 0.4443229675292969, | |
| "memory(GiB)": 77.63, | |
| "step": 2705, | |
| "token_acc": 0.9038031319910514, | |
| "train_speed(iter/s)": 0.029561 | |
| }, | |
| { | |
| "epoch": 0.8671653133874645, | |
| "grad_norm": 0.055723745057826034, | |
| "learning_rate": 9.144963529124163e-07, | |
| "loss": 0.42942004203796386, | |
| "memory(GiB)": 77.63, | |
| "step": 2710, | |
| "token_acc": 0.8680161943319838, | |
| "train_speed(iter/s)": 0.029557 | |
| }, | |
| { | |
| "epoch": 0.8687652493900244, | |
| "grad_norm": 0.0589746686821641, | |
| "learning_rate": 8.929250589160166e-07, | |
| "loss": 0.4397599220275879, | |
| "memory(GiB)": 77.63, | |
| "step": 2715, | |
| "token_acc": 0.8713878713878714, | |
| "train_speed(iter/s)": 0.029552 | |
| }, | |
| { | |
| "epoch": 0.8703651853925843, | |
| "grad_norm": 0.04909314017257213, | |
| "learning_rate": 8.715993301814174e-07, | |
| "loss": 0.44155421257019045, | |
| "memory(GiB)": 77.63, | |
| "step": 2720, | |
| "token_acc": 0.8710053650571495, | |
| "train_speed(iter/s)": 0.029561 | |
| }, | |
| { | |
| "epoch": 0.8719651213951441, | |
| "grad_norm": 0.05047518453544575, | |
| "learning_rate": 8.505197417404687e-07, | |
| "loss": 0.43677616119384766, | |
| "memory(GiB)": 77.63, | |
| "step": 2725, | |
| "token_acc": 0.8809886575249704, | |
| "train_speed(iter/s)": 0.029556 | |
| }, | |
| { | |
| "epoch": 0.8735650573977041, | |
| "grad_norm": 0.05102151204327215, | |
| "learning_rate": 8.296868619880372e-07, | |
| "loss": 0.44188566207885743, | |
| "memory(GiB)": 77.63, | |
| "step": 2730, | |
| "token_acc": 0.8547172833573602, | |
| "train_speed(iter/s)": 0.029553 | |
| }, | |
| { | |
| "epoch": 0.875164993400264, | |
| "grad_norm": 0.04729834705444575, | |
| "learning_rate": 8.091012526666797e-07, | |
| "loss": 0.4441237926483154, | |
| "memory(GiB)": 77.63, | |
| "step": 2735, | |
| "token_acc": 0.8537975972307066, | |
| "train_speed(iter/s)": 0.029561 | |
| }, | |
| { | |
| "epoch": 0.8767649294028239, | |
| "grad_norm": 0.047668539210598965, | |
| "learning_rate": 7.887634688515e-07, | |
| "loss": 0.4462736129760742, | |
| "memory(GiB)": 77.63, | |
| "step": 2740, | |
| "token_acc": 0.903437815975733, | |
| "train_speed(iter/s)": 0.029554 | |
| }, | |
| { | |
| "epoch": 0.8783648654053838, | |
| "grad_norm": 0.052216823887528664, | |
| "learning_rate": 7.686740589351704e-07, | |
| "loss": 0.44857120513916016, | |
| "memory(GiB)": 77.63, | |
| "step": 2745, | |
| "token_acc": 0.8033573141486811, | |
| "train_speed(iter/s)": 0.029556 | |
| }, | |
| { | |
| "epoch": 0.8799648014079436, | |
| "grad_norm": 0.055862979558343906, | |
| "learning_rate": 7.488335646131628e-07, | |
| "loss": 0.44959425926208496, | |
| "memory(GiB)": 77.63, | |
| "step": 2750, | |
| "token_acc": 0.8605054151624548, | |
| "train_speed(iter/s)": 0.029558 | |
| }, | |
| { | |
| "epoch": 0.8799648014079436, | |
| "eval_loss": 0.6577034592628479, | |
| "eval_runtime": 106.4875, | |
| "eval_samples_per_second": 188.642, | |
| "eval_steps_per_second": 0.948, | |
| "eval_token_acc": 0.8722958612553617, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.8815647374105036, | |
| "grad_norm": 0.051528350081992934, | |
| "learning_rate": 7.292425208691212e-07, | |
| "loss": 0.43878631591796874, | |
| "memory(GiB)": 77.63, | |
| "step": 2755, | |
| "token_acc": 0.8812832745626772, | |
| "train_speed(iter/s)": 0.029532 | |
| }, | |
| { | |
| "epoch": 0.8831646734130635, | |
| "grad_norm": 0.05310175482611414, | |
| "learning_rate": 7.099014559604556e-07, | |
| "loss": 0.45635418891906737, | |
| "memory(GiB)": 77.63, | |
| "step": 2760, | |
| "token_acc": 0.8894999360532038, | |
| "train_speed(iter/s)": 0.029537 | |
| }, | |
| { | |
| "epoch": 0.8847646094156234, | |
| "grad_norm": 0.04975738369955541, | |
| "learning_rate": 6.908108914040823e-07, | |
| "loss": 0.4421397686004639, | |
| "memory(GiB)": 77.63, | |
| "step": 2765, | |
| "token_acc": 0.9070493575117089, | |
| "train_speed(iter/s)": 0.029548 | |
| }, | |
| { | |
| "epoch": 0.8863645454181833, | |
| "grad_norm": 0.053564472486824076, | |
| "learning_rate": 6.71971341962373e-07, | |
| "loss": 0.4513510227203369, | |
| "memory(GiB)": 77.63, | |
| "step": 2770, | |
| "token_acc": 0.8660589060308556, | |
| "train_speed(iter/s)": 0.02955 | |
| }, | |
| { | |
| "epoch": 0.8879644814207431, | |
| "grad_norm": 0.06332925713320489, | |
| "learning_rate": 6.53383315629268e-07, | |
| "loss": 0.4404273509979248, | |
| "memory(GiB)": 77.63, | |
| "step": 2775, | |
| "token_acc": 0.8507806501151779, | |
| "train_speed(iter/s)": 0.029546 | |
| }, | |
| { | |
| "epoch": 0.889564417423303, | |
| "grad_norm": 0.063294227744794, | |
| "learning_rate": 6.350473136165836e-07, | |
| "loss": 0.4379493236541748, | |
| "memory(GiB)": 77.63, | |
| "step": 2780, | |
| "token_acc": 0.8879898461050294, | |
| "train_speed(iter/s)": 0.029561 | |
| }, | |
| { | |
| "epoch": 0.891164353425863, | |
| "grad_norm": 0.05151642870451994, | |
| "learning_rate": 6.169638303404912e-07, | |
| "loss": 0.4380655765533447, | |
| "memory(GiB)": 77.63, | |
| "step": 2785, | |
| "token_acc": 0.8904059040590406, | |
| "train_speed(iter/s)": 0.02956 | |
| }, | |
| { | |
| "epoch": 0.8927642894284229, | |
| "grad_norm": 0.05406168921762394, | |
| "learning_rate": 5.991333534081878e-07, | |
| "loss": 0.4479250907897949, | |
| "memory(GiB)": 77.63, | |
| "step": 2790, | |
| "token_acc": 0.8831118813787792, | |
| "train_speed(iter/s)": 0.02956 | |
| }, | |
| { | |
| "epoch": 0.8943642254309827, | |
| "grad_norm": 0.054911478200183335, | |
| "learning_rate": 5.815563636047539e-07, | |
| "loss": 0.43634886741638185, | |
| "memory(GiB)": 77.63, | |
| "step": 2795, | |
| "token_acc": 0.8714865708931917, | |
| "train_speed(iter/s)": 0.02957 | |
| }, | |
| { | |
| "epoch": 0.8959641614335426, | |
| "grad_norm": 0.05257509941727236, | |
| "learning_rate": 5.64233334880181e-07, | |
| "loss": 0.44048466682434084, | |
| "memory(GiB)": 77.63, | |
| "step": 2800, | |
| "token_acc": 0.891296869625043, | |
| "train_speed(iter/s)": 0.029568 | |
| }, | |
| { | |
| "epoch": 0.8975640974361025, | |
| "grad_norm": 0.05532206374459385, | |
| "learning_rate": 5.471647343365982e-07, | |
| "loss": 0.44726853370666503, | |
| "memory(GiB)": 77.63, | |
| "step": 2805, | |
| "token_acc": 0.8828892005610098, | |
| "train_speed(iter/s)": 0.029576 | |
| }, | |
| { | |
| "epoch": 0.8991640334386625, | |
| "grad_norm": 0.05107208389411162, | |
| "learning_rate": 5.303510222156716e-07, | |
| "loss": 0.4470540523529053, | |
| "memory(GiB)": 77.63, | |
| "step": 2810, | |
| "token_acc": 0.8489765812281025, | |
| "train_speed(iter/s)": 0.029574 | |
| }, | |
| { | |
| "epoch": 0.9007639694412224, | |
| "grad_norm": 0.055452171728558604, | |
| "learning_rate": 5.137926518862013e-07, | |
| "loss": 0.4417248249053955, | |
| "memory(GiB)": 77.63, | |
| "step": 2815, | |
| "token_acc": 0.8739084132055378, | |
| "train_speed(iter/s)": 0.029569 | |
| }, | |
| { | |
| "epoch": 0.9023639054437822, | |
| "grad_norm": 0.04866974543895629, | |
| "learning_rate": 4.974900698318885e-07, | |
| "loss": 0.4414045810699463, | |
| "memory(GiB)": 77.63, | |
| "step": 2820, | |
| "token_acc": 0.8659420289855072, | |
| "train_speed(iter/s)": 0.029581 | |
| }, | |
| { | |
| "epoch": 0.9039638414463421, | |
| "grad_norm": 0.05275403785935388, | |
| "learning_rate": 4.814437156393048e-07, | |
| "loss": 0.4543337821960449, | |
| "memory(GiB)": 77.63, | |
| "step": 2825, | |
| "token_acc": 0.8396122896854425, | |
| "train_speed(iter/s)": 0.029574 | |
| }, | |
| { | |
| "epoch": 0.905563777448902, | |
| "grad_norm": 0.05372217346495556, | |
| "learning_rate": 4.656540219860317e-07, | |
| "loss": 0.45271754264831543, | |
| "memory(GiB)": 77.63, | |
| "step": 2830, | |
| "token_acc": 0.8707617789520036, | |
| "train_speed(iter/s)": 0.029571 | |
| }, | |
| { | |
| "epoch": 0.907163713451462, | |
| "grad_norm": 0.05715427837146615, | |
| "learning_rate": 4.501214146289956e-07, | |
| "loss": 0.4418344497680664, | |
| "memory(GiB)": 77.63, | |
| "step": 2835, | |
| "token_acc": 0.8788416882939489, | |
| "train_speed(iter/s)": 0.02958 | |
| }, | |
| { | |
| "epoch": 0.9087636494540219, | |
| "grad_norm": 0.04986985198768239, | |
| "learning_rate": 4.3484631239299356e-07, | |
| "loss": 0.4437891960144043, | |
| "memory(GiB)": 77.63, | |
| "step": 2840, | |
| "token_acc": 0.8431597023468803, | |
| "train_speed(iter/s)": 0.029572 | |
| }, | |
| { | |
| "epoch": 0.9103635854565817, | |
| "grad_norm": 0.04999552116510165, | |
| "learning_rate": 4.198291271593924e-07, | |
| "loss": 0.44283204078674315, | |
| "memory(GiB)": 77.63, | |
| "step": 2845, | |
| "token_acc": 0.8843727072633896, | |
| "train_speed(iter/s)": 0.029575 | |
| }, | |
| { | |
| "epoch": 0.9119635214591416, | |
| "grad_norm": 0.047844581858855956, | |
| "learning_rate": 4.0507026385502747e-07, | |
| "loss": 0.4449836254119873, | |
| "memory(GiB)": 77.63, | |
| "step": 2850, | |
| "token_acc": 0.9034812490661885, | |
| "train_speed(iter/s)": 0.029576 | |
| }, | |
| { | |
| "epoch": 0.9135634574617015, | |
| "grad_norm": 0.05675006848739315, | |
| "learning_rate": 3.9057012044127817e-07, | |
| "loss": 0.44204154014587405, | |
| "memory(GiB)": 77.63, | |
| "step": 2855, | |
| "token_acc": 0.8655569782330346, | |
| "train_speed(iter/s)": 0.029569 | |
| }, | |
| { | |
| "epoch": 0.9151633934642615, | |
| "grad_norm": 0.054077703664588216, | |
| "learning_rate": 3.7632908790334656e-07, | |
| "loss": 0.4383398532867432, | |
| "memory(GiB)": 77.63, | |
| "step": 2860, | |
| "token_acc": 0.8899396929824561, | |
| "train_speed(iter/s)": 0.029578 | |
| }, | |
| { | |
| "epoch": 0.9167633294668214, | |
| "grad_norm": 0.055142279260510525, | |
| "learning_rate": 3.6234755023970447e-07, | |
| "loss": 0.4388674259185791, | |
| "memory(GiB)": 77.63, | |
| "step": 2865, | |
| "token_acc": 0.8406133828996283, | |
| "train_speed(iter/s)": 0.029574 | |
| }, | |
| { | |
| "epoch": 0.9183632654693812, | |
| "grad_norm": 0.05242909357272202, | |
| "learning_rate": 3.4862588445174985e-07, | |
| "loss": 0.44350008964538573, | |
| "memory(GiB)": 77.63, | |
| "step": 2870, | |
| "token_acc": 0.8773854961832062, | |
| "train_speed(iter/s)": 0.029568 | |
| }, | |
| { | |
| "epoch": 0.9199632014719411, | |
| "grad_norm": 0.051431062939426, | |
| "learning_rate": 3.3516446053363015e-07, | |
| "loss": 0.43948516845703123, | |
| "memory(GiB)": 77.63, | |
| "step": 2875, | |
| "token_acc": 0.8812238692512353, | |
| "train_speed(iter/s)": 0.029578 | |
| }, | |
| { | |
| "epoch": 0.921563137474501, | |
| "grad_norm": 0.04842108410960974, | |
| "learning_rate": 3.219636414622751e-07, | |
| "loss": 0.44395694732666013, | |
| "memory(GiB)": 77.63, | |
| "step": 2880, | |
| "token_acc": 0.872349158571624, | |
| "train_speed(iter/s)": 0.029573 | |
| }, | |
| { | |
| "epoch": 0.923163073477061, | |
| "grad_norm": 0.050854557927068264, | |
| "learning_rate": 3.090237831876053e-07, | |
| "loss": 0.4437469482421875, | |
| "memory(GiB)": 77.63, | |
| "step": 2885, | |
| "token_acc": 0.8500566251415629, | |
| "train_speed(iter/s)": 0.029572 | |
| }, | |
| { | |
| "epoch": 0.9247630094796209, | |
| "grad_norm": 0.05289158935619814, | |
| "learning_rate": 2.9634523462293005e-07, | |
| "loss": 0.439394474029541, | |
| "memory(GiB)": 77.63, | |
| "step": 2890, | |
| "token_acc": 0.9039268013724743, | |
| "train_speed(iter/s)": 0.029576 | |
| }, | |
| { | |
| "epoch": 0.9263629454821807, | |
| "grad_norm": 0.0532633695315608, | |
| "learning_rate": 2.839283376355506e-07, | |
| "loss": 0.4414195537567139, | |
| "memory(GiB)": 77.63, | |
| "step": 2895, | |
| "token_acc": 0.8979206049149339, | |
| "train_speed(iter/s)": 0.029567 | |
| }, | |
| { | |
| "epoch": 0.9279628814847406, | |
| "grad_norm": 0.05197494947423765, | |
| "learning_rate": 2.717734270375272e-07, | |
| "loss": 0.4303572177886963, | |
| "memory(GiB)": 77.63, | |
| "step": 2900, | |
| "token_acc": 0.8932318992654774, | |
| "train_speed(iter/s)": 0.029573 | |
| }, | |
| { | |
| "epoch": 0.9295628174873005, | |
| "grad_norm": 0.05715199054523632, | |
| "learning_rate": 2.5988083057666534e-07, | |
| "loss": 0.4488718032836914, | |
| "memory(GiB)": 77.63, | |
| "step": 2905, | |
| "token_acc": 0.8734599589322382, | |
| "train_speed(iter/s)": 0.029572 | |
| }, | |
| { | |
| "epoch": 0.9311627534898604, | |
| "grad_norm": 0.052180972248620894, | |
| "learning_rate": 2.4825086892766745e-07, | |
| "loss": 0.44499683380126953, | |
| "memory(GiB)": 77.63, | |
| "step": 2910, | |
| "token_acc": 0.8798773215198501, | |
| "train_speed(iter/s)": 0.029568 | |
| }, | |
| { | |
| "epoch": 0.9327626894924202, | |
| "grad_norm": 0.05195036988843101, | |
| "learning_rate": 2.3688385568349515e-07, | |
| "loss": 0.4348268508911133, | |
| "memory(GiB)": 77.63, | |
| "step": 2915, | |
| "token_acc": 0.8380835380835381, | |
| "train_speed(iter/s)": 0.029578 | |
| }, | |
| { | |
| "epoch": 0.9343626254949802, | |
| "grad_norm": 0.05340475761749026, | |
| "learning_rate": 2.2578009734690264e-07, | |
| "loss": 0.4533662796020508, | |
| "memory(GiB)": 77.63, | |
| "step": 2920, | |
| "token_acc": 0.8854700854700854, | |
| "train_speed(iter/s)": 0.029569 | |
| }, | |
| { | |
| "epoch": 0.9359625614975401, | |
| "grad_norm": 0.047434143218971526, | |
| "learning_rate": 2.1493989332218468e-07, | |
| "loss": 0.4382183074951172, | |
| "memory(GiB)": 77.63, | |
| "step": 2925, | |
| "token_acc": 0.880750496121234, | |
| "train_speed(iter/s)": 0.029565 | |
| }, | |
| { | |
| "epoch": 0.9375624975001, | |
| "grad_norm": 0.052509681961836426, | |
| "learning_rate": 2.043635359070928e-07, | |
| "loss": 0.44708704948425293, | |
| "memory(GiB)": 77.63, | |
| "step": 2930, | |
| "token_acc": 0.8942012598933936, | |
| "train_speed(iter/s)": 0.029576 | |
| }, | |
| { | |
| "epoch": 0.9391624335026599, | |
| "grad_norm": 0.0531539858574761, | |
| "learning_rate": 1.9405131028495838e-07, | |
| "loss": 0.45058341026306153, | |
| "memory(GiB)": 77.63, | |
| "step": 2935, | |
| "token_acc": 0.8614694335389792, | |
| "train_speed(iter/s)": 0.029569 | |
| }, | |
| { | |
| "epoch": 0.9407623695052197, | |
| "grad_norm": 0.05931701393554437, | |
| "learning_rate": 1.8400349451700438e-07, | |
| "loss": 0.44367156028747556, | |
| "memory(GiB)": 77.63, | |
| "step": 2940, | |
| "token_acc": 0.881156184096266, | |
| "train_speed(iter/s)": 0.029573 | |
| }, | |
| { | |
| "epoch": 0.9423623055077797, | |
| "grad_norm": 0.05380572562718686, | |
| "learning_rate": 1.742203595348435e-07, | |
| "loss": 0.4424111843109131, | |
| "memory(GiB)": 77.63, | |
| "step": 2945, | |
| "token_acc": 0.8769617074701821, | |
| "train_speed(iter/s)": 0.029576 | |
| }, | |
| { | |
| "epoch": 0.9439622415103396, | |
| "grad_norm": 0.06156259021000458, | |
| "learning_rate": 1.6470216913317628e-07, | |
| "loss": 0.4509577751159668, | |
| "memory(GiB)": 77.63, | |
| "step": 2950, | |
| "token_acc": 0.8544532947139754, | |
| "train_speed(iter/s)": 0.029567 | |
| }, | |
| { | |
| "epoch": 0.9455621775128995, | |
| "grad_norm": 0.050697077683688974, | |
| "learning_rate": 1.5544917996267562e-07, | |
| "loss": 0.44117283821105957, | |
| "memory(GiB)": 77.63, | |
| "step": 2955, | |
| "token_acc": 0.8515226026101759, | |
| "train_speed(iter/s)": 0.029575 | |
| }, | |
| { | |
| "epoch": 0.9471621135154594, | |
| "grad_norm": 0.05205296295346333, | |
| "learning_rate": 1.464616415230702e-07, | |
| "loss": 0.4488182067871094, | |
| "memory(GiB)": 77.63, | |
| "step": 2960, | |
| "token_acc": 0.874439461883408, | |
| "train_speed(iter/s)": 0.029569 | |
| }, | |
| { | |
| "epoch": 0.9487620495180192, | |
| "grad_norm": 0.047632029637579856, | |
| "learning_rate": 1.3773979615640976e-07, | |
| "loss": 0.4415272235870361, | |
| "memory(GiB)": 77.63, | |
| "step": 2965, | |
| "token_acc": 0.8889883616830797, | |
| "train_speed(iter/s)": 0.029564 | |
| }, | |
| { | |
| "epoch": 0.9503619855205792, | |
| "grad_norm": 0.051226023380966074, | |
| "learning_rate": 1.292838790405393e-07, | |
| "loss": 0.4453396797180176, | |
| "memory(GiB)": 77.63, | |
| "step": 2970, | |
| "token_acc": 0.8701866977829639, | |
| "train_speed(iter/s)": 0.029572 | |
| }, | |
| { | |
| "epoch": 0.9519619215231391, | |
| "grad_norm": 0.05296626405711913, | |
| "learning_rate": 1.2109411818274851e-07, | |
| "loss": 0.44417614936828614, | |
| "memory(GiB)": 77.63, | |
| "step": 2975, | |
| "token_acc": 0.8997547959036493, | |
| "train_speed(iter/s)": 0.029565 | |
| }, | |
| { | |
| "epoch": 0.953561857525699, | |
| "grad_norm": 0.0535756814263424, | |
| "learning_rate": 1.1317073441363458e-07, | |
| "loss": 0.444796085357666, | |
| "memory(GiB)": 77.63, | |
| "step": 2980, | |
| "token_acc": 0.9101887677336147, | |
| "train_speed(iter/s)": 0.029563 | |
| }, | |
| { | |
| "epoch": 0.9551617935282589, | |
| "grad_norm": 0.05005027675979017, | |
| "learning_rate": 1.055139413811379e-07, | |
| "loss": 0.45203323364257814, | |
| "memory(GiB)": 77.63, | |
| "step": 2985, | |
| "token_acc": 0.8569892473118279, | |
| "train_speed(iter/s)": 0.029568 | |
| }, | |
| { | |
| "epoch": 0.9567617295308187, | |
| "grad_norm": 0.04934807615166247, | |
| "learning_rate": 9.812394554478355e-08, | |
| "loss": 0.43912034034729003, | |
| "memory(GiB)": 77.63, | |
| "step": 2990, | |
| "token_acc": 0.8557346268189642, | |
| "train_speed(iter/s)": 0.029559 | |
| }, | |
| { | |
| "epoch": 0.9583616655333786, | |
| "grad_norm": 0.051618825055470385, | |
| "learning_rate": 9.10009461701189e-08, | |
| "loss": 0.4506105899810791, | |
| "memory(GiB)": 77.63, | |
| "step": 2995, | |
| "token_acc": 0.7809948032665182, | |
| "train_speed(iter/s)": 0.029565 | |
| }, | |
| { | |
| "epoch": 0.9599616015359386, | |
| "grad_norm": 0.054833054855342726, | |
| "learning_rate": 8.41451353233369e-08, | |
| "loss": 0.442844820022583, | |
| "memory(GiB)": 77.63, | |
| "step": 3000, | |
| "token_acc": 0.8733862959285005, | |
| "train_speed(iter/s)": 0.029563 | |
| }, | |
| { | |
| "epoch": 0.9599616015359386, | |
| "eval_loss": 0.6573547720909119, | |
| "eval_runtime": 106.0877, | |
| "eval_samples_per_second": 189.353, | |
| "eval_steps_per_second": 0.952, | |
| "eval_token_acc": 0.8724322608695082, | |
| "step": 3000 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 3125, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.944159044825703e+20, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |