End of training
Browse files
all_results.json
CHANGED
|
@@ -1,15 +1,15 @@
|
|
| 1 |
{
|
| 2 |
-
"epoch":
|
| 3 |
-
"eval_cer": 0.
|
| 4 |
-
"eval_loss": 0.
|
| 5 |
-
"eval_runtime":
|
| 6 |
"eval_samples": 1647,
|
| 7 |
-
"eval_samples_per_second":
|
| 8 |
-
"eval_steps_per_second": 2.
|
| 9 |
-
"eval_wer": 0.
|
| 10 |
-
"train_loss": 0.
|
| 11 |
-
"train_runtime":
|
| 12 |
"train_samples": 3478,
|
| 13 |
-
"train_samples_per_second":
|
| 14 |
-
"train_steps_per_second":
|
| 15 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"epoch": 96.33,
|
| 3 |
+
"eval_cer": 0.13390005847164124,
|
| 4 |
+
"eval_loss": 0.7359636425971985,
|
| 5 |
+
"eval_runtime": 83.055,
|
| 6 |
"eval_samples": 1647,
|
| 7 |
+
"eval_samples_per_second": 19.83,
|
| 8 |
+
"eval_steps_per_second": 2.48,
|
| 9 |
+
"eval_wer": 0.4754366254723726,
|
| 10 |
+
"train_loss": 0.0,
|
| 11 |
+
"train_runtime": 113.9955,
|
| 12 |
"train_samples": 3478,
|
| 13 |
+
"train_samples_per_second": 152.55,
|
| 14 |
+
"train_steps_per_second": 4.781
|
| 15 |
}
|
eval_results.json
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
{
|
| 2 |
-
"epoch":
|
| 3 |
-
"eval_cer": 0.
|
| 4 |
-
"eval_loss": 0.
|
| 5 |
-
"eval_runtime":
|
| 6 |
"eval_samples": 1647,
|
| 7 |
-
"eval_samples_per_second":
|
| 8 |
-
"eval_steps_per_second": 2.
|
| 9 |
-
"eval_wer": 0.
|
| 10 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"epoch": 96.33,
|
| 3 |
+
"eval_cer": 0.13390005847164124,
|
| 4 |
+
"eval_loss": 0.7359636425971985,
|
| 5 |
+
"eval_runtime": 83.055,
|
| 6 |
"eval_samples": 1647,
|
| 7 |
+
"eval_samples_per_second": 19.83,
|
| 8 |
+
"eval_steps_per_second": 2.48,
|
| 9 |
+
"eval_wer": 0.4754366254723726
|
| 10 |
}
|
runs/Feb02_08-46-53_job-98a7eb1d-c6a6-4e4b-936b-3898557b5e73/events.out.tfevents.1643792324.job-98a7eb1d-c6a6-4e4b-936b-3898557b5e73.544082.2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:08a917ccd708b36361a9ce5fa79328ee4416e8a8f0e92f2d35ec8d196cb7b111
|
| 3 |
+
size 405
|
train_results.json
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"epoch":
|
| 3 |
-
"train_loss": 0.
|
| 4 |
-
"train_runtime":
|
| 5 |
"train_samples": 3478,
|
| 6 |
-
"train_samples_per_second":
|
| 7 |
-
"train_steps_per_second":
|
| 8 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"epoch": 96.33,
|
| 3 |
+
"train_loss": 0.0,
|
| 4 |
+
"train_runtime": 113.9955,
|
| 5 |
"train_samples": 3478,
|
| 6 |
+
"train_samples_per_second": 152.55,
|
| 7 |
+
"train_steps_per_second": 4.781
|
| 8 |
}
|
trainer_state.json
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -63218,2418 +63218,18 @@
|
|
| 63218 |
"step": 10500
|
| 63219 |
},
|
| 63220 |
{
|
| 63221 |
-
"epoch": 96.
|
| 63222 |
-
"
|
| 63223 |
-
"
|
| 63224 |
-
"
|
| 63225 |
-
|
| 63226 |
-
|
| 63227 |
-
"
|
| 63228 |
-
"learning_rate": 1.932692307692308e-05,
|
| 63229 |
-
"loss": 0.1755,
|
| 63230 |
-
"step": 10502
|
| 63231 |
-
},
|
| 63232 |
-
{
|
| 63233 |
-
"epoch": 96.36,
|
| 63234 |
-
"learning_rate": 1.9278846153846155e-05,
|
| 63235 |
-
"loss": 0.0978,
|
| 63236 |
-
"step": 10503
|
| 63237 |
-
},
|
| 63238 |
-
{
|
| 63239 |
-
"epoch": 96.37,
|
| 63240 |
-
"learning_rate": 1.923076923076923e-05,
|
| 63241 |
-
"loss": 0.1605,
|
| 63242 |
-
"step": 10504
|
| 63243 |
-
},
|
| 63244 |
-
{
|
| 63245 |
-
"epoch": 96.38,
|
| 63246 |
-
"learning_rate": 1.9182692307692307e-05,
|
| 63247 |
-
"loss": 0.1321,
|
| 63248 |
-
"step": 10505
|
| 63249 |
-
},
|
| 63250 |
-
{
|
| 63251 |
-
"epoch": 96.39,
|
| 63252 |
-
"learning_rate": 1.9134615384615387e-05,
|
| 63253 |
-
"loss": 0.1023,
|
| 63254 |
-
"step": 10506
|
| 63255 |
-
},
|
| 63256 |
-
{
|
| 63257 |
-
"epoch": 96.39,
|
| 63258 |
-
"learning_rate": 1.908653846153846e-05,
|
| 63259 |
-
"loss": 0.0784,
|
| 63260 |
-
"step": 10507
|
| 63261 |
-
},
|
| 63262 |
-
{
|
| 63263 |
-
"epoch": 96.4,
|
| 63264 |
-
"learning_rate": 1.903846153846154e-05,
|
| 63265 |
-
"loss": 0.069,
|
| 63266 |
-
"step": 10508
|
| 63267 |
-
},
|
| 63268 |
-
{
|
| 63269 |
-
"epoch": 96.41,
|
| 63270 |
-
"learning_rate": 1.8990384615384615e-05,
|
| 63271 |
-
"loss": 0.0694,
|
| 63272 |
-
"step": 10509
|
| 63273 |
-
},
|
| 63274 |
-
{
|
| 63275 |
-
"epoch": 96.42,
|
| 63276 |
-
"learning_rate": 1.8942307692307694e-05,
|
| 63277 |
-
"loss": 0.1372,
|
| 63278 |
-
"step": 10510
|
| 63279 |
-
},
|
| 63280 |
-
{
|
| 63281 |
-
"epoch": 96.43,
|
| 63282 |
-
"learning_rate": 1.889423076923077e-05,
|
| 63283 |
-
"loss": 0.0912,
|
| 63284 |
-
"step": 10511
|
| 63285 |
-
},
|
| 63286 |
-
{
|
| 63287 |
-
"epoch": 96.44,
|
| 63288 |
-
"learning_rate": 1.8846153846153846e-05,
|
| 63289 |
-
"loss": 0.1011,
|
| 63290 |
-
"step": 10512
|
| 63291 |
-
},
|
| 63292 |
-
{
|
| 63293 |
-
"epoch": 96.45,
|
| 63294 |
-
"learning_rate": 1.8798076923076926e-05,
|
| 63295 |
-
"loss": 0.1284,
|
| 63296 |
-
"step": 10513
|
| 63297 |
-
},
|
| 63298 |
-
{
|
| 63299 |
-
"epoch": 96.46,
|
| 63300 |
-
"learning_rate": 1.875e-05,
|
| 63301 |
-
"loss": 0.1563,
|
| 63302 |
-
"step": 10514
|
| 63303 |
-
},
|
| 63304 |
-
{
|
| 63305 |
-
"epoch": 96.47,
|
| 63306 |
-
"learning_rate": 1.8701923076923078e-05,
|
| 63307 |
-
"loss": 0.0666,
|
| 63308 |
-
"step": 10515
|
| 63309 |
-
},
|
| 63310 |
-
{
|
| 63311 |
-
"epoch": 96.48,
|
| 63312 |
-
"learning_rate": 1.8653846153846154e-05,
|
| 63313 |
-
"loss": 0.1044,
|
| 63314 |
-
"step": 10516
|
| 63315 |
-
},
|
| 63316 |
-
{
|
| 63317 |
-
"epoch": 96.49,
|
| 63318 |
-
"learning_rate": 1.8605769230769233e-05,
|
| 63319 |
-
"loss": 0.0935,
|
| 63320 |
-
"step": 10517
|
| 63321 |
-
},
|
| 63322 |
-
{
|
| 63323 |
-
"epoch": 96.5,
|
| 63324 |
-
"learning_rate": 1.8557692307692306e-05,
|
| 63325 |
-
"loss": 0.112,
|
| 63326 |
-
"step": 10518
|
| 63327 |
-
},
|
| 63328 |
-
{
|
| 63329 |
-
"epoch": 96.5,
|
| 63330 |
-
"learning_rate": 1.8509615384615385e-05,
|
| 63331 |
-
"loss": 0.126,
|
| 63332 |
-
"step": 10519
|
| 63333 |
-
},
|
| 63334 |
-
{
|
| 63335 |
-
"epoch": 96.51,
|
| 63336 |
-
"learning_rate": 1.8461538461538465e-05,
|
| 63337 |
-
"loss": 0.1627,
|
| 63338 |
-
"step": 10520
|
| 63339 |
-
},
|
| 63340 |
-
{
|
| 63341 |
-
"epoch": 96.52,
|
| 63342 |
-
"learning_rate": 1.8413461538461537e-05,
|
| 63343 |
-
"loss": 0.1363,
|
| 63344 |
-
"step": 10521
|
| 63345 |
-
},
|
| 63346 |
-
{
|
| 63347 |
-
"epoch": 96.53,
|
| 63348 |
-
"learning_rate": 1.8365384615384617e-05,
|
| 63349 |
-
"loss": 0.1524,
|
| 63350 |
-
"step": 10522
|
| 63351 |
-
},
|
| 63352 |
-
{
|
| 63353 |
-
"epoch": 96.54,
|
| 63354 |
-
"learning_rate": 1.8317307692307693e-05,
|
| 63355 |
-
"loss": 0.1185,
|
| 63356 |
-
"step": 10523
|
| 63357 |
-
},
|
| 63358 |
-
{
|
| 63359 |
-
"epoch": 96.55,
|
| 63360 |
-
"learning_rate": 1.8269230769230772e-05,
|
| 63361 |
-
"loss": 0.111,
|
| 63362 |
-
"step": 10524
|
| 63363 |
-
},
|
| 63364 |
-
{
|
| 63365 |
-
"epoch": 96.56,
|
| 63366 |
-
"learning_rate": 1.8221153846153845e-05,
|
| 63367 |
-
"loss": 0.1093,
|
| 63368 |
-
"step": 10525
|
| 63369 |
-
},
|
| 63370 |
-
{
|
| 63371 |
-
"epoch": 96.57,
|
| 63372 |
-
"learning_rate": 1.8173076923076924e-05,
|
| 63373 |
-
"loss": 0.058,
|
| 63374 |
-
"step": 10526
|
| 63375 |
-
},
|
| 63376 |
-
{
|
| 63377 |
-
"epoch": 96.58,
|
| 63378 |
-
"learning_rate": 1.8125e-05,
|
| 63379 |
-
"loss": 0.0667,
|
| 63380 |
-
"step": 10527
|
| 63381 |
-
},
|
| 63382 |
-
{
|
| 63383 |
-
"epoch": 96.59,
|
| 63384 |
-
"learning_rate": 1.8076923076923076e-05,
|
| 63385 |
-
"loss": 0.1126,
|
| 63386 |
-
"step": 10528
|
| 63387 |
-
},
|
| 63388 |
-
{
|
| 63389 |
-
"epoch": 96.6,
|
| 63390 |
-
"learning_rate": 1.8028846153846152e-05,
|
| 63391 |
-
"loss": 0.0928,
|
| 63392 |
-
"step": 10529
|
| 63393 |
-
},
|
| 63394 |
-
{
|
| 63395 |
-
"epoch": 96.61,
|
| 63396 |
-
"learning_rate": 1.798076923076923e-05,
|
| 63397 |
-
"loss": 0.0966,
|
| 63398 |
-
"step": 10530
|
| 63399 |
-
},
|
| 63400 |
-
{
|
| 63401 |
-
"epoch": 96.61,
|
| 63402 |
-
"learning_rate": 1.793269230769231e-05,
|
| 63403 |
-
"loss": 0.0637,
|
| 63404 |
-
"step": 10531
|
| 63405 |
-
},
|
| 63406 |
-
{
|
| 63407 |
-
"epoch": 96.62,
|
| 63408 |
-
"learning_rate": 1.7884615384615384e-05,
|
| 63409 |
-
"loss": 0.0786,
|
| 63410 |
-
"step": 10532
|
| 63411 |
-
},
|
| 63412 |
-
{
|
| 63413 |
-
"epoch": 96.63,
|
| 63414 |
-
"learning_rate": 1.7836538461538463e-05,
|
| 63415 |
-
"loss": 0.1502,
|
| 63416 |
-
"step": 10533
|
| 63417 |
-
},
|
| 63418 |
-
{
|
| 63419 |
-
"epoch": 96.64,
|
| 63420 |
-
"learning_rate": 1.778846153846154e-05,
|
| 63421 |
-
"loss": 0.0982,
|
| 63422 |
-
"step": 10534
|
| 63423 |
-
},
|
| 63424 |
-
{
|
| 63425 |
-
"epoch": 96.65,
|
| 63426 |
-
"learning_rate": 1.7740384615384615e-05,
|
| 63427 |
-
"loss": 0.1326,
|
| 63428 |
-
"step": 10535
|
| 63429 |
-
},
|
| 63430 |
-
{
|
| 63431 |
-
"epoch": 96.66,
|
| 63432 |
-
"learning_rate": 1.769230769230769e-05,
|
| 63433 |
-
"loss": 0.0662,
|
| 63434 |
-
"step": 10536
|
| 63435 |
-
},
|
| 63436 |
-
{
|
| 63437 |
-
"epoch": 96.67,
|
| 63438 |
-
"learning_rate": 1.764423076923077e-05,
|
| 63439 |
-
"loss": 0.1025,
|
| 63440 |
-
"step": 10537
|
| 63441 |
-
},
|
| 63442 |
-
{
|
| 63443 |
-
"epoch": 96.68,
|
| 63444 |
-
"learning_rate": 1.7596153846153846e-05,
|
| 63445 |
-
"loss": 0.1385,
|
| 63446 |
-
"step": 10538
|
| 63447 |
-
},
|
| 63448 |
-
{
|
| 63449 |
-
"epoch": 96.69,
|
| 63450 |
-
"learning_rate": 1.7548076923076922e-05,
|
| 63451 |
-
"loss": 0.0851,
|
| 63452 |
-
"step": 10539
|
| 63453 |
-
},
|
| 63454 |
-
{
|
| 63455 |
-
"epoch": 96.7,
|
| 63456 |
-
"learning_rate": 1.7500000000000002e-05,
|
| 63457 |
-
"loss": 0.0788,
|
| 63458 |
-
"step": 10540
|
| 63459 |
-
},
|
| 63460 |
-
{
|
| 63461 |
-
"epoch": 96.71,
|
| 63462 |
-
"learning_rate": 1.7451923076923078e-05,
|
| 63463 |
-
"loss": 0.0495,
|
| 63464 |
-
"step": 10541
|
| 63465 |
-
},
|
| 63466 |
-
{
|
| 63467 |
-
"epoch": 96.72,
|
| 63468 |
-
"learning_rate": 1.7403846153846154e-05,
|
| 63469 |
-
"loss": 0.0613,
|
| 63470 |
-
"step": 10542
|
| 63471 |
-
},
|
| 63472 |
-
{
|
| 63473 |
-
"epoch": 96.72,
|
| 63474 |
-
"learning_rate": 1.735576923076923e-05,
|
| 63475 |
-
"loss": 0.083,
|
| 63476 |
-
"step": 10543
|
| 63477 |
-
},
|
| 63478 |
-
{
|
| 63479 |
-
"epoch": 96.73,
|
| 63480 |
-
"learning_rate": 1.730769230769231e-05,
|
| 63481 |
-
"loss": 0.1314,
|
| 63482 |
-
"step": 10544
|
| 63483 |
-
},
|
| 63484 |
-
{
|
| 63485 |
-
"epoch": 96.74,
|
| 63486 |
-
"learning_rate": 1.7259615384615385e-05,
|
| 63487 |
-
"loss": 0.1278,
|
| 63488 |
-
"step": 10545
|
| 63489 |
-
},
|
| 63490 |
-
{
|
| 63491 |
-
"epoch": 96.75,
|
| 63492 |
-
"learning_rate": 1.721153846153846e-05,
|
| 63493 |
-
"loss": 0.1405,
|
| 63494 |
-
"step": 10546
|
| 63495 |
-
},
|
| 63496 |
-
{
|
| 63497 |
-
"epoch": 96.76,
|
| 63498 |
-
"learning_rate": 1.7163461538461537e-05,
|
| 63499 |
-
"loss": 0.0983,
|
| 63500 |
-
"step": 10547
|
| 63501 |
-
},
|
| 63502 |
-
{
|
| 63503 |
-
"epoch": 96.77,
|
| 63504 |
-
"learning_rate": 1.7115384615384617e-05,
|
| 63505 |
-
"loss": 0.1169,
|
| 63506 |
-
"step": 10548
|
| 63507 |
-
},
|
| 63508 |
-
{
|
| 63509 |
-
"epoch": 96.78,
|
| 63510 |
-
"learning_rate": 1.706730769230769e-05,
|
| 63511 |
-
"loss": 0.0914,
|
| 63512 |
-
"step": 10549
|
| 63513 |
-
},
|
| 63514 |
-
{
|
| 63515 |
-
"epoch": 96.79,
|
| 63516 |
-
"learning_rate": 1.701923076923077e-05,
|
| 63517 |
-
"loss": 0.0971,
|
| 63518 |
-
"step": 10550
|
| 63519 |
-
},
|
| 63520 |
-
{
|
| 63521 |
-
"epoch": 96.8,
|
| 63522 |
-
"learning_rate": 1.6971153846153848e-05,
|
| 63523 |
-
"loss": 0.0951,
|
| 63524 |
-
"step": 10551
|
| 63525 |
-
},
|
| 63526 |
-
{
|
| 63527 |
-
"epoch": 96.81,
|
| 63528 |
-
"learning_rate": 1.6923076923076924e-05,
|
| 63529 |
-
"loss": 0.0813,
|
| 63530 |
-
"step": 10552
|
| 63531 |
-
},
|
| 63532 |
-
{
|
| 63533 |
-
"epoch": 96.82,
|
| 63534 |
-
"learning_rate": 1.6875e-05,
|
| 63535 |
-
"loss": 0.0793,
|
| 63536 |
-
"step": 10553
|
| 63537 |
-
},
|
| 63538 |
-
{
|
| 63539 |
-
"epoch": 96.83,
|
| 63540 |
-
"learning_rate": 1.6826923076923076e-05,
|
| 63541 |
-
"loss": 0.0885,
|
| 63542 |
-
"step": 10554
|
| 63543 |
-
},
|
| 63544 |
-
{
|
| 63545 |
-
"epoch": 96.83,
|
| 63546 |
-
"learning_rate": 1.6778846153846156e-05,
|
| 63547 |
-
"loss": 0.073,
|
| 63548 |
-
"step": 10555
|
| 63549 |
-
},
|
| 63550 |
-
{
|
| 63551 |
-
"epoch": 96.84,
|
| 63552 |
-
"learning_rate": 1.673076923076923e-05,
|
| 63553 |
-
"loss": 0.0892,
|
| 63554 |
-
"step": 10556
|
| 63555 |
-
},
|
| 63556 |
-
{
|
| 63557 |
-
"epoch": 96.85,
|
| 63558 |
-
"learning_rate": 1.6682692307692308e-05,
|
| 63559 |
-
"loss": 0.0787,
|
| 63560 |
-
"step": 10557
|
| 63561 |
-
},
|
| 63562 |
-
{
|
| 63563 |
-
"epoch": 96.86,
|
| 63564 |
-
"learning_rate": 1.6634615384615384e-05,
|
| 63565 |
-
"loss": 0.0782,
|
| 63566 |
-
"step": 10558
|
| 63567 |
-
},
|
| 63568 |
-
{
|
| 63569 |
-
"epoch": 96.87,
|
| 63570 |
-
"learning_rate": 1.6586538461538463e-05,
|
| 63571 |
-
"loss": 0.0622,
|
| 63572 |
-
"step": 10559
|
| 63573 |
-
},
|
| 63574 |
-
{
|
| 63575 |
-
"epoch": 96.88,
|
| 63576 |
-
"learning_rate": 1.653846153846154e-05,
|
| 63577 |
-
"loss": 0.0956,
|
| 63578 |
-
"step": 10560
|
| 63579 |
-
},
|
| 63580 |
-
{
|
| 63581 |
-
"epoch": 96.89,
|
| 63582 |
-
"learning_rate": 1.6490384615384615e-05,
|
| 63583 |
-
"loss": 0.0741,
|
| 63584 |
-
"step": 10561
|
| 63585 |
-
},
|
| 63586 |
-
{
|
| 63587 |
-
"epoch": 96.9,
|
| 63588 |
-
"learning_rate": 1.6442307692307695e-05,
|
| 63589 |
-
"loss": 0.1073,
|
| 63590 |
-
"step": 10562
|
| 63591 |
-
},
|
| 63592 |
-
{
|
| 63593 |
-
"epoch": 96.91,
|
| 63594 |
-
"learning_rate": 1.639423076923077e-05,
|
| 63595 |
-
"loss": 0.0747,
|
| 63596 |
-
"step": 10563
|
| 63597 |
-
},
|
| 63598 |
-
{
|
| 63599 |
-
"epoch": 96.92,
|
| 63600 |
-
"learning_rate": 1.6346153846153847e-05,
|
| 63601 |
-
"loss": 0.1143,
|
| 63602 |
-
"step": 10564
|
| 63603 |
-
},
|
| 63604 |
-
{
|
| 63605 |
-
"epoch": 96.93,
|
| 63606 |
-
"learning_rate": 1.6298076923076923e-05,
|
| 63607 |
-
"loss": 0.0906,
|
| 63608 |
-
"step": 10565
|
| 63609 |
-
},
|
| 63610 |
-
{
|
| 63611 |
-
"epoch": 96.94,
|
| 63612 |
-
"learning_rate": 1.6250000000000002e-05,
|
| 63613 |
-
"loss": 0.1303,
|
| 63614 |
-
"step": 10566
|
| 63615 |
-
},
|
| 63616 |
-
{
|
| 63617 |
-
"epoch": 96.94,
|
| 63618 |
-
"learning_rate": 1.6201923076923075e-05,
|
| 63619 |
-
"loss": 0.1188,
|
| 63620 |
-
"step": 10567
|
| 63621 |
-
},
|
| 63622 |
-
{
|
| 63623 |
-
"epoch": 96.95,
|
| 63624 |
-
"learning_rate": 1.6153846153846154e-05,
|
| 63625 |
-
"loss": 0.0773,
|
| 63626 |
-
"step": 10568
|
| 63627 |
-
},
|
| 63628 |
-
{
|
| 63629 |
-
"epoch": 96.96,
|
| 63630 |
-
"learning_rate": 1.610576923076923e-05,
|
| 63631 |
-
"loss": 0.1376,
|
| 63632 |
-
"step": 10569
|
| 63633 |
-
},
|
| 63634 |
-
{
|
| 63635 |
-
"epoch": 96.97,
|
| 63636 |
-
"learning_rate": 1.605769230769231e-05,
|
| 63637 |
-
"loss": 0.091,
|
| 63638 |
-
"step": 10570
|
| 63639 |
-
},
|
| 63640 |
-
{
|
| 63641 |
-
"epoch": 96.98,
|
| 63642 |
-
"learning_rate": 1.6009615384615385e-05,
|
| 63643 |
-
"loss": 0.0843,
|
| 63644 |
-
"step": 10571
|
| 63645 |
-
},
|
| 63646 |
-
{
|
| 63647 |
-
"epoch": 96.99,
|
| 63648 |
-
"learning_rate": 1.596153846153846e-05,
|
| 63649 |
-
"loss": 0.1066,
|
| 63650 |
-
"step": 10572
|
| 63651 |
-
},
|
| 63652 |
-
{
|
| 63653 |
-
"epoch": 97.0,
|
| 63654 |
-
"learning_rate": 1.591346153846154e-05,
|
| 63655 |
-
"loss": 0.0846,
|
| 63656 |
-
"step": 10573
|
| 63657 |
-
},
|
| 63658 |
-
{
|
| 63659 |
-
"epoch": 97.01,
|
| 63660 |
-
"learning_rate": 1.5865384615384613e-05,
|
| 63661 |
-
"loss": 0.1501,
|
| 63662 |
-
"step": 10574
|
| 63663 |
-
},
|
| 63664 |
-
{
|
| 63665 |
-
"epoch": 97.02,
|
| 63666 |
-
"learning_rate": 1.5817307692307693e-05,
|
| 63667 |
-
"loss": 0.1045,
|
| 63668 |
-
"step": 10575
|
| 63669 |
-
},
|
| 63670 |
-
{
|
| 63671 |
-
"epoch": 97.03,
|
| 63672 |
-
"learning_rate": 1.576923076923077e-05,
|
| 63673 |
-
"loss": 0.1179,
|
| 63674 |
-
"step": 10576
|
| 63675 |
-
},
|
| 63676 |
-
{
|
| 63677 |
-
"epoch": 97.04,
|
| 63678 |
-
"learning_rate": 1.5721153846153848e-05,
|
| 63679 |
-
"loss": 0.1085,
|
| 63680 |
-
"step": 10577
|
| 63681 |
-
},
|
| 63682 |
-
{
|
| 63683 |
-
"epoch": 97.05,
|
| 63684 |
-
"learning_rate": 1.567307692307692e-05,
|
| 63685 |
-
"loss": 0.1578,
|
| 63686 |
-
"step": 10578
|
| 63687 |
-
},
|
| 63688 |
-
{
|
| 63689 |
-
"epoch": 97.06,
|
| 63690 |
-
"learning_rate": 1.5625e-05,
|
| 63691 |
-
"loss": 0.0982,
|
| 63692 |
-
"step": 10579
|
| 63693 |
-
},
|
| 63694 |
-
{
|
| 63695 |
-
"epoch": 97.06,
|
| 63696 |
-
"learning_rate": 1.5576923076923076e-05,
|
| 63697 |
-
"loss": 0.0764,
|
| 63698 |
-
"step": 10580
|
| 63699 |
-
},
|
| 63700 |
-
{
|
| 63701 |
-
"epoch": 97.07,
|
| 63702 |
-
"learning_rate": 1.5528846153846152e-05,
|
| 63703 |
-
"loss": 0.1044,
|
| 63704 |
-
"step": 10581
|
| 63705 |
-
},
|
| 63706 |
-
{
|
| 63707 |
-
"epoch": 97.08,
|
| 63708 |
-
"learning_rate": 1.5480769230769232e-05,
|
| 63709 |
-
"loss": 0.1084,
|
| 63710 |
-
"step": 10582
|
| 63711 |
-
},
|
| 63712 |
-
{
|
| 63713 |
-
"epoch": 97.09,
|
| 63714 |
-
"learning_rate": 1.5432692307692308e-05,
|
| 63715 |
-
"loss": 0.1081,
|
| 63716 |
-
"step": 10583
|
| 63717 |
-
},
|
| 63718 |
-
{
|
| 63719 |
-
"epoch": 97.1,
|
| 63720 |
-
"learning_rate": 1.5384615384615387e-05,
|
| 63721 |
-
"loss": 0.0863,
|
| 63722 |
-
"step": 10584
|
| 63723 |
-
},
|
| 63724 |
-
{
|
| 63725 |
-
"epoch": 97.11,
|
| 63726 |
-
"learning_rate": 1.5336538461538463e-05,
|
| 63727 |
-
"loss": 0.1192,
|
| 63728 |
-
"step": 10585
|
| 63729 |
-
},
|
| 63730 |
-
{
|
| 63731 |
-
"epoch": 97.12,
|
| 63732 |
-
"learning_rate": 1.528846153846154e-05,
|
| 63733 |
-
"loss": 0.125,
|
| 63734 |
-
"step": 10586
|
| 63735 |
-
},
|
| 63736 |
-
{
|
| 63737 |
-
"epoch": 97.13,
|
| 63738 |
-
"learning_rate": 1.5240384615384615e-05,
|
| 63739 |
-
"loss": 0.0623,
|
| 63740 |
-
"step": 10587
|
| 63741 |
-
},
|
| 63742 |
-
{
|
| 63743 |
-
"epoch": 97.14,
|
| 63744 |
-
"learning_rate": 1.5192307692307693e-05,
|
| 63745 |
-
"loss": 0.0727,
|
| 63746 |
-
"step": 10588
|
| 63747 |
-
},
|
| 63748 |
-
{
|
| 63749 |
-
"epoch": 97.15,
|
| 63750 |
-
"learning_rate": 1.5144230769230769e-05,
|
| 63751 |
-
"loss": 0.0703,
|
| 63752 |
-
"step": 10589
|
| 63753 |
-
},
|
| 63754 |
-
{
|
| 63755 |
-
"epoch": 97.16,
|
| 63756 |
-
"learning_rate": 1.5096153846153847e-05,
|
| 63757 |
-
"loss": 0.113,
|
| 63758 |
-
"step": 10590
|
| 63759 |
-
},
|
| 63760 |
-
{
|
| 63761 |
-
"epoch": 97.17,
|
| 63762 |
-
"learning_rate": 1.5048076923076923e-05,
|
| 63763 |
-
"loss": 0.0971,
|
| 63764 |
-
"step": 10591
|
| 63765 |
-
},
|
| 63766 |
-
{
|
| 63767 |
-
"epoch": 97.17,
|
| 63768 |
-
"learning_rate": 1.5e-05,
|
| 63769 |
-
"loss": 0.0691,
|
| 63770 |
-
"step": 10592
|
| 63771 |
-
},
|
| 63772 |
-
{
|
| 63773 |
-
"epoch": 97.18,
|
| 63774 |
-
"learning_rate": 1.4951923076923076e-05,
|
| 63775 |
-
"loss": 0.1618,
|
| 63776 |
-
"step": 10593
|
| 63777 |
-
},
|
| 63778 |
-
{
|
| 63779 |
-
"epoch": 97.19,
|
| 63780 |
-
"learning_rate": 1.4903846153846156e-05,
|
| 63781 |
-
"loss": 0.086,
|
| 63782 |
-
"step": 10594
|
| 63783 |
-
},
|
| 63784 |
-
{
|
| 63785 |
-
"epoch": 97.2,
|
| 63786 |
-
"learning_rate": 1.4855769230769232e-05,
|
| 63787 |
-
"loss": 0.1307,
|
| 63788 |
-
"step": 10595
|
| 63789 |
-
},
|
| 63790 |
-
{
|
| 63791 |
-
"epoch": 97.21,
|
| 63792 |
-
"learning_rate": 1.4807692307692308e-05,
|
| 63793 |
-
"loss": 0.0762,
|
| 63794 |
-
"step": 10596
|
| 63795 |
-
},
|
| 63796 |
-
{
|
| 63797 |
-
"epoch": 97.22,
|
| 63798 |
-
"learning_rate": 1.4759615384615386e-05,
|
| 63799 |
-
"loss": 0.1148,
|
| 63800 |
-
"step": 10597
|
| 63801 |
-
},
|
| 63802 |
-
{
|
| 63803 |
-
"epoch": 97.23,
|
| 63804 |
-
"learning_rate": 1.4711538461538462e-05,
|
| 63805 |
-
"loss": 0.0723,
|
| 63806 |
-
"step": 10598
|
| 63807 |
-
},
|
| 63808 |
-
{
|
| 63809 |
-
"epoch": 97.24,
|
| 63810 |
-
"learning_rate": 1.466346153846154e-05,
|
| 63811 |
-
"loss": 0.0957,
|
| 63812 |
-
"step": 10599
|
| 63813 |
-
},
|
| 63814 |
-
{
|
| 63815 |
-
"epoch": 97.25,
|
| 63816 |
-
"learning_rate": 1.4615384615384615e-05,
|
| 63817 |
-
"loss": 0.085,
|
| 63818 |
-
"step": 10600
|
| 63819 |
-
},
|
| 63820 |
-
{
|
| 63821 |
-
"epoch": 97.26,
|
| 63822 |
-
"learning_rate": 1.4567307692307693e-05,
|
| 63823 |
-
"loss": 0.1458,
|
| 63824 |
-
"step": 10601
|
| 63825 |
-
},
|
| 63826 |
-
{
|
| 63827 |
-
"epoch": 97.27,
|
| 63828 |
-
"learning_rate": 1.4519230769230769e-05,
|
| 63829 |
-
"loss": 0.1268,
|
| 63830 |
-
"step": 10602
|
| 63831 |
-
},
|
| 63832 |
-
{
|
| 63833 |
-
"epoch": 97.28,
|
| 63834 |
-
"learning_rate": 1.4471153846153845e-05,
|
| 63835 |
-
"loss": 0.1339,
|
| 63836 |
-
"step": 10603
|
| 63837 |
-
},
|
| 63838 |
-
{
|
| 63839 |
-
"epoch": 97.28,
|
| 63840 |
-
"learning_rate": 1.4423076923076924e-05,
|
| 63841 |
-
"loss": 0.1093,
|
| 63842 |
-
"step": 10604
|
| 63843 |
-
},
|
| 63844 |
-
{
|
| 63845 |
-
"epoch": 97.29,
|
| 63846 |
-
"learning_rate": 1.4375e-05,
|
| 63847 |
-
"loss": 0.1123,
|
| 63848 |
-
"step": 10605
|
| 63849 |
-
},
|
| 63850 |
-
{
|
| 63851 |
-
"epoch": 97.3,
|
| 63852 |
-
"learning_rate": 1.4326923076923078e-05,
|
| 63853 |
-
"loss": 0.1232,
|
| 63854 |
-
"step": 10606
|
| 63855 |
-
},
|
| 63856 |
-
{
|
| 63857 |
-
"epoch": 97.31,
|
| 63858 |
-
"learning_rate": 1.4278846153846154e-05,
|
| 63859 |
-
"loss": 0.0946,
|
| 63860 |
-
"step": 10607
|
| 63861 |
-
},
|
| 63862 |
-
{
|
| 63863 |
-
"epoch": 97.32,
|
| 63864 |
-
"learning_rate": 1.4230769230769232e-05,
|
| 63865 |
-
"loss": 0.0789,
|
| 63866 |
-
"step": 10608
|
| 63867 |
-
},
|
| 63868 |
-
{
|
| 63869 |
-
"epoch": 97.33,
|
| 63870 |
-
"learning_rate": 1.4182692307692308e-05,
|
| 63871 |
-
"loss": 0.1177,
|
| 63872 |
-
"step": 10609
|
| 63873 |
-
},
|
| 63874 |
-
{
|
| 63875 |
-
"epoch": 97.34,
|
| 63876 |
-
"learning_rate": 1.4134615384615384e-05,
|
| 63877 |
-
"loss": 0.0587,
|
| 63878 |
-
"step": 10610
|
| 63879 |
-
},
|
| 63880 |
-
{
|
| 63881 |
-
"epoch": 97.35,
|
| 63882 |
-
"learning_rate": 1.4086538461538462e-05,
|
| 63883 |
-
"loss": 0.088,
|
| 63884 |
-
"step": 10611
|
| 63885 |
-
},
|
| 63886 |
-
{
|
| 63887 |
-
"epoch": 97.36,
|
| 63888 |
-
"learning_rate": 1.4038461538461538e-05,
|
| 63889 |
-
"loss": 0.0568,
|
| 63890 |
-
"step": 10612
|
| 63891 |
-
},
|
| 63892 |
-
{
|
| 63893 |
-
"epoch": 97.37,
|
| 63894 |
-
"learning_rate": 1.3990384615384615e-05,
|
| 63895 |
-
"loss": 0.096,
|
| 63896 |
-
"step": 10613
|
| 63897 |
-
},
|
| 63898 |
-
{
|
| 63899 |
-
"epoch": 97.38,
|
| 63900 |
-
"learning_rate": 1.3942307692307693e-05,
|
| 63901 |
-
"loss": 0.1626,
|
| 63902 |
-
"step": 10614
|
| 63903 |
-
},
|
| 63904 |
-
{
|
| 63905 |
-
"epoch": 97.39,
|
| 63906 |
-
"learning_rate": 1.389423076923077e-05,
|
| 63907 |
-
"loss": 0.0847,
|
| 63908 |
-
"step": 10615
|
| 63909 |
-
},
|
| 63910 |
-
{
|
| 63911 |
-
"epoch": 97.39,
|
| 63912 |
-
"learning_rate": 1.3846153846153847e-05,
|
| 63913 |
-
"loss": 0.1289,
|
| 63914 |
-
"step": 10616
|
| 63915 |
-
},
|
| 63916 |
-
{
|
| 63917 |
-
"epoch": 97.4,
|
| 63918 |
-
"learning_rate": 1.3798076923076923e-05,
|
| 63919 |
-
"loss": 0.0813,
|
| 63920 |
-
"step": 10617
|
| 63921 |
-
},
|
| 63922 |
-
{
|
| 63923 |
-
"epoch": 97.41,
|
| 63924 |
-
"learning_rate": 1.375e-05,
|
| 63925 |
-
"loss": 0.0931,
|
| 63926 |
-
"step": 10618
|
| 63927 |
-
},
|
| 63928 |
-
{
|
| 63929 |
-
"epoch": 97.42,
|
| 63930 |
-
"learning_rate": 1.3701923076923077e-05,
|
| 63931 |
-
"loss": 0.1229,
|
| 63932 |
-
"step": 10619
|
| 63933 |
-
},
|
| 63934 |
-
{
|
| 63935 |
-
"epoch": 97.43,
|
| 63936 |
-
"learning_rate": 1.3653846153846154e-05,
|
| 63937 |
-
"loss": 0.1042,
|
| 63938 |
-
"step": 10620
|
| 63939 |
-
},
|
| 63940 |
-
{
|
| 63941 |
-
"epoch": 97.44,
|
| 63942 |
-
"learning_rate": 1.360576923076923e-05,
|
| 63943 |
-
"loss": 0.1809,
|
| 63944 |
-
"step": 10621
|
| 63945 |
-
},
|
| 63946 |
-
{
|
| 63947 |
-
"epoch": 97.45,
|
| 63948 |
-
"learning_rate": 1.3557692307692308e-05,
|
| 63949 |
-
"loss": 0.163,
|
| 63950 |
-
"step": 10622
|
| 63951 |
-
},
|
| 63952 |
-
{
|
| 63953 |
-
"epoch": 97.46,
|
| 63954 |
-
"learning_rate": 1.3509615384615384e-05,
|
| 63955 |
-
"loss": 0.0389,
|
| 63956 |
-
"step": 10623
|
| 63957 |
-
},
|
| 63958 |
-
{
|
| 63959 |
-
"epoch": 97.47,
|
| 63960 |
-
"learning_rate": 1.3461538461538463e-05,
|
| 63961 |
-
"loss": 0.1089,
|
| 63962 |
-
"step": 10624
|
| 63963 |
-
},
|
| 63964 |
-
{
|
| 63965 |
-
"epoch": 97.48,
|
| 63966 |
-
"learning_rate": 1.341346153846154e-05,
|
| 63967 |
-
"loss": 0.0626,
|
| 63968 |
-
"step": 10625
|
| 63969 |
-
},
|
| 63970 |
-
{
|
| 63971 |
-
"epoch": 97.49,
|
| 63972 |
-
"learning_rate": 1.3365384615384615e-05,
|
| 63973 |
-
"loss": 0.0596,
|
| 63974 |
-
"step": 10626
|
| 63975 |
-
},
|
| 63976 |
-
{
|
| 63977 |
-
"epoch": 97.5,
|
| 63978 |
-
"learning_rate": 1.3317307692307693e-05,
|
| 63979 |
-
"loss": 0.0524,
|
| 63980 |
-
"step": 10627
|
| 63981 |
-
},
|
| 63982 |
-
{
|
| 63983 |
-
"epoch": 97.5,
|
| 63984 |
-
"learning_rate": 1.3269230769230769e-05,
|
| 63985 |
-
"loss": 0.1355,
|
| 63986 |
-
"step": 10628
|
| 63987 |
-
},
|
| 63988 |
-
{
|
| 63989 |
-
"epoch": 97.51,
|
| 63990 |
-
"learning_rate": 1.3221153846153847e-05,
|
| 63991 |
-
"loss": 0.1323,
|
| 63992 |
-
"step": 10629
|
| 63993 |
-
},
|
| 63994 |
-
{
|
| 63995 |
-
"epoch": 97.52,
|
| 63996 |
-
"learning_rate": 1.3173076923076923e-05,
|
| 63997 |
-
"loss": 0.1384,
|
| 63998 |
-
"step": 10630
|
| 63999 |
-
},
|
| 64000 |
-
{
|
| 64001 |
-
"epoch": 97.53,
|
| 64002 |
-
"learning_rate": 1.3125e-05,
|
| 64003 |
-
"loss": 0.0982,
|
| 64004 |
-
"step": 10631
|
| 64005 |
-
},
|
| 64006 |
-
{
|
| 64007 |
-
"epoch": 97.54,
|
| 64008 |
-
"learning_rate": 1.3076923076923077e-05,
|
| 64009 |
-
"loss": 0.0849,
|
| 64010 |
-
"step": 10632
|
| 64011 |
-
},
|
| 64012 |
-
{
|
| 64013 |
-
"epoch": 97.55,
|
| 64014 |
-
"learning_rate": 1.3028846153846153e-05,
|
| 64015 |
-
"loss": 0.0658,
|
| 64016 |
-
"step": 10633
|
| 64017 |
-
},
|
| 64018 |
-
{
|
| 64019 |
-
"epoch": 97.56,
|
| 64020 |
-
"learning_rate": 1.2980769230769232e-05,
|
| 64021 |
-
"loss": 0.0823,
|
| 64022 |
-
"step": 10634
|
| 64023 |
-
},
|
| 64024 |
-
{
|
| 64025 |
-
"epoch": 97.57,
|
| 64026 |
-
"learning_rate": 1.2932692307692308e-05,
|
| 64027 |
-
"loss": 0.1178,
|
| 64028 |
-
"step": 10635
|
| 64029 |
-
},
|
| 64030 |
-
{
|
| 64031 |
-
"epoch": 97.58,
|
| 64032 |
-
"learning_rate": 1.2884615384615386e-05,
|
| 64033 |
-
"loss": 0.0725,
|
| 64034 |
-
"step": 10636
|
| 64035 |
-
},
|
| 64036 |
-
{
|
| 64037 |
-
"epoch": 97.59,
|
| 64038 |
-
"learning_rate": 1.2836538461538462e-05,
|
| 64039 |
-
"loss": 0.0602,
|
| 64040 |
-
"step": 10637
|
| 64041 |
-
},
|
| 64042 |
-
{
|
| 64043 |
-
"epoch": 97.6,
|
| 64044 |
-
"learning_rate": 1.278846153846154e-05,
|
| 64045 |
-
"loss": 0.0946,
|
| 64046 |
-
"step": 10638
|
| 64047 |
-
},
|
| 64048 |
-
{
|
| 64049 |
-
"epoch": 97.61,
|
| 64050 |
-
"learning_rate": 1.2740384615384615e-05,
|
| 64051 |
-
"loss": 0.0841,
|
| 64052 |
-
"step": 10639
|
| 64053 |
-
},
|
| 64054 |
-
{
|
| 64055 |
-
"epoch": 97.61,
|
| 64056 |
-
"learning_rate": 1.2692307692307691e-05,
|
| 64057 |
-
"loss": 0.0951,
|
| 64058 |
-
"step": 10640
|
| 64059 |
-
},
|
| 64060 |
-
{
|
| 64061 |
-
"epoch": 97.62,
|
| 64062 |
-
"learning_rate": 1.264423076923077e-05,
|
| 64063 |
-
"loss": 0.121,
|
| 64064 |
-
"step": 10641
|
| 64065 |
-
},
|
| 64066 |
-
{
|
| 64067 |
-
"epoch": 97.63,
|
| 64068 |
-
"learning_rate": 1.2596153846153845e-05,
|
| 64069 |
-
"loss": 0.0961,
|
| 64070 |
-
"step": 10642
|
| 64071 |
-
},
|
| 64072 |
-
{
|
| 64073 |
-
"epoch": 97.64,
|
| 64074 |
-
"learning_rate": 1.2548076923076923e-05,
|
| 64075 |
-
"loss": 0.0875,
|
| 64076 |
-
"step": 10643
|
| 64077 |
-
},
|
| 64078 |
-
{
|
| 64079 |
-
"epoch": 97.65,
|
| 64080 |
-
"learning_rate": 1.25e-05,
|
| 64081 |
-
"loss": 0.0905,
|
| 64082 |
-
"step": 10644
|
| 64083 |
-
},
|
| 64084 |
-
{
|
| 64085 |
-
"epoch": 97.66,
|
| 64086 |
-
"learning_rate": 1.2451923076923078e-05,
|
| 64087 |
-
"loss": 0.087,
|
| 64088 |
-
"step": 10645
|
| 64089 |
-
},
|
| 64090 |
-
{
|
| 64091 |
-
"epoch": 97.67,
|
| 64092 |
-
"learning_rate": 1.2403846153846154e-05,
|
| 64093 |
-
"loss": 0.0773,
|
| 64094 |
-
"step": 10646
|
| 64095 |
-
},
|
| 64096 |
-
{
|
| 64097 |
-
"epoch": 97.68,
|
| 64098 |
-
"learning_rate": 1.2355769230769232e-05,
|
| 64099 |
-
"loss": 0.0643,
|
| 64100 |
-
"step": 10647
|
| 64101 |
-
},
|
| 64102 |
-
{
|
| 64103 |
-
"epoch": 97.69,
|
| 64104 |
-
"learning_rate": 1.2307692307692308e-05,
|
| 64105 |
-
"loss": 0.0863,
|
| 64106 |
-
"step": 10648
|
| 64107 |
-
},
|
| 64108 |
-
{
|
| 64109 |
-
"epoch": 97.7,
|
| 64110 |
-
"learning_rate": 1.2259615384615384e-05,
|
| 64111 |
-
"loss": 0.0755,
|
| 64112 |
-
"step": 10649
|
| 64113 |
-
},
|
| 64114 |
-
{
|
| 64115 |
-
"epoch": 97.71,
|
| 64116 |
-
"learning_rate": 1.2211538461538462e-05,
|
| 64117 |
-
"loss": 0.094,
|
| 64118 |
-
"step": 10650
|
| 64119 |
-
},
|
| 64120 |
-
{
|
| 64121 |
-
"epoch": 97.72,
|
| 64122 |
-
"learning_rate": 1.2163461538461538e-05,
|
| 64123 |
-
"loss": 0.0653,
|
| 64124 |
-
"step": 10651
|
| 64125 |
-
},
|
| 64126 |
-
{
|
| 64127 |
-
"epoch": 97.72,
|
| 64128 |
-
"learning_rate": 1.2115384615384615e-05,
|
| 64129 |
-
"loss": 0.0869,
|
| 64130 |
-
"step": 10652
|
| 64131 |
-
},
|
| 64132 |
-
{
|
| 64133 |
-
"epoch": 97.73,
|
| 64134 |
-
"learning_rate": 1.2067307692307693e-05,
|
| 64135 |
-
"loss": 0.0762,
|
| 64136 |
-
"step": 10653
|
| 64137 |
-
},
|
| 64138 |
-
{
|
| 64139 |
-
"epoch": 97.74,
|
| 64140 |
-
"learning_rate": 1.2019230769230771e-05,
|
| 64141 |
-
"loss": 0.0212,
|
| 64142 |
-
"step": 10654
|
| 64143 |
-
},
|
| 64144 |
-
{
|
| 64145 |
-
"epoch": 97.75,
|
| 64146 |
-
"learning_rate": 1.1971153846153847e-05,
|
| 64147 |
-
"loss": 0.1531,
|
| 64148 |
-
"step": 10655
|
| 64149 |
-
},
|
| 64150 |
-
{
|
| 64151 |
-
"epoch": 97.76,
|
| 64152 |
-
"learning_rate": 1.1923076923076923e-05,
|
| 64153 |
-
"loss": 0.1183,
|
| 64154 |
-
"step": 10656
|
| 64155 |
-
},
|
| 64156 |
-
{
|
| 64157 |
-
"epoch": 97.77,
|
| 64158 |
-
"learning_rate": 1.1875e-05,
|
| 64159 |
-
"loss": 0.0916,
|
| 64160 |
-
"step": 10657
|
| 64161 |
-
},
|
| 64162 |
-
{
|
| 64163 |
-
"epoch": 97.78,
|
| 64164 |
-
"learning_rate": 1.1826923076923077e-05,
|
| 64165 |
-
"loss": 0.086,
|
| 64166 |
-
"step": 10658
|
| 64167 |
-
},
|
| 64168 |
-
{
|
| 64169 |
-
"epoch": 97.79,
|
| 64170 |
-
"learning_rate": 1.1778846153846154e-05,
|
| 64171 |
-
"loss": 0.1233,
|
| 64172 |
-
"step": 10659
|
| 64173 |
-
},
|
| 64174 |
-
{
|
| 64175 |
-
"epoch": 97.8,
|
| 64176 |
-
"learning_rate": 1.173076923076923e-05,
|
| 64177 |
-
"loss": 0.0986,
|
| 64178 |
-
"step": 10660
|
| 64179 |
-
},
|
| 64180 |
-
{
|
| 64181 |
-
"epoch": 97.81,
|
| 64182 |
-
"learning_rate": 1.1682692307692308e-05,
|
| 64183 |
-
"loss": 0.1258,
|
| 64184 |
-
"step": 10661
|
| 64185 |
-
},
|
| 64186 |
-
{
|
| 64187 |
-
"epoch": 97.82,
|
| 64188 |
-
"learning_rate": 1.1634615384615384e-05,
|
| 64189 |
-
"loss": 0.1512,
|
| 64190 |
-
"step": 10662
|
| 64191 |
-
},
|
| 64192 |
-
{
|
| 64193 |
-
"epoch": 97.83,
|
| 64194 |
-
"learning_rate": 1.1586538461538462e-05,
|
| 64195 |
-
"loss": 0.0741,
|
| 64196 |
-
"step": 10663
|
| 64197 |
-
},
|
| 64198 |
-
{
|
| 64199 |
-
"epoch": 97.83,
|
| 64200 |
-
"learning_rate": 1.153846153846154e-05,
|
| 64201 |
-
"loss": 0.1224,
|
| 64202 |
-
"step": 10664
|
| 64203 |
-
},
|
| 64204 |
-
{
|
| 64205 |
-
"epoch": 97.84,
|
| 64206 |
-
"learning_rate": 1.1490384615384616e-05,
|
| 64207 |
-
"loss": 0.1029,
|
| 64208 |
-
"step": 10665
|
| 64209 |
-
},
|
| 64210 |
-
{
|
| 64211 |
-
"epoch": 97.85,
|
| 64212 |
-
"learning_rate": 1.1442307692307693e-05,
|
| 64213 |
-
"loss": 0.1199,
|
| 64214 |
-
"step": 10666
|
| 64215 |
-
},
|
| 64216 |
-
{
|
| 64217 |
-
"epoch": 97.86,
|
| 64218 |
-
"learning_rate": 1.139423076923077e-05,
|
| 64219 |
-
"loss": 0.0765,
|
| 64220 |
-
"step": 10667
|
| 64221 |
-
},
|
| 64222 |
-
{
|
| 64223 |
-
"epoch": 97.87,
|
| 64224 |
-
"learning_rate": 1.1346153846153847e-05,
|
| 64225 |
-
"loss": 0.1132,
|
| 64226 |
-
"step": 10668
|
| 64227 |
-
},
|
| 64228 |
-
{
|
| 64229 |
-
"epoch": 97.88,
|
| 64230 |
-
"learning_rate": 1.1298076923076923e-05,
|
| 64231 |
-
"loss": 0.1571,
|
| 64232 |
-
"step": 10669
|
| 64233 |
-
},
|
| 64234 |
-
{
|
| 64235 |
-
"epoch": 97.89,
|
| 64236 |
-
"learning_rate": 1.1249999999999999e-05,
|
| 64237 |
-
"loss": 0.0853,
|
| 64238 |
-
"step": 10670
|
| 64239 |
-
},
|
| 64240 |
-
{
|
| 64241 |
-
"epoch": 97.9,
|
| 64242 |
-
"learning_rate": 1.1201923076923077e-05,
|
| 64243 |
-
"loss": 0.1177,
|
| 64244 |
-
"step": 10671
|
| 64245 |
-
},
|
| 64246 |
-
{
|
| 64247 |
-
"epoch": 97.91,
|
| 64248 |
-
"learning_rate": 1.1153846153846153e-05,
|
| 64249 |
-
"loss": 0.1029,
|
| 64250 |
-
"step": 10672
|
| 64251 |
-
},
|
| 64252 |
-
{
|
| 64253 |
-
"epoch": 97.92,
|
| 64254 |
-
"learning_rate": 1.1105769230769232e-05,
|
| 64255 |
-
"loss": 0.0737,
|
| 64256 |
-
"step": 10673
|
| 64257 |
-
},
|
| 64258 |
-
{
|
| 64259 |
-
"epoch": 97.93,
|
| 64260 |
-
"learning_rate": 1.1057692307692308e-05,
|
| 64261 |
-
"loss": 0.071,
|
| 64262 |
-
"step": 10674
|
| 64263 |
-
},
|
| 64264 |
-
{
|
| 64265 |
-
"epoch": 97.94,
|
| 64266 |
-
"learning_rate": 1.1009615384615386e-05,
|
| 64267 |
-
"loss": 0.0793,
|
| 64268 |
-
"step": 10675
|
| 64269 |
-
},
|
| 64270 |
-
{
|
| 64271 |
-
"epoch": 97.94,
|
| 64272 |
-
"learning_rate": 1.0961538461538462e-05,
|
| 64273 |
-
"loss": 0.0541,
|
| 64274 |
-
"step": 10676
|
| 64275 |
-
},
|
| 64276 |
-
{
|
| 64277 |
-
"epoch": 97.95,
|
| 64278 |
-
"learning_rate": 1.091346153846154e-05,
|
| 64279 |
-
"loss": 0.0715,
|
| 64280 |
-
"step": 10677
|
| 64281 |
-
},
|
| 64282 |
-
{
|
| 64283 |
-
"epoch": 97.96,
|
| 64284 |
-
"learning_rate": 1.0865384615384616e-05,
|
| 64285 |
-
"loss": 0.0896,
|
| 64286 |
-
"step": 10678
|
| 64287 |
-
},
|
| 64288 |
-
{
|
| 64289 |
-
"epoch": 97.97,
|
| 64290 |
-
"learning_rate": 1.0817307692307692e-05,
|
| 64291 |
-
"loss": 0.1121,
|
| 64292 |
-
"step": 10679
|
| 64293 |
-
},
|
| 64294 |
-
{
|
| 64295 |
-
"epoch": 97.98,
|
| 64296 |
-
"learning_rate": 1.076923076923077e-05,
|
| 64297 |
-
"loss": 0.0806,
|
| 64298 |
-
"step": 10680
|
| 64299 |
-
},
|
| 64300 |
-
{
|
| 64301 |
-
"epoch": 97.99,
|
| 64302 |
-
"learning_rate": 1.0721153846153845e-05,
|
| 64303 |
-
"loss": 0.039,
|
| 64304 |
-
"step": 10681
|
| 64305 |
-
},
|
| 64306 |
-
{
|
| 64307 |
-
"epoch": 98.0,
|
| 64308 |
-
"learning_rate": 1.0673076923076923e-05,
|
| 64309 |
-
"loss": 0.0613,
|
| 64310 |
-
"step": 10682
|
| 64311 |
-
},
|
| 64312 |
-
{
|
| 64313 |
-
"epoch": 98.01,
|
| 64314 |
-
"learning_rate": 1.0625e-05,
|
| 64315 |
-
"loss": 0.1759,
|
| 64316 |
-
"step": 10683
|
| 64317 |
-
},
|
| 64318 |
-
{
|
| 64319 |
-
"epoch": 98.02,
|
| 64320 |
-
"learning_rate": 1.0576923076923078e-05,
|
| 64321 |
-
"loss": 0.1243,
|
| 64322 |
-
"step": 10684
|
| 64323 |
-
},
|
| 64324 |
-
{
|
| 64325 |
-
"epoch": 98.03,
|
| 64326 |
-
"learning_rate": 1.0528846153846154e-05,
|
| 64327 |
-
"loss": 0.1195,
|
| 64328 |
-
"step": 10685
|
| 64329 |
-
},
|
| 64330 |
-
{
|
| 64331 |
-
"epoch": 98.04,
|
| 64332 |
-
"learning_rate": 1.048076923076923e-05,
|
| 64333 |
-
"loss": 0.0845,
|
| 64334 |
-
"step": 10686
|
| 64335 |
-
},
|
| 64336 |
-
{
|
| 64337 |
-
"epoch": 98.05,
|
| 64338 |
-
"learning_rate": 1.0432692307692308e-05,
|
| 64339 |
-
"loss": 0.1433,
|
| 64340 |
-
"step": 10687
|
| 64341 |
-
},
|
| 64342 |
-
{
|
| 64343 |
-
"epoch": 98.06,
|
| 64344 |
-
"learning_rate": 1.0384615384615384e-05,
|
| 64345 |
-
"loss": 0.1045,
|
| 64346 |
-
"step": 10688
|
| 64347 |
-
},
|
| 64348 |
-
{
|
| 64349 |
-
"epoch": 98.06,
|
| 64350 |
-
"learning_rate": 1.0336538461538462e-05,
|
| 64351 |
-
"loss": 0.1342,
|
| 64352 |
-
"step": 10689
|
| 64353 |
-
},
|
| 64354 |
-
{
|
| 64355 |
-
"epoch": 98.07,
|
| 64356 |
-
"learning_rate": 1.0288461538461538e-05,
|
| 64357 |
-
"loss": 0.1045,
|
| 64358 |
-
"step": 10690
|
| 64359 |
-
},
|
| 64360 |
-
{
|
| 64361 |
-
"epoch": 98.08,
|
| 64362 |
-
"learning_rate": 1.0240384615384616e-05,
|
| 64363 |
-
"loss": 0.1304,
|
| 64364 |
-
"step": 10691
|
| 64365 |
-
},
|
| 64366 |
-
{
|
| 64367 |
-
"epoch": 98.09,
|
| 64368 |
-
"learning_rate": 1.0192307692307692e-05,
|
| 64369 |
-
"loss": 0.1098,
|
| 64370 |
-
"step": 10692
|
| 64371 |
-
},
|
| 64372 |
-
{
|
| 64373 |
-
"epoch": 98.1,
|
| 64374 |
-
"learning_rate": 1.014423076923077e-05,
|
| 64375 |
-
"loss": 0.1077,
|
| 64376 |
-
"step": 10693
|
| 64377 |
-
},
|
| 64378 |
-
{
|
| 64379 |
-
"epoch": 98.11,
|
| 64380 |
-
"learning_rate": 1.0096153846153847e-05,
|
| 64381 |
-
"loss": 0.1068,
|
| 64382 |
-
"step": 10694
|
| 64383 |
-
},
|
| 64384 |
-
{
|
| 64385 |
-
"epoch": 98.12,
|
| 64386 |
-
"learning_rate": 1.0048076923076923e-05,
|
| 64387 |
-
"loss": 0.1021,
|
| 64388 |
-
"step": 10695
|
| 64389 |
-
},
|
| 64390 |
-
{
|
| 64391 |
-
"epoch": 98.13,
|
| 64392 |
-
"learning_rate": 1e-05,
|
| 64393 |
-
"loss": 0.0923,
|
| 64394 |
-
"step": 10696
|
| 64395 |
-
},
|
| 64396 |
-
{
|
| 64397 |
-
"epoch": 98.14,
|
| 64398 |
-
"learning_rate": 9.951923076923077e-06,
|
| 64399 |
-
"loss": 0.1061,
|
| 64400 |
-
"step": 10697
|
| 64401 |
-
},
|
| 64402 |
-
{
|
| 64403 |
-
"epoch": 98.15,
|
| 64404 |
-
"learning_rate": 9.903846153846155e-06,
|
| 64405 |
-
"loss": 0.1195,
|
| 64406 |
-
"step": 10698
|
| 64407 |
-
},
|
| 64408 |
-
{
|
| 64409 |
-
"epoch": 98.16,
|
| 64410 |
-
"learning_rate": 9.85576923076923e-06,
|
| 64411 |
-
"loss": 0.0543,
|
| 64412 |
-
"step": 10699
|
| 64413 |
-
},
|
| 64414 |
-
{
|
| 64415 |
-
"epoch": 98.17,
|
| 64416 |
-
"learning_rate": 9.807692307692307e-06,
|
| 64417 |
-
"loss": 0.1391,
|
| 64418 |
-
"step": 10700
|
| 64419 |
-
},
|
| 64420 |
-
{
|
| 64421 |
-
"epoch": 98.17,
|
| 64422 |
-
"learning_rate": 9.759615384615384e-06,
|
| 64423 |
-
"loss": 0.0912,
|
| 64424 |
-
"step": 10701
|
| 64425 |
-
},
|
| 64426 |
-
{
|
| 64427 |
-
"epoch": 98.18,
|
| 64428 |
-
"learning_rate": 9.71153846153846e-06,
|
| 64429 |
-
"loss": 0.0989,
|
| 64430 |
-
"step": 10702
|
| 64431 |
-
},
|
| 64432 |
-
{
|
| 64433 |
-
"epoch": 98.19,
|
| 64434 |
-
"learning_rate": 9.66346153846154e-06,
|
| 64435 |
-
"loss": 0.1448,
|
| 64436 |
-
"step": 10703
|
| 64437 |
-
},
|
| 64438 |
-
{
|
| 64439 |
-
"epoch": 98.2,
|
| 64440 |
-
"learning_rate": 9.615384615384616e-06,
|
| 64441 |
-
"loss": 0.0309,
|
| 64442 |
-
"step": 10704
|
| 64443 |
-
},
|
| 64444 |
-
{
|
| 64445 |
-
"epoch": 98.21,
|
| 64446 |
-
"learning_rate": 9.567307692307693e-06,
|
| 64447 |
-
"loss": 0.0515,
|
| 64448 |
-
"step": 10705
|
| 64449 |
-
},
|
| 64450 |
-
{
|
| 64451 |
-
"epoch": 98.22,
|
| 64452 |
-
"learning_rate": 9.51923076923077e-06,
|
| 64453 |
-
"loss": 0.0897,
|
| 64454 |
-
"step": 10706
|
| 64455 |
-
},
|
| 64456 |
-
{
|
| 64457 |
-
"epoch": 98.23,
|
| 64458 |
-
"learning_rate": 9.471153846153847e-06,
|
| 64459 |
-
"loss": 0.0825,
|
| 64460 |
-
"step": 10707
|
| 64461 |
-
},
|
| 64462 |
-
{
|
| 64463 |
-
"epoch": 98.24,
|
| 64464 |
-
"learning_rate": 9.423076923076923e-06,
|
| 64465 |
-
"loss": 0.07,
|
| 64466 |
-
"step": 10708
|
| 64467 |
-
},
|
| 64468 |
-
{
|
| 64469 |
-
"epoch": 98.25,
|
| 64470 |
-
"learning_rate": 9.375e-06,
|
| 64471 |
-
"loss": 0.0725,
|
| 64472 |
-
"step": 10709
|
| 64473 |
-
},
|
| 64474 |
-
{
|
| 64475 |
-
"epoch": 98.26,
|
| 64476 |
-
"learning_rate": 9.326923076923077e-06,
|
| 64477 |
-
"loss": 0.1557,
|
| 64478 |
-
"step": 10710
|
| 64479 |
-
},
|
| 64480 |
-
{
|
| 64481 |
-
"epoch": 98.27,
|
| 64482 |
-
"learning_rate": 9.278846153846153e-06,
|
| 64483 |
-
"loss": 0.1375,
|
| 64484 |
-
"step": 10711
|
| 64485 |
-
},
|
| 64486 |
-
{
|
| 64487 |
-
"epoch": 98.28,
|
| 64488 |
-
"learning_rate": 9.230769230769232e-06,
|
| 64489 |
-
"loss": 0.1014,
|
| 64490 |
-
"step": 10712
|
| 64491 |
-
},
|
| 64492 |
-
{
|
| 64493 |
-
"epoch": 98.28,
|
| 64494 |
-
"learning_rate": 9.182692307692308e-06,
|
| 64495 |
-
"loss": 0.1071,
|
| 64496 |
-
"step": 10713
|
| 64497 |
-
},
|
| 64498 |
-
{
|
| 64499 |
-
"epoch": 98.29,
|
| 64500 |
-
"learning_rate": 9.134615384615386e-06,
|
| 64501 |
-
"loss": 0.1428,
|
| 64502 |
-
"step": 10714
|
| 64503 |
-
},
|
| 64504 |
-
{
|
| 64505 |
-
"epoch": 98.3,
|
| 64506 |
-
"learning_rate": 9.086538461538462e-06,
|
| 64507 |
-
"loss": 0.1182,
|
| 64508 |
-
"step": 10715
|
| 64509 |
-
},
|
| 64510 |
-
{
|
| 64511 |
-
"epoch": 98.31,
|
| 64512 |
-
"learning_rate": 9.038461538461538e-06,
|
| 64513 |
-
"loss": 0.1034,
|
| 64514 |
-
"step": 10716
|
| 64515 |
-
},
|
| 64516 |
-
{
|
| 64517 |
-
"epoch": 98.32,
|
| 64518 |
-
"learning_rate": 8.990384615384616e-06,
|
| 64519 |
-
"loss": 0.1138,
|
| 64520 |
-
"step": 10717
|
| 64521 |
-
},
|
| 64522 |
-
{
|
| 64523 |
-
"epoch": 98.33,
|
| 64524 |
-
"learning_rate": 8.942307692307692e-06,
|
| 64525 |
-
"loss": 0.1355,
|
| 64526 |
-
"step": 10718
|
| 64527 |
-
},
|
| 64528 |
-
{
|
| 64529 |
-
"epoch": 98.34,
|
| 64530 |
-
"learning_rate": 8.89423076923077e-06,
|
| 64531 |
-
"loss": 0.0971,
|
| 64532 |
-
"step": 10719
|
| 64533 |
-
},
|
| 64534 |
-
{
|
| 64535 |
-
"epoch": 98.35,
|
| 64536 |
-
"learning_rate": 8.846153846153846e-06,
|
| 64537 |
-
"loss": 0.1325,
|
| 64538 |
-
"step": 10720
|
| 64539 |
-
},
|
| 64540 |
-
{
|
| 64541 |
-
"epoch": 98.36,
|
| 64542 |
-
"learning_rate": 8.798076923076923e-06,
|
| 64543 |
-
"loss": 0.1429,
|
| 64544 |
-
"step": 10721
|
| 64545 |
-
},
|
| 64546 |
-
{
|
| 64547 |
-
"epoch": 98.37,
|
| 64548 |
-
"learning_rate": 8.750000000000001e-06,
|
| 64549 |
-
"loss": 0.1403,
|
| 64550 |
-
"step": 10722
|
| 64551 |
-
},
|
| 64552 |
-
{
|
| 64553 |
-
"epoch": 98.38,
|
| 64554 |
-
"learning_rate": 8.701923076923077e-06,
|
| 64555 |
-
"loss": 0.1079,
|
| 64556 |
-
"step": 10723
|
| 64557 |
-
},
|
| 64558 |
-
{
|
| 64559 |
-
"epoch": 98.39,
|
| 64560 |
-
"learning_rate": 8.653846153846155e-06,
|
| 64561 |
-
"loss": 0.1083,
|
| 64562 |
-
"step": 10724
|
| 64563 |
-
},
|
| 64564 |
-
{
|
| 64565 |
-
"epoch": 98.39,
|
| 64566 |
-
"learning_rate": 8.60576923076923e-06,
|
| 64567 |
-
"loss": 0.1192,
|
| 64568 |
-
"step": 10725
|
| 64569 |
-
},
|
| 64570 |
-
{
|
| 64571 |
-
"epoch": 98.4,
|
| 64572 |
-
"learning_rate": 8.557692307692308e-06,
|
| 64573 |
-
"loss": 0.1056,
|
| 64574 |
-
"step": 10726
|
| 64575 |
-
},
|
| 64576 |
-
{
|
| 64577 |
-
"epoch": 98.41,
|
| 64578 |
-
"learning_rate": 8.509615384615384e-06,
|
| 64579 |
-
"loss": 0.1136,
|
| 64580 |
-
"step": 10727
|
| 64581 |
-
},
|
| 64582 |
-
{
|
| 64583 |
-
"epoch": 98.42,
|
| 64584 |
-
"learning_rate": 8.461538461538462e-06,
|
| 64585 |
-
"loss": 0.088,
|
| 64586 |
-
"step": 10728
|
| 64587 |
-
},
|
| 64588 |
-
{
|
| 64589 |
-
"epoch": 98.43,
|
| 64590 |
-
"learning_rate": 8.413461538461538e-06,
|
| 64591 |
-
"loss": 0.0983,
|
| 64592 |
-
"step": 10729
|
| 64593 |
-
},
|
| 64594 |
-
{
|
| 64595 |
-
"epoch": 98.44,
|
| 64596 |
-
"learning_rate": 8.365384615384616e-06,
|
| 64597 |
-
"loss": 0.1053,
|
| 64598 |
-
"step": 10730
|
| 64599 |
-
},
|
| 64600 |
-
{
|
| 64601 |
-
"epoch": 98.45,
|
| 64602 |
-
"learning_rate": 8.317307692307692e-06,
|
| 64603 |
-
"loss": 0.0869,
|
| 64604 |
-
"step": 10731
|
| 64605 |
-
},
|
| 64606 |
-
{
|
| 64607 |
-
"epoch": 98.46,
|
| 64608 |
-
"learning_rate": 8.26923076923077e-06,
|
| 64609 |
-
"loss": 0.071,
|
| 64610 |
-
"step": 10732
|
| 64611 |
-
},
|
| 64612 |
-
{
|
| 64613 |
-
"epoch": 98.47,
|
| 64614 |
-
"learning_rate": 8.221153846153847e-06,
|
| 64615 |
-
"loss": 0.1736,
|
| 64616 |
-
"step": 10733
|
| 64617 |
-
},
|
| 64618 |
-
{
|
| 64619 |
-
"epoch": 98.48,
|
| 64620 |
-
"learning_rate": 8.173076923076923e-06,
|
| 64621 |
-
"loss": 0.0977,
|
| 64622 |
-
"step": 10734
|
| 64623 |
-
},
|
| 64624 |
-
{
|
| 64625 |
-
"epoch": 98.49,
|
| 64626 |
-
"learning_rate": 8.125000000000001e-06,
|
| 64627 |
-
"loss": 0.0914,
|
| 64628 |
-
"step": 10735
|
| 64629 |
-
},
|
| 64630 |
-
{
|
| 64631 |
-
"epoch": 98.5,
|
| 64632 |
-
"learning_rate": 8.076923076923077e-06,
|
| 64633 |
-
"loss": 0.0306,
|
| 64634 |
-
"step": 10736
|
| 64635 |
-
},
|
| 64636 |
-
{
|
| 64637 |
-
"epoch": 98.5,
|
| 64638 |
-
"learning_rate": 8.028846153846155e-06,
|
| 64639 |
-
"loss": 0.1421,
|
| 64640 |
-
"step": 10737
|
| 64641 |
-
},
|
| 64642 |
-
{
|
| 64643 |
-
"epoch": 98.51,
|
| 64644 |
-
"learning_rate": 7.98076923076923e-06,
|
| 64645 |
-
"loss": 0.1401,
|
| 64646 |
-
"step": 10738
|
| 64647 |
-
},
|
| 64648 |
-
{
|
| 64649 |
-
"epoch": 98.52,
|
| 64650 |
-
"learning_rate": 7.932692307692307e-06,
|
| 64651 |
-
"loss": 0.1104,
|
| 64652 |
-
"step": 10739
|
| 64653 |
-
},
|
| 64654 |
-
{
|
| 64655 |
-
"epoch": 98.53,
|
| 64656 |
-
"learning_rate": 7.884615384615384e-06,
|
| 64657 |
-
"loss": 0.1197,
|
| 64658 |
-
"step": 10740
|
| 64659 |
-
},
|
| 64660 |
-
{
|
| 64661 |
-
"epoch": 98.54,
|
| 64662 |
-
"learning_rate": 7.83653846153846e-06,
|
| 64663 |
-
"loss": 0.133,
|
| 64664 |
-
"step": 10741
|
| 64665 |
-
},
|
| 64666 |
-
{
|
| 64667 |
-
"epoch": 98.55,
|
| 64668 |
-
"learning_rate": 7.788461538461538e-06,
|
| 64669 |
-
"loss": 0.1172,
|
| 64670 |
-
"step": 10742
|
| 64671 |
-
},
|
| 64672 |
-
{
|
| 64673 |
-
"epoch": 98.56,
|
| 64674 |
-
"learning_rate": 7.740384615384616e-06,
|
| 64675 |
-
"loss": 0.1163,
|
| 64676 |
-
"step": 10743
|
| 64677 |
-
},
|
| 64678 |
-
{
|
| 64679 |
-
"epoch": 98.57,
|
| 64680 |
-
"learning_rate": 7.692307692307694e-06,
|
| 64681 |
-
"loss": 0.1027,
|
| 64682 |
-
"step": 10744
|
| 64683 |
-
},
|
| 64684 |
-
{
|
| 64685 |
-
"epoch": 98.58,
|
| 64686 |
-
"learning_rate": 7.64423076923077e-06,
|
| 64687 |
-
"loss": 0.1016,
|
| 64688 |
-
"step": 10745
|
| 64689 |
-
},
|
| 64690 |
-
{
|
| 64691 |
-
"epoch": 98.59,
|
| 64692 |
-
"learning_rate": 7.5961538461538465e-06,
|
| 64693 |
-
"loss": 0.123,
|
| 64694 |
-
"step": 10746
|
| 64695 |
-
},
|
| 64696 |
-
{
|
| 64697 |
-
"epoch": 98.6,
|
| 64698 |
-
"learning_rate": 7.548076923076923e-06,
|
| 64699 |
-
"loss": 0.135,
|
| 64700 |
-
"step": 10747
|
| 64701 |
-
},
|
| 64702 |
-
{
|
| 64703 |
-
"epoch": 98.61,
|
| 64704 |
-
"learning_rate": 7.5e-06,
|
| 64705 |
-
"loss": 0.1509,
|
| 64706 |
-
"step": 10748
|
| 64707 |
-
},
|
| 64708 |
-
{
|
| 64709 |
-
"epoch": 98.61,
|
| 64710 |
-
"learning_rate": 7.451923076923078e-06,
|
| 64711 |
-
"loss": 0.0623,
|
| 64712 |
-
"step": 10749
|
| 64713 |
-
},
|
| 64714 |
-
{
|
| 64715 |
-
"epoch": 98.62,
|
| 64716 |
-
"learning_rate": 7.403846153846154e-06,
|
| 64717 |
-
"loss": 0.0884,
|
| 64718 |
-
"step": 10750
|
| 64719 |
-
},
|
| 64720 |
-
{
|
| 64721 |
-
"epoch": 98.63,
|
| 64722 |
-
"learning_rate": 7.355769230769231e-06,
|
| 64723 |
-
"loss": 0.1031,
|
| 64724 |
-
"step": 10751
|
| 64725 |
-
},
|
| 64726 |
-
{
|
| 64727 |
-
"epoch": 98.64,
|
| 64728 |
-
"learning_rate": 7.307692307692308e-06,
|
| 64729 |
-
"loss": 0.0448,
|
| 64730 |
-
"step": 10752
|
| 64731 |
-
},
|
| 64732 |
-
{
|
| 64733 |
-
"epoch": 98.65,
|
| 64734 |
-
"learning_rate": 7.2596153846153845e-06,
|
| 64735 |
-
"loss": 0.0493,
|
| 64736 |
-
"step": 10753
|
| 64737 |
-
},
|
| 64738 |
-
{
|
| 64739 |
-
"epoch": 98.66,
|
| 64740 |
-
"learning_rate": 7.211538461538462e-06,
|
| 64741 |
-
"loss": 0.0913,
|
| 64742 |
-
"step": 10754
|
| 64743 |
-
},
|
| 64744 |
-
{
|
| 64745 |
-
"epoch": 98.67,
|
| 64746 |
-
"learning_rate": 7.163461538461539e-06,
|
| 64747 |
-
"loss": 0.0801,
|
| 64748 |
-
"step": 10755
|
| 64749 |
-
},
|
| 64750 |
-
{
|
| 64751 |
-
"epoch": 98.68,
|
| 64752 |
-
"learning_rate": 7.115384615384616e-06,
|
| 64753 |
-
"loss": 0.1115,
|
| 64754 |
-
"step": 10756
|
| 64755 |
-
},
|
| 64756 |
-
{
|
| 64757 |
-
"epoch": 98.69,
|
| 64758 |
-
"learning_rate": 7.067307692307692e-06,
|
| 64759 |
-
"loss": 0.0649,
|
| 64760 |
-
"step": 10757
|
| 64761 |
-
},
|
| 64762 |
-
{
|
| 64763 |
-
"epoch": 98.7,
|
| 64764 |
-
"learning_rate": 7.019230769230769e-06,
|
| 64765 |
-
"loss": 0.1154,
|
| 64766 |
-
"step": 10758
|
| 64767 |
-
},
|
| 64768 |
-
{
|
| 64769 |
-
"epoch": 98.71,
|
| 64770 |
-
"learning_rate": 6.9711538461538465e-06,
|
| 64771 |
-
"loss": 0.0921,
|
| 64772 |
-
"step": 10759
|
| 64773 |
-
},
|
| 64774 |
-
{
|
| 64775 |
-
"epoch": 98.72,
|
| 64776 |
-
"learning_rate": 6.923076923076923e-06,
|
| 64777 |
-
"loss": 0.1203,
|
| 64778 |
-
"step": 10760
|
| 64779 |
-
},
|
| 64780 |
-
{
|
| 64781 |
-
"epoch": 98.72,
|
| 64782 |
-
"learning_rate": 6.875e-06,
|
| 64783 |
-
"loss": 0.07,
|
| 64784 |
-
"step": 10761
|
| 64785 |
-
},
|
| 64786 |
-
{
|
| 64787 |
-
"epoch": 98.73,
|
| 64788 |
-
"learning_rate": 6.826923076923077e-06,
|
| 64789 |
-
"loss": 0.0516,
|
| 64790 |
-
"step": 10762
|
| 64791 |
-
},
|
| 64792 |
-
{
|
| 64793 |
-
"epoch": 98.74,
|
| 64794 |
-
"learning_rate": 6.778846153846154e-06,
|
| 64795 |
-
"loss": 0.0659,
|
| 64796 |
-
"step": 10763
|
| 64797 |
-
},
|
| 64798 |
-
{
|
| 64799 |
-
"epoch": 98.75,
|
| 64800 |
-
"learning_rate": 6.730769230769232e-06,
|
| 64801 |
-
"loss": 0.1547,
|
| 64802 |
-
"step": 10764
|
| 64803 |
-
},
|
| 64804 |
-
{
|
| 64805 |
-
"epoch": 98.76,
|
| 64806 |
-
"learning_rate": 6.682692307692308e-06,
|
| 64807 |
-
"loss": 0.0945,
|
| 64808 |
-
"step": 10765
|
| 64809 |
-
},
|
| 64810 |
-
{
|
| 64811 |
-
"epoch": 98.77,
|
| 64812 |
-
"learning_rate": 6.6346153846153846e-06,
|
| 64813 |
-
"loss": 0.1401,
|
| 64814 |
-
"step": 10766
|
| 64815 |
-
},
|
| 64816 |
-
{
|
| 64817 |
-
"epoch": 98.78,
|
| 64818 |
-
"learning_rate": 6.586538461538461e-06,
|
| 64819 |
-
"loss": 0.1126,
|
| 64820 |
-
"step": 10767
|
| 64821 |
-
},
|
| 64822 |
-
{
|
| 64823 |
-
"epoch": 98.79,
|
| 64824 |
-
"learning_rate": 6.538461538461538e-06,
|
| 64825 |
-
"loss": 0.1139,
|
| 64826 |
-
"step": 10768
|
| 64827 |
-
},
|
| 64828 |
-
{
|
| 64829 |
-
"epoch": 98.8,
|
| 64830 |
-
"learning_rate": 6.490384615384616e-06,
|
| 64831 |
-
"loss": 0.1408,
|
| 64832 |
-
"step": 10769
|
| 64833 |
-
},
|
| 64834 |
-
{
|
| 64835 |
-
"epoch": 98.81,
|
| 64836 |
-
"learning_rate": 6.442307692307693e-06,
|
| 64837 |
-
"loss": 0.112,
|
| 64838 |
-
"step": 10770
|
| 64839 |
-
},
|
| 64840 |
-
{
|
| 64841 |
-
"epoch": 98.82,
|
| 64842 |
-
"learning_rate": 6.39423076923077e-06,
|
| 64843 |
-
"loss": 0.0731,
|
| 64844 |
-
"step": 10771
|
| 64845 |
-
},
|
| 64846 |
-
{
|
| 64847 |
-
"epoch": 98.83,
|
| 64848 |
-
"learning_rate": 6.346153846153846e-06,
|
| 64849 |
-
"loss": 0.1398,
|
| 64850 |
-
"step": 10772
|
| 64851 |
-
},
|
| 64852 |
-
{
|
| 64853 |
-
"epoch": 98.83,
|
| 64854 |
-
"learning_rate": 6.298076923076923e-06,
|
| 64855 |
-
"loss": 0.0956,
|
| 64856 |
-
"step": 10773
|
| 64857 |
-
},
|
| 64858 |
-
{
|
| 64859 |
-
"epoch": 98.84,
|
| 64860 |
-
"learning_rate": 6.25e-06,
|
| 64861 |
-
"loss": 0.1047,
|
| 64862 |
-
"step": 10774
|
| 64863 |
-
},
|
| 64864 |
-
{
|
| 64865 |
-
"epoch": 98.85,
|
| 64866 |
-
"learning_rate": 6.201923076923077e-06,
|
| 64867 |
-
"loss": 0.0632,
|
| 64868 |
-
"step": 10775
|
| 64869 |
-
},
|
| 64870 |
-
{
|
| 64871 |
-
"epoch": 98.86,
|
| 64872 |
-
"learning_rate": 6.153846153846154e-06,
|
| 64873 |
-
"loss": 0.0971,
|
| 64874 |
-
"step": 10776
|
| 64875 |
-
},
|
| 64876 |
-
{
|
| 64877 |
-
"epoch": 98.87,
|
| 64878 |
-
"learning_rate": 6.105769230769231e-06,
|
| 64879 |
-
"loss": 0.1108,
|
| 64880 |
-
"step": 10777
|
| 64881 |
-
},
|
| 64882 |
-
{
|
| 64883 |
-
"epoch": 98.88,
|
| 64884 |
-
"learning_rate": 6.057692307692308e-06,
|
| 64885 |
-
"loss": 0.154,
|
| 64886 |
-
"step": 10778
|
| 64887 |
-
},
|
| 64888 |
-
{
|
| 64889 |
-
"epoch": 98.89,
|
| 64890 |
-
"learning_rate": 6.0096153846153855e-06,
|
| 64891 |
-
"loss": 0.0705,
|
| 64892 |
-
"step": 10779
|
| 64893 |
-
},
|
| 64894 |
-
{
|
| 64895 |
-
"epoch": 98.9,
|
| 64896 |
-
"learning_rate": 5.9615384615384615e-06,
|
| 64897 |
-
"loss": 0.1322,
|
| 64898 |
-
"step": 10780
|
| 64899 |
-
},
|
| 64900 |
-
{
|
| 64901 |
-
"epoch": 98.91,
|
| 64902 |
-
"learning_rate": 5.913461538461538e-06,
|
| 64903 |
-
"loss": 0.0784,
|
| 64904 |
-
"step": 10781
|
| 64905 |
-
},
|
| 64906 |
-
{
|
| 64907 |
-
"epoch": 98.92,
|
| 64908 |
-
"learning_rate": 5.865384615384615e-06,
|
| 64909 |
-
"loss": 0.0775,
|
| 64910 |
-
"step": 10782
|
| 64911 |
-
},
|
| 64912 |
-
{
|
| 64913 |
-
"epoch": 98.93,
|
| 64914 |
-
"learning_rate": 5.817307692307692e-06,
|
| 64915 |
-
"loss": 0.1013,
|
| 64916 |
-
"step": 10783
|
| 64917 |
-
},
|
| 64918 |
-
{
|
| 64919 |
-
"epoch": 98.94,
|
| 64920 |
-
"learning_rate": 5.76923076923077e-06,
|
| 64921 |
-
"loss": 0.0877,
|
| 64922 |
-
"step": 10784
|
| 64923 |
-
},
|
| 64924 |
-
{
|
| 64925 |
-
"epoch": 98.94,
|
| 64926 |
-
"learning_rate": 5.721153846153847e-06,
|
| 64927 |
-
"loss": 0.1856,
|
| 64928 |
-
"step": 10785
|
| 64929 |
-
},
|
| 64930 |
-
{
|
| 64931 |
-
"epoch": 98.95,
|
| 64932 |
-
"learning_rate": 5.6730769230769235e-06,
|
| 64933 |
-
"loss": 0.0554,
|
| 64934 |
-
"step": 10786
|
| 64935 |
-
},
|
| 64936 |
-
{
|
| 64937 |
-
"epoch": 98.96,
|
| 64938 |
-
"learning_rate": 5.6249999999999995e-06,
|
| 64939 |
-
"loss": 0.0956,
|
| 64940 |
-
"step": 10787
|
| 64941 |
-
},
|
| 64942 |
-
{
|
| 64943 |
-
"epoch": 98.97,
|
| 64944 |
-
"learning_rate": 5.576923076923076e-06,
|
| 64945 |
-
"loss": 0.0672,
|
| 64946 |
-
"step": 10788
|
| 64947 |
-
},
|
| 64948 |
-
{
|
| 64949 |
-
"epoch": 98.98,
|
| 64950 |
-
"learning_rate": 5.528846153846154e-06,
|
| 64951 |
-
"loss": 0.1047,
|
| 64952 |
-
"step": 10789
|
| 64953 |
-
},
|
| 64954 |
-
{
|
| 64955 |
-
"epoch": 98.99,
|
| 64956 |
-
"learning_rate": 5.480769230769231e-06,
|
| 64957 |
-
"loss": 0.0257,
|
| 64958 |
-
"step": 10790
|
| 64959 |
-
},
|
| 64960 |
-
{
|
| 64961 |
-
"epoch": 99.0,
|
| 64962 |
-
"learning_rate": 5.432692307692308e-06,
|
| 64963 |
-
"loss": 0.1263,
|
| 64964 |
-
"step": 10791
|
| 64965 |
-
},
|
| 64966 |
-
{
|
| 64967 |
-
"epoch": 99.01,
|
| 64968 |
-
"learning_rate": 5.384615384615385e-06,
|
| 64969 |
-
"loss": 0.1653,
|
| 64970 |
-
"step": 10792
|
| 64971 |
-
},
|
| 64972 |
-
{
|
| 64973 |
-
"epoch": 99.02,
|
| 64974 |
-
"learning_rate": 5.3365384615384615e-06,
|
| 64975 |
-
"loss": 0.1337,
|
| 64976 |
-
"step": 10793
|
| 64977 |
-
},
|
| 64978 |
-
{
|
| 64979 |
-
"epoch": 99.03,
|
| 64980 |
-
"learning_rate": 5.288461538461539e-06,
|
| 64981 |
-
"loss": 0.1166,
|
| 64982 |
-
"step": 10794
|
| 64983 |
-
},
|
| 64984 |
-
{
|
| 64985 |
-
"epoch": 99.04,
|
| 64986 |
-
"learning_rate": 5.240384615384615e-06,
|
| 64987 |
-
"loss": 0.1185,
|
| 64988 |
-
"step": 10795
|
| 64989 |
-
},
|
| 64990 |
-
{
|
| 64991 |
-
"epoch": 99.05,
|
| 64992 |
-
"learning_rate": 5.192307692307692e-06,
|
| 64993 |
-
"loss": 0.1101,
|
| 64994 |
-
"step": 10796
|
| 64995 |
-
},
|
| 64996 |
-
{
|
| 64997 |
-
"epoch": 99.06,
|
| 64998 |
-
"learning_rate": 5.144230769230769e-06,
|
| 64999 |
-
"loss": 0.1459,
|
| 65000 |
-
"step": 10797
|
| 65001 |
-
},
|
| 65002 |
-
{
|
| 65003 |
-
"epoch": 99.06,
|
| 65004 |
-
"learning_rate": 5.096153846153846e-06,
|
| 65005 |
-
"loss": 0.1221,
|
| 65006 |
-
"step": 10798
|
| 65007 |
-
},
|
| 65008 |
-
{
|
| 65009 |
-
"epoch": 99.07,
|
| 65010 |
-
"learning_rate": 5.0480769230769235e-06,
|
| 65011 |
-
"loss": 0.106,
|
| 65012 |
-
"step": 10799
|
| 65013 |
-
},
|
| 65014 |
-
{
|
| 65015 |
-
"epoch": 99.08,
|
| 65016 |
-
"learning_rate": 5e-06,
|
| 65017 |
-
"loss": 0.1366,
|
| 65018 |
-
"step": 10800
|
| 65019 |
-
},
|
| 65020 |
-
{
|
| 65021 |
-
"epoch": 99.09,
|
| 65022 |
-
"learning_rate": 4.951923076923077e-06,
|
| 65023 |
-
"loss": 0.0836,
|
| 65024 |
-
"step": 10801
|
| 65025 |
-
},
|
| 65026 |
-
{
|
| 65027 |
-
"epoch": 99.1,
|
| 65028 |
-
"learning_rate": 4.903846153846153e-06,
|
| 65029 |
-
"loss": 0.1291,
|
| 65030 |
-
"step": 10802
|
| 65031 |
-
},
|
| 65032 |
-
{
|
| 65033 |
-
"epoch": 99.11,
|
| 65034 |
-
"learning_rate": 4.85576923076923e-06,
|
| 65035 |
-
"loss": 0.1328,
|
| 65036 |
-
"step": 10803
|
| 65037 |
-
},
|
| 65038 |
-
{
|
| 65039 |
-
"epoch": 99.12,
|
| 65040 |
-
"learning_rate": 4.807692307692308e-06,
|
| 65041 |
-
"loss": 0.0745,
|
| 65042 |
-
"step": 10804
|
| 65043 |
-
},
|
| 65044 |
-
{
|
| 65045 |
-
"epoch": 99.13,
|
| 65046 |
-
"learning_rate": 4.759615384615385e-06,
|
| 65047 |
-
"loss": 0.1366,
|
| 65048 |
-
"step": 10805
|
| 65049 |
-
},
|
| 65050 |
-
{
|
| 65051 |
-
"epoch": 99.14,
|
| 65052 |
-
"learning_rate": 4.711538461538462e-06,
|
| 65053 |
-
"loss": 0.0914,
|
| 65054 |
-
"step": 10806
|
| 65055 |
-
},
|
| 65056 |
-
{
|
| 65057 |
-
"epoch": 99.15,
|
| 65058 |
-
"learning_rate": 4.6634615384615384e-06,
|
| 65059 |
-
"loss": 0.1165,
|
| 65060 |
-
"step": 10807
|
| 65061 |
-
},
|
| 65062 |
-
{
|
| 65063 |
-
"epoch": 99.16,
|
| 65064 |
-
"learning_rate": 4.615384615384616e-06,
|
| 65065 |
-
"loss": 0.0281,
|
| 65066 |
-
"step": 10808
|
| 65067 |
-
},
|
| 65068 |
-
{
|
| 65069 |
-
"epoch": 99.17,
|
| 65070 |
-
"learning_rate": 4.567307692307693e-06,
|
| 65071 |
-
"loss": 0.1348,
|
| 65072 |
-
"step": 10809
|
| 65073 |
-
},
|
| 65074 |
-
{
|
| 65075 |
-
"epoch": 99.17,
|
| 65076 |
-
"learning_rate": 4.519230769230769e-06,
|
| 65077 |
-
"loss": 0.0689,
|
| 65078 |
-
"step": 10810
|
| 65079 |
-
},
|
| 65080 |
-
{
|
| 65081 |
-
"epoch": 99.18,
|
| 65082 |
-
"learning_rate": 4.471153846153846e-06,
|
| 65083 |
-
"loss": 0.1174,
|
| 65084 |
-
"step": 10811
|
| 65085 |
-
},
|
| 65086 |
-
{
|
| 65087 |
-
"epoch": 99.19,
|
| 65088 |
-
"learning_rate": 4.423076923076923e-06,
|
| 65089 |
-
"loss": 0.079,
|
| 65090 |
-
"step": 10812
|
| 65091 |
-
},
|
| 65092 |
-
{
|
| 65093 |
-
"epoch": 99.2,
|
| 65094 |
-
"learning_rate": 4.3750000000000005e-06,
|
| 65095 |
-
"loss": 0.1269,
|
| 65096 |
-
"step": 10813
|
| 65097 |
-
},
|
| 65098 |
-
{
|
| 65099 |
-
"epoch": 99.21,
|
| 65100 |
-
"learning_rate": 4.326923076923077e-06,
|
| 65101 |
-
"loss": 0.0556,
|
| 65102 |
-
"step": 10814
|
| 65103 |
-
},
|
| 65104 |
-
{
|
| 65105 |
-
"epoch": 99.22,
|
| 65106 |
-
"learning_rate": 4.278846153846154e-06,
|
| 65107 |
-
"loss": 0.0863,
|
| 65108 |
-
"step": 10815
|
| 65109 |
-
},
|
| 65110 |
-
{
|
| 65111 |
-
"epoch": 99.23,
|
| 65112 |
-
"learning_rate": 4.230769230769231e-06,
|
| 65113 |
-
"loss": 0.0922,
|
| 65114 |
-
"step": 10816
|
| 65115 |
-
},
|
| 65116 |
-
{
|
| 65117 |
-
"epoch": 99.24,
|
| 65118 |
-
"learning_rate": 4.182692307692308e-06,
|
| 65119 |
-
"loss": 0.0968,
|
| 65120 |
-
"step": 10817
|
| 65121 |
-
},
|
| 65122 |
-
{
|
| 65123 |
-
"epoch": 99.25,
|
| 65124 |
-
"learning_rate": 4.134615384615385e-06,
|
| 65125 |
-
"loss": 0.077,
|
| 65126 |
-
"step": 10818
|
| 65127 |
-
},
|
| 65128 |
-
{
|
| 65129 |
-
"epoch": 99.26,
|
| 65130 |
-
"learning_rate": 4.086538461538462e-06,
|
| 65131 |
-
"loss": 0.1543,
|
| 65132 |
-
"step": 10819
|
| 65133 |
-
},
|
| 65134 |
-
{
|
| 65135 |
-
"epoch": 99.27,
|
| 65136 |
-
"learning_rate": 4.0384615384615385e-06,
|
| 65137 |
-
"loss": 0.1635,
|
| 65138 |
-
"step": 10820
|
| 65139 |
-
},
|
| 65140 |
-
{
|
| 65141 |
-
"epoch": 99.28,
|
| 65142 |
-
"learning_rate": 3.990384615384615e-06,
|
| 65143 |
-
"loss": 0.0929,
|
| 65144 |
-
"step": 10821
|
| 65145 |
-
},
|
| 65146 |
-
{
|
| 65147 |
-
"epoch": 99.28,
|
| 65148 |
-
"learning_rate": 3.942307692307692e-06,
|
| 65149 |
-
"loss": 0.1135,
|
| 65150 |
-
"step": 10822
|
| 65151 |
-
},
|
| 65152 |
-
{
|
| 65153 |
-
"epoch": 99.29,
|
| 65154 |
-
"learning_rate": 3.894230769230769e-06,
|
| 65155 |
-
"loss": 0.1106,
|
| 65156 |
-
"step": 10823
|
| 65157 |
-
},
|
| 65158 |
-
{
|
| 65159 |
-
"epoch": 99.3,
|
| 65160 |
-
"learning_rate": 3.846153846153847e-06,
|
| 65161 |
-
"loss": 0.1107,
|
| 65162 |
-
"step": 10824
|
| 65163 |
-
},
|
| 65164 |
-
{
|
| 65165 |
-
"epoch": 99.31,
|
| 65166 |
-
"learning_rate": 3.7980769230769232e-06,
|
| 65167 |
-
"loss": 0.1289,
|
| 65168 |
-
"step": 10825
|
| 65169 |
-
},
|
| 65170 |
-
{
|
| 65171 |
-
"epoch": 99.32,
|
| 65172 |
-
"learning_rate": 3.75e-06,
|
| 65173 |
-
"loss": 0.0919,
|
| 65174 |
-
"step": 10826
|
| 65175 |
-
},
|
| 65176 |
-
{
|
| 65177 |
-
"epoch": 99.33,
|
| 65178 |
-
"learning_rate": 3.701923076923077e-06,
|
| 65179 |
-
"loss": 0.1318,
|
| 65180 |
-
"step": 10827
|
| 65181 |
-
},
|
| 65182 |
-
{
|
| 65183 |
-
"epoch": 99.34,
|
| 65184 |
-
"learning_rate": 3.653846153846154e-06,
|
| 65185 |
-
"loss": 0.0655,
|
| 65186 |
-
"step": 10828
|
| 65187 |
-
},
|
| 65188 |
-
{
|
| 65189 |
-
"epoch": 99.35,
|
| 65190 |
-
"learning_rate": 3.605769230769231e-06,
|
| 65191 |
-
"loss": 0.0838,
|
| 65192 |
-
"step": 10829
|
| 65193 |
-
},
|
| 65194 |
-
{
|
| 65195 |
-
"epoch": 99.36,
|
| 65196 |
-
"learning_rate": 3.557692307692308e-06,
|
| 65197 |
-
"loss": 0.0583,
|
| 65198 |
-
"step": 10830
|
| 65199 |
-
},
|
| 65200 |
-
{
|
| 65201 |
-
"epoch": 99.37,
|
| 65202 |
-
"learning_rate": 3.5096153846153844e-06,
|
| 65203 |
-
"loss": 0.0635,
|
| 65204 |
-
"step": 10831
|
| 65205 |
-
},
|
| 65206 |
-
{
|
| 65207 |
-
"epoch": 99.38,
|
| 65208 |
-
"learning_rate": 3.4615384615384617e-06,
|
| 65209 |
-
"loss": 0.1053,
|
| 65210 |
-
"step": 10832
|
| 65211 |
-
},
|
| 65212 |
-
{
|
| 65213 |
-
"epoch": 99.39,
|
| 65214 |
-
"learning_rate": 3.4134615384615386e-06,
|
| 65215 |
-
"loss": 0.1396,
|
| 65216 |
-
"step": 10833
|
| 65217 |
-
},
|
| 65218 |
-
{
|
| 65219 |
-
"epoch": 99.39,
|
| 65220 |
-
"learning_rate": 3.365384615384616e-06,
|
| 65221 |
-
"loss": 0.0984,
|
| 65222 |
-
"step": 10834
|
| 65223 |
-
},
|
| 65224 |
-
{
|
| 65225 |
-
"epoch": 99.4,
|
| 65226 |
-
"learning_rate": 3.3173076923076923e-06,
|
| 65227 |
-
"loss": 0.0749,
|
| 65228 |
-
"step": 10835
|
| 65229 |
-
},
|
| 65230 |
-
{
|
| 65231 |
-
"epoch": 99.41,
|
| 65232 |
-
"learning_rate": 3.269230769230769e-06,
|
| 65233 |
-
"loss": 0.1174,
|
| 65234 |
-
"step": 10836
|
| 65235 |
-
},
|
| 65236 |
-
{
|
| 65237 |
-
"epoch": 99.42,
|
| 65238 |
-
"learning_rate": 3.2211538461538464e-06,
|
| 65239 |
-
"loss": 0.0928,
|
| 65240 |
-
"step": 10837
|
| 65241 |
-
},
|
| 65242 |
-
{
|
| 65243 |
-
"epoch": 99.43,
|
| 65244 |
-
"learning_rate": 3.173076923076923e-06,
|
| 65245 |
-
"loss": 0.0549,
|
| 65246 |
-
"step": 10838
|
| 65247 |
-
},
|
| 65248 |
-
{
|
| 65249 |
-
"epoch": 99.44,
|
| 65250 |
-
"learning_rate": 3.125e-06,
|
| 65251 |
-
"loss": 0.0872,
|
| 65252 |
-
"step": 10839
|
| 65253 |
-
},
|
| 65254 |
-
{
|
| 65255 |
-
"epoch": 99.45,
|
| 65256 |
-
"learning_rate": 3.076923076923077e-06,
|
| 65257 |
-
"loss": 0.0704,
|
| 65258 |
-
"step": 10840
|
| 65259 |
-
},
|
| 65260 |
-
{
|
| 65261 |
-
"epoch": 99.46,
|
| 65262 |
-
"learning_rate": 3.028846153846154e-06,
|
| 65263 |
-
"loss": 0.0398,
|
| 65264 |
-
"step": 10841
|
| 65265 |
-
},
|
| 65266 |
-
{
|
| 65267 |
-
"epoch": 99.47,
|
| 65268 |
-
"learning_rate": 2.9807692307692307e-06,
|
| 65269 |
-
"loss": 0.094,
|
| 65270 |
-
"step": 10842
|
| 65271 |
-
},
|
| 65272 |
-
{
|
| 65273 |
-
"epoch": 99.48,
|
| 65274 |
-
"learning_rate": 2.9326923076923076e-06,
|
| 65275 |
-
"loss": 0.0902,
|
| 65276 |
-
"step": 10843
|
| 65277 |
-
},
|
| 65278 |
-
{
|
| 65279 |
-
"epoch": 99.49,
|
| 65280 |
-
"learning_rate": 2.884615384615385e-06,
|
| 65281 |
-
"loss": 0.0994,
|
| 65282 |
-
"step": 10844
|
| 65283 |
-
},
|
| 65284 |
-
{
|
| 65285 |
-
"epoch": 99.5,
|
| 65286 |
-
"learning_rate": 2.8365384615384617e-06,
|
| 65287 |
-
"loss": 0.0513,
|
| 65288 |
-
"step": 10845
|
| 65289 |
-
},
|
| 65290 |
-
{
|
| 65291 |
-
"epoch": 99.5,
|
| 65292 |
-
"learning_rate": 2.788461538461538e-06,
|
| 65293 |
-
"loss": 0.2055,
|
| 65294 |
-
"step": 10846
|
| 65295 |
-
},
|
| 65296 |
-
{
|
| 65297 |
-
"epoch": 99.51,
|
| 65298 |
-
"learning_rate": 2.7403846153846155e-06,
|
| 65299 |
-
"loss": 0.1316,
|
| 65300 |
-
"step": 10847
|
| 65301 |
-
},
|
| 65302 |
-
{
|
| 65303 |
-
"epoch": 99.52,
|
| 65304 |
-
"learning_rate": 2.6923076923076923e-06,
|
| 65305 |
-
"loss": 0.1219,
|
| 65306 |
-
"step": 10848
|
| 65307 |
-
},
|
| 65308 |
-
{
|
| 65309 |
-
"epoch": 99.53,
|
| 65310 |
-
"learning_rate": 2.6442307692307696e-06,
|
| 65311 |
-
"loss": 0.1132,
|
| 65312 |
-
"step": 10849
|
| 65313 |
-
},
|
| 65314 |
-
{
|
| 65315 |
-
"epoch": 99.54,
|
| 65316 |
-
"learning_rate": 2.596153846153846e-06,
|
| 65317 |
-
"loss": 0.1072,
|
| 65318 |
-
"step": 10850
|
| 65319 |
-
},
|
| 65320 |
-
{
|
| 65321 |
-
"epoch": 99.55,
|
| 65322 |
-
"learning_rate": 2.548076923076923e-06,
|
| 65323 |
-
"loss": 0.0983,
|
| 65324 |
-
"step": 10851
|
| 65325 |
-
},
|
| 65326 |
-
{
|
| 65327 |
-
"epoch": 99.56,
|
| 65328 |
-
"learning_rate": 2.5e-06,
|
| 65329 |
-
"loss": 0.1457,
|
| 65330 |
-
"step": 10852
|
| 65331 |
-
},
|
| 65332 |
-
{
|
| 65333 |
-
"epoch": 99.57,
|
| 65334 |
-
"learning_rate": 2.4519230769230766e-06,
|
| 65335 |
-
"loss": 0.0745,
|
| 65336 |
-
"step": 10853
|
| 65337 |
-
},
|
| 65338 |
-
{
|
| 65339 |
-
"epoch": 99.58,
|
| 65340 |
-
"learning_rate": 2.403846153846154e-06,
|
| 65341 |
-
"loss": 0.0876,
|
| 65342 |
-
"step": 10854
|
| 65343 |
-
},
|
| 65344 |
-
{
|
| 65345 |
-
"epoch": 99.59,
|
| 65346 |
-
"learning_rate": 2.355769230769231e-06,
|
| 65347 |
-
"loss": 0.1684,
|
| 65348 |
-
"step": 10855
|
| 65349 |
-
},
|
| 65350 |
-
{
|
| 65351 |
-
"epoch": 99.6,
|
| 65352 |
-
"learning_rate": 2.307692307692308e-06,
|
| 65353 |
-
"loss": 0.0859,
|
| 65354 |
-
"step": 10856
|
| 65355 |
-
},
|
| 65356 |
-
{
|
| 65357 |
-
"epoch": 99.61,
|
| 65358 |
-
"learning_rate": 2.2596153846153845e-06,
|
| 65359 |
-
"loss": 0.1476,
|
| 65360 |
-
"step": 10857
|
| 65361 |
-
},
|
| 65362 |
-
{
|
| 65363 |
-
"epoch": 99.61,
|
| 65364 |
-
"learning_rate": 2.2115384615384614e-06,
|
| 65365 |
-
"loss": 0.0793,
|
| 65366 |
-
"step": 10858
|
| 65367 |
-
},
|
| 65368 |
-
{
|
| 65369 |
-
"epoch": 99.62,
|
| 65370 |
-
"learning_rate": 2.1634615384615387e-06,
|
| 65371 |
-
"loss": 0.1262,
|
| 65372 |
-
"step": 10859
|
| 65373 |
-
},
|
| 65374 |
-
{
|
| 65375 |
-
"epoch": 99.63,
|
| 65376 |
-
"learning_rate": 2.1153846153846155e-06,
|
| 65377 |
-
"loss": 0.1145,
|
| 65378 |
-
"step": 10860
|
| 65379 |
-
},
|
| 65380 |
-
{
|
| 65381 |
-
"epoch": 99.64,
|
| 65382 |
-
"learning_rate": 2.0673076923076924e-06,
|
| 65383 |
-
"loss": 0.0778,
|
| 65384 |
-
"step": 10861
|
| 65385 |
-
},
|
| 65386 |
-
{
|
| 65387 |
-
"epoch": 99.65,
|
| 65388 |
-
"learning_rate": 2.0192307692307692e-06,
|
| 65389 |
-
"loss": 0.067,
|
| 65390 |
-
"step": 10862
|
| 65391 |
-
},
|
| 65392 |
-
{
|
| 65393 |
-
"epoch": 99.66,
|
| 65394 |
-
"learning_rate": 1.971153846153846e-06,
|
| 65395 |
-
"loss": 0.0686,
|
| 65396 |
-
"step": 10863
|
| 65397 |
-
},
|
| 65398 |
-
{
|
| 65399 |
-
"epoch": 99.67,
|
| 65400 |
-
"learning_rate": 1.9230769230769234e-06,
|
| 65401 |
-
"loss": 0.0865,
|
| 65402 |
-
"step": 10864
|
| 65403 |
-
},
|
| 65404 |
-
{
|
| 65405 |
-
"epoch": 99.68,
|
| 65406 |
-
"learning_rate": 1.875e-06,
|
| 65407 |
-
"loss": 0.1209,
|
| 65408 |
-
"step": 10865
|
| 65409 |
-
},
|
| 65410 |
-
{
|
| 65411 |
-
"epoch": 99.69,
|
| 65412 |
-
"learning_rate": 1.826923076923077e-06,
|
| 65413 |
-
"loss": 0.0729,
|
| 65414 |
-
"step": 10866
|
| 65415 |
-
},
|
| 65416 |
-
{
|
| 65417 |
-
"epoch": 99.7,
|
| 65418 |
-
"learning_rate": 1.778846153846154e-06,
|
| 65419 |
-
"loss": 0.1472,
|
| 65420 |
-
"step": 10867
|
| 65421 |
-
},
|
| 65422 |
-
{
|
| 65423 |
-
"epoch": 99.71,
|
| 65424 |
-
"learning_rate": 1.7307692307692308e-06,
|
| 65425 |
-
"loss": 0.0516,
|
| 65426 |
-
"step": 10868
|
| 65427 |
-
},
|
| 65428 |
-
{
|
| 65429 |
-
"epoch": 99.72,
|
| 65430 |
-
"learning_rate": 1.682692307692308e-06,
|
| 65431 |
-
"loss": 0.0901,
|
| 65432 |
-
"step": 10869
|
| 65433 |
-
},
|
| 65434 |
-
{
|
| 65435 |
-
"epoch": 99.72,
|
| 65436 |
-
"learning_rate": 1.6346153846153846e-06,
|
| 65437 |
-
"loss": 0.0505,
|
| 65438 |
-
"step": 10870
|
| 65439 |
-
},
|
| 65440 |
-
{
|
| 65441 |
-
"epoch": 99.73,
|
| 65442 |
-
"learning_rate": 1.5865384615384614e-06,
|
| 65443 |
-
"loss": 0.0629,
|
| 65444 |
-
"step": 10871
|
| 65445 |
-
},
|
| 65446 |
-
{
|
| 65447 |
-
"epoch": 99.74,
|
| 65448 |
-
"learning_rate": 1.5384615384615385e-06,
|
| 65449 |
-
"loss": 0.2219,
|
| 65450 |
-
"step": 10872
|
| 65451 |
-
},
|
| 65452 |
-
{
|
| 65453 |
-
"epoch": 99.75,
|
| 65454 |
-
"learning_rate": 1.4903846153846154e-06,
|
| 65455 |
-
"loss": 0.1274,
|
| 65456 |
-
"step": 10873
|
| 65457 |
-
},
|
| 65458 |
-
{
|
| 65459 |
-
"epoch": 99.76,
|
| 65460 |
-
"learning_rate": 1.4423076923076924e-06,
|
| 65461 |
-
"loss": 0.1312,
|
| 65462 |
-
"step": 10874
|
| 65463 |
-
},
|
| 65464 |
-
{
|
| 65465 |
-
"epoch": 99.77,
|
| 65466 |
-
"learning_rate": 1.394230769230769e-06,
|
| 65467 |
-
"loss": 0.1167,
|
| 65468 |
-
"step": 10875
|
| 65469 |
-
},
|
| 65470 |
-
{
|
| 65471 |
-
"epoch": 99.78,
|
| 65472 |
-
"learning_rate": 1.3461538461538462e-06,
|
| 65473 |
-
"loss": 0.0919,
|
| 65474 |
-
"step": 10876
|
| 65475 |
-
},
|
| 65476 |
-
{
|
| 65477 |
-
"epoch": 99.79,
|
| 65478 |
-
"learning_rate": 1.298076923076923e-06,
|
| 65479 |
-
"loss": 0.17,
|
| 65480 |
-
"step": 10877
|
| 65481 |
-
},
|
| 65482 |
-
{
|
| 65483 |
-
"epoch": 99.8,
|
| 65484 |
-
"learning_rate": 1.25e-06,
|
| 65485 |
-
"loss": 0.1601,
|
| 65486 |
-
"step": 10878
|
| 65487 |
-
},
|
| 65488 |
-
{
|
| 65489 |
-
"epoch": 99.81,
|
| 65490 |
-
"learning_rate": 1.201923076923077e-06,
|
| 65491 |
-
"loss": 0.0664,
|
| 65492 |
-
"step": 10879
|
| 65493 |
-
},
|
| 65494 |
-
{
|
| 65495 |
-
"epoch": 99.82,
|
| 65496 |
-
"learning_rate": 1.153846153846154e-06,
|
| 65497 |
-
"loss": 0.08,
|
| 65498 |
-
"step": 10880
|
| 65499 |
-
},
|
| 65500 |
-
{
|
| 65501 |
-
"epoch": 99.83,
|
| 65502 |
-
"learning_rate": 1.1057692307692307e-06,
|
| 65503 |
-
"loss": 0.1484,
|
| 65504 |
-
"step": 10881
|
| 65505 |
-
},
|
| 65506 |
-
{
|
| 65507 |
-
"epoch": 99.83,
|
| 65508 |
-
"learning_rate": 1.0576923076923078e-06,
|
| 65509 |
-
"loss": 0.1291,
|
| 65510 |
-
"step": 10882
|
| 65511 |
-
},
|
| 65512 |
-
{
|
| 65513 |
-
"epoch": 99.84,
|
| 65514 |
-
"learning_rate": 1.0096153846153846e-06,
|
| 65515 |
-
"loss": 0.1387,
|
| 65516 |
-
"step": 10883
|
| 65517 |
-
},
|
| 65518 |
-
{
|
| 65519 |
-
"epoch": 99.85,
|
| 65520 |
-
"learning_rate": 9.615384615384617e-07,
|
| 65521 |
-
"loss": 0.1047,
|
| 65522 |
-
"step": 10884
|
| 65523 |
-
},
|
| 65524 |
-
{
|
| 65525 |
-
"epoch": 99.86,
|
| 65526 |
-
"learning_rate": 9.134615384615385e-07,
|
| 65527 |
-
"loss": 0.1096,
|
| 65528 |
-
"step": 10885
|
| 65529 |
-
},
|
| 65530 |
-
{
|
| 65531 |
-
"epoch": 99.87,
|
| 65532 |
-
"learning_rate": 8.653846153846154e-07,
|
| 65533 |
-
"loss": 0.1193,
|
| 65534 |
-
"step": 10886
|
| 65535 |
-
},
|
| 65536 |
-
{
|
| 65537 |
-
"epoch": 99.88,
|
| 65538 |
-
"learning_rate": 8.173076923076923e-07,
|
| 65539 |
-
"loss": 0.1043,
|
| 65540 |
-
"step": 10887
|
| 65541 |
-
},
|
| 65542 |
-
{
|
| 65543 |
-
"epoch": 99.89,
|
| 65544 |
-
"learning_rate": 7.692307692307693e-07,
|
| 65545 |
-
"loss": 0.178,
|
| 65546 |
-
"step": 10888
|
| 65547 |
-
},
|
| 65548 |
-
{
|
| 65549 |
-
"epoch": 99.9,
|
| 65550 |
-
"learning_rate": 7.211538461538462e-07,
|
| 65551 |
-
"loss": 0.0839,
|
| 65552 |
-
"step": 10889
|
| 65553 |
-
},
|
| 65554 |
-
{
|
| 65555 |
-
"epoch": 99.91,
|
| 65556 |
-
"learning_rate": 6.730769230769231e-07,
|
| 65557 |
-
"loss": 0.0829,
|
| 65558 |
-
"step": 10890
|
| 65559 |
-
},
|
| 65560 |
-
{
|
| 65561 |
-
"epoch": 99.92,
|
| 65562 |
-
"learning_rate": 6.25e-07,
|
| 65563 |
-
"loss": 0.1001,
|
| 65564 |
-
"step": 10891
|
| 65565 |
-
},
|
| 65566 |
-
{
|
| 65567 |
-
"epoch": 99.93,
|
| 65568 |
-
"learning_rate": 5.76923076923077e-07,
|
| 65569 |
-
"loss": 0.0748,
|
| 65570 |
-
"step": 10892
|
| 65571 |
-
},
|
| 65572 |
-
{
|
| 65573 |
-
"epoch": 99.94,
|
| 65574 |
-
"learning_rate": 5.288461538461539e-07,
|
| 65575 |
-
"loss": 0.0612,
|
| 65576 |
-
"step": 10893
|
| 65577 |
-
},
|
| 65578 |
-
{
|
| 65579 |
-
"epoch": 99.94,
|
| 65580 |
-
"learning_rate": 4.807692307692308e-07,
|
| 65581 |
-
"loss": 0.1034,
|
| 65582 |
-
"step": 10894
|
| 65583 |
-
},
|
| 65584 |
-
{
|
| 65585 |
-
"epoch": 99.95,
|
| 65586 |
-
"learning_rate": 4.326923076923077e-07,
|
| 65587 |
-
"loss": 0.0369,
|
| 65588 |
-
"step": 10895
|
| 65589 |
-
},
|
| 65590 |
-
{
|
| 65591 |
-
"epoch": 99.96,
|
| 65592 |
-
"learning_rate": 3.8461538461538463e-07,
|
| 65593 |
-
"loss": 0.1068,
|
| 65594 |
-
"step": 10896
|
| 65595 |
-
},
|
| 65596 |
-
{
|
| 65597 |
-
"epoch": 99.97,
|
| 65598 |
-
"learning_rate": 3.3653846153846154e-07,
|
| 65599 |
-
"loss": 0.0879,
|
| 65600 |
-
"step": 10897
|
| 65601 |
-
},
|
| 65602 |
-
{
|
| 65603 |
-
"epoch": 99.98,
|
| 65604 |
-
"learning_rate": 2.884615384615385e-07,
|
| 65605 |
-
"loss": 0.1357,
|
| 65606 |
-
"step": 10898
|
| 65607 |
-
},
|
| 65608 |
-
{
|
| 65609 |
-
"epoch": 99.99,
|
| 65610 |
-
"learning_rate": 2.403846153846154e-07,
|
| 65611 |
-
"loss": 0.0615,
|
| 65612 |
-
"step": 10899
|
| 65613 |
-
},
|
| 65614 |
-
{
|
| 65615 |
-
"epoch": 100.0,
|
| 65616 |
-
"learning_rate": 1.9230769230769231e-07,
|
| 65617 |
-
"loss": 0.0967,
|
| 65618 |
-
"step": 10900
|
| 65619 |
-
},
|
| 65620 |
-
{
|
| 65621 |
-
"epoch": 100.0,
|
| 65622 |
-
"step": 10900,
|
| 65623 |
-
"total_flos": 1.3360714276118677e+20,
|
| 65624 |
-
"train_loss": 0.44451899920906357,
|
| 65625 |
-
"train_runtime": 28212.0393,
|
| 65626 |
-
"train_samples_per_second": 12.328,
|
| 65627 |
-
"train_steps_per_second": 0.386
|
| 65628 |
}
|
| 65629 |
],
|
| 65630 |
-
"max_steps":
|
| 65631 |
-
"num_train_epochs":
|
| 65632 |
-
"total_flos": 1.
|
| 65633 |
"trial_name": null,
|
| 65634 |
"trial_params": null
|
| 65635 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 96.3302752293578,
|
| 5 |
+
"global_step": 10500,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 63218 |
"step": 10500
|
| 63219 |
},
|
| 63220 |
{
|
| 63221 |
+
"epoch": 96.33,
|
| 63222 |
+
"step": 10500,
|
| 63223 |
+
"total_flos": 1.2873670788396168e+20,
|
| 63224 |
+
"train_loss": 0.0,
|
| 63225 |
+
"train_runtime": 113.9955,
|
| 63226 |
+
"train_samples_per_second": 152.55,
|
| 63227 |
+
"train_steps_per_second": 4.781
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 63228 |
}
|
| 63229 |
],
|
| 63230 |
+
"max_steps": 545,
|
| 63231 |
+
"num_train_epochs": 5,
|
| 63232 |
+
"total_flos": 1.2873670788396168e+20,
|
| 63233 |
"trial_name": null,
|
| 63234 |
"trial_params": null
|
| 63235 |
}
|