mirror of
https://www.modelscope.cn/shiertier/ComfyUI-segformer_b2_clothes.git
synced 2026-04-02 14:22:54 +08:00
12227 lines
284 KiB
JSON
12227 lines
284 KiB
JSON
{
|
|
"best_metric": null,
|
|
"best_model_checkpoint": null,
|
|
"epoch": 4.993191103041307,
|
|
"global_step": 11000,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.0,
|
|
"learning_rate": 7.992737176577395e-05,
|
|
"loss": 2.2448,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.01,
|
|
"learning_rate": 7.985474353154789e-05,
|
|
"loss": 1.2019,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.01,
|
|
"learning_rate": 7.978211529732184e-05,
|
|
"loss": 0.8748,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.02,
|
|
"learning_rate": 7.970948706309578e-05,
|
|
"loss": 0.6175,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 0.02,
|
|
"learning_rate": 7.963685882886972e-05,
|
|
"loss": 0.5248,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.03,
|
|
"learning_rate": 7.956423059464368e-05,
|
|
"loss": 0.4516,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 0.03,
|
|
"learning_rate": 7.949160236041762e-05,
|
|
"loss": 0.4111,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 0.04,
|
|
"learning_rate": 7.941897412619157e-05,
|
|
"loss": 0.3828,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 0.04,
|
|
"learning_rate": 7.934634589196551e-05,
|
|
"loss": 0.3985,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"learning_rate": 7.927371765773945e-05,
|
|
"loss": 0.3005,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"eval_loss": 0.30581027269363403,
|
|
"eval_mean_accuracy": 0.5292847927678034,
|
|
"eval_mean_iou": 0.41129172532009384,
|
|
"eval_overall_accuracy": 0.9130313316088045,
|
|
"eval_per_category_accuracy": [
|
|
0.9867883273464839,
|
|
0.5434349649443806,
|
|
0.780135770881479,
|
|
0.0,
|
|
0.8113983778550663,
|
|
0.5378490454382858,
|
|
0.7404555846704737,
|
|
0.2147733081885067,
|
|
0.0,
|
|
0.40405372405372403,
|
|
0.25176572429125105,
|
|
0.9192750649381457,
|
|
0.5542416776073233,
|
|
0.7225067314291748,
|
|
0.6504408087601311,
|
|
0.6555691161356629,
|
|
0.754438043280372,
|
|
0.0
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9692330591898085,
|
|
0.41746377262267553,
|
|
0.6905409848042081,
|
|
0.0,
|
|
0.6326664141511412,
|
|
0.3040235250652781,
|
|
0.50160063365291,
|
|
0.19516779272624604,
|
|
0.0,
|
|
0.2624766014150195,
|
|
0.18683524714023958,
|
|
0.7177784569221793,
|
|
0.4419055676444967,
|
|
0.49617640216212294,
|
|
0.5149447942797659,
|
|
0.4906163771139591,
|
|
0.5818214268716386,
|
|
0.0
|
|
],
|
|
"eval_runtime": 80.836,
|
|
"eval_samples_per_second": 1.101,
|
|
"eval_steps_per_second": 0.148,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"learning_rate": 7.92010894235134e-05,
|
|
"loss": 0.3054,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"learning_rate": 7.912846118928734e-05,
|
|
"loss": 0.3008,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 0.06,
|
|
"learning_rate": 7.905583295506128e-05,
|
|
"loss": 0.28,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 0.06,
|
|
"learning_rate": 7.898320472083523e-05,
|
|
"loss": 0.2834,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 0.07,
|
|
"learning_rate": 7.891057648660917e-05,
|
|
"loss": 0.2652,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.07,
|
|
"learning_rate": 7.883794825238313e-05,
|
|
"loss": 0.2602,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 0.08,
|
|
"learning_rate": 7.876532001815707e-05,
|
|
"loss": 0.2308,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 0.08,
|
|
"learning_rate": 7.869269178393101e-05,
|
|
"loss": 0.2486,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 0.09,
|
|
"learning_rate": 7.862006354970496e-05,
|
|
"loss": 0.2408,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 0.09,
|
|
"learning_rate": 7.85474353154789e-05,
|
|
"loss": 0.2458,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.09,
|
|
"eval_loss": 0.23244497179985046,
|
|
"eval_mean_accuracy": 0.5973882303923332,
|
|
"eval_mean_iou": 0.4794550446888198,
|
|
"eval_overall_accuracy": 0.926452679580517,
|
|
"eval_per_category_accuracy": [
|
|
0.9907792090081089,
|
|
0.5952492754092521,
|
|
0.8835458622207549,
|
|
0.3454964731416169,
|
|
0.7917318774601994,
|
|
0.3479833420921308,
|
|
0.7821879338963764,
|
|
0.5380378725887068,
|
|
0.015571650284531816,
|
|
0.5074578754578755,
|
|
0.43118101870979164,
|
|
0.9046111025449427,
|
|
0.6666366922648058,
|
|
0.7831467891522799,
|
|
0.6957018451457148,
|
|
0.7076567317574511,
|
|
0.7660125959274607,
|
|
0.0
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9718178183715939,
|
|
0.4710299054403813,
|
|
0.7456040814786654,
|
|
0.30835351089588375,
|
|
0.6790629628823184,
|
|
0.2858956985209414,
|
|
0.5603084033173548,
|
|
0.39362178720525015,
|
|
0.015510666804081212,
|
|
0.3084005627682499,
|
|
0.29075652548837516,
|
|
0.7718823584975677,
|
|
0.5398866830767438,
|
|
0.5739773379508709,
|
|
0.5774116215829398,
|
|
0.5246370047954892,
|
|
0.6120338753220498,
|
|
0.0
|
|
],
|
|
"eval_runtime": 79.6125,
|
|
"eval_samples_per_second": 1.118,
|
|
"eval_steps_per_second": 0.151,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.1,
|
|
"learning_rate": 7.847480708125284e-05,
|
|
"loss": 0.2293,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 0.1,
|
|
"learning_rate": 7.840217884702679e-05,
|
|
"loss": 0.2194,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 0.1,
|
|
"learning_rate": 7.832955061280073e-05,
|
|
"loss": 0.2144,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 0.11,
|
|
"learning_rate": 7.825692237857467e-05,
|
|
"loss": 0.219,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 0.11,
|
|
"learning_rate": 7.818429414434862e-05,
|
|
"loss": 0.2142,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 0.12,
|
|
"learning_rate": 7.811166591012257e-05,
|
|
"loss": 0.1925,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 0.12,
|
|
"learning_rate": 7.803903767589652e-05,
|
|
"loss": 0.2189,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 0.13,
|
|
"learning_rate": 7.796640944167046e-05,
|
|
"loss": 0.2039,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 0.13,
|
|
"learning_rate": 7.78937812074444e-05,
|
|
"loss": 0.2164,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 0.14,
|
|
"learning_rate": 7.782115297321835e-05,
|
|
"loss": 0.1981,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 0.14,
|
|
"eval_loss": 0.18461237847805023,
|
|
"eval_mean_accuracy": 0.6624681867698782,
|
|
"eval_mean_iou": 0.5471385687290596,
|
|
"eval_overall_accuracy": 0.9407998417200667,
|
|
"eval_per_category_accuracy": [
|
|
0.9879307267213177,
|
|
0.60820285302547,
|
|
0.8681323594145482,
|
|
0.5095632121540966,
|
|
0.8110457078204237,
|
|
0.792409876390551,
|
|
0.8195600051502087,
|
|
0.6689390942443275,
|
|
0.08535954474909467,
|
|
0.4324884004884005,
|
|
0.5378032225697928,
|
|
0.9206627842155828,
|
|
0.7629039800010791,
|
|
0.7839079246079932,
|
|
0.7553457492671151,
|
|
0.712108170606372,
|
|
0.8469511498952397,
|
|
0.02111260053619303
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9761332103456535,
|
|
0.506465313827277,
|
|
0.7501968813986455,
|
|
0.4387153284671533,
|
|
0.7108649071563323,
|
|
0.5753079215555509,
|
|
0.6404749829013227,
|
|
0.5486831870207801,
|
|
0.08104125736738703,
|
|
0.31538757425954156,
|
|
0.3560139873801012,
|
|
0.7866483347577722,
|
|
0.6376416739319966,
|
|
0.665543581794072,
|
|
0.61706426945715,
|
|
0.5785792942920218,
|
|
0.6433084891951437,
|
|
0.020424042015172145
|
|
],
|
|
"eval_runtime": 80.127,
|
|
"eval_samples_per_second": 1.111,
|
|
"eval_steps_per_second": 0.15,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 0.14,
|
|
"learning_rate": 7.774852473899229e-05,
|
|
"loss": 0.1953,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 0.15,
|
|
"learning_rate": 7.767589650476623e-05,
|
|
"loss": 0.216,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 0.15,
|
|
"learning_rate": 7.760326827054018e-05,
|
|
"loss": 0.2082,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 0.15,
|
|
"learning_rate": 7.753064003631412e-05,
|
|
"loss": 0.1968,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 0.16,
|
|
"learning_rate": 7.745801180208806e-05,
|
|
"loss": 0.2073,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 0.16,
|
|
"learning_rate": 7.738538356786201e-05,
|
|
"loss": 0.1935,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 0.17,
|
|
"learning_rate": 7.731275533363596e-05,
|
|
"loss": 0.2165,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 0.17,
|
|
"learning_rate": 7.724012709940991e-05,
|
|
"loss": 0.1938,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 0.18,
|
|
"learning_rate": 7.716749886518385e-05,
|
|
"loss": 0.1906,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 0.18,
|
|
"learning_rate": 7.70948706309578e-05,
|
|
"loss": 0.213,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 0.18,
|
|
"eval_loss": 0.19584733247756958,
|
|
"eval_mean_accuracy": 0.6836090641053443,
|
|
"eval_mean_iou": 0.5613970854128544,
|
|
"eval_overall_accuracy": 0.937813962443491,
|
|
"eval_per_category_accuracy": [
|
|
0.9872100638090652,
|
|
0.6182904515940997,
|
|
0.8384488142401233,
|
|
0.6422951709169832,
|
|
0.7890436261066395,
|
|
0.5009693559854972,
|
|
0.8500088736702545,
|
|
0.8491183956444536,
|
|
0.09529229177444387,
|
|
0.4957167277167277,
|
|
0.40987428436522455,
|
|
0.9088928688625051,
|
|
0.7835323633616891,
|
|
0.7806346021857347,
|
|
0.7904002845318159,
|
|
0.7346993833504625,
|
|
0.8303345233944204,
|
|
0.40020107238605895
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9775908413156985,
|
|
0.5344072327264076,
|
|
0.7521034650296529,
|
|
0.5101820924469346,
|
|
0.7105627872001641,
|
|
0.40911795785441996,
|
|
0.6171770972037284,
|
|
0.5161374621460837,
|
|
0.09092704116892092,
|
|
0.3422119125808344,
|
|
0.3093433312298155,
|
|
0.7932841949798822,
|
|
0.6599897744702607,
|
|
0.6609991431658161,
|
|
0.6341190305315129,
|
|
0.5955388476397755,
|
|
0.6818525050425194,
|
|
0.3096028206989526
|
|
],
|
|
"eval_runtime": 80.4083,
|
|
"eval_samples_per_second": 1.107,
|
|
"eval_steps_per_second": 0.149,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 0.19,
|
|
"learning_rate": 7.702224239673174e-05,
|
|
"loss": 0.2075,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 0.19,
|
|
"learning_rate": 7.694961416250568e-05,
|
|
"loss": 0.184,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"learning_rate": 7.687698592827962e-05,
|
|
"loss": 0.2022,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"learning_rate": 7.680435769405357e-05,
|
|
"loss": 0.2174,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"learning_rate": 7.673172945982751e-05,
|
|
"loss": 0.1752,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 0.21,
|
|
"learning_rate": 7.665910122560145e-05,
|
|
"loss": 0.1899,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 0.21,
|
|
"learning_rate": 7.658647299137541e-05,
|
|
"loss": 0.172,
|
|
"step": 470
|
|
},
|
|
{
|
|
"epoch": 0.22,
|
|
"learning_rate": 7.651384475714935e-05,
|
|
"loss": 0.1834,
|
|
"step": 480
|
|
},
|
|
{
|
|
"epoch": 0.22,
|
|
"learning_rate": 7.64412165229233e-05,
|
|
"loss": 0.1858,
|
|
"step": 490
|
|
},
|
|
{
|
|
"epoch": 0.23,
|
|
"learning_rate": 7.636858828869724e-05,
|
|
"loss": 0.2084,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 0.23,
|
|
"eval_loss": 0.1708747297525406,
|
|
"eval_mean_accuracy": 0.6622155280724535,
|
|
"eval_mean_iou": 0.5659747611486945,
|
|
"eval_overall_accuracy": 0.9442667586080143,
|
|
"eval_per_category_accuracy": [
|
|
0.9866500967447802,
|
|
0.5469648148448,
|
|
0.8757806481787168,
|
|
0.4290558871405317,
|
|
0.9149488868594733,
|
|
0.7030833911119551,
|
|
0.8079024383453911,
|
|
0.6823532812011712,
|
|
0.04759441282979824,
|
|
0.4875702075702076,
|
|
0.5283204469889237,
|
|
0.8952884756042114,
|
|
0.7860142438357642,
|
|
0.8228842194181933,
|
|
0.7062640110363856,
|
|
0.7089606885919836,
|
|
0.8114900037954411,
|
|
0.17875335120643432
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9778258858246885,
|
|
0.5132802528414271,
|
|
0.7644472861480458,
|
|
0.391654284299158,
|
|
0.7185696664929735,
|
|
0.5427059364933272,
|
|
0.6772939108618439,
|
|
0.5690865593212168,
|
|
0.0467979042677654,
|
|
0.3500497917163415,
|
|
0.38005066587412784,
|
|
0.7938747177854749,
|
|
0.6864153225173156,
|
|
0.7103305267235585,
|
|
0.6054474564369792,
|
|
0.5835536333643516,
|
|
0.7060943062886968,
|
|
0.17006759341920674
|
|
],
|
|
"eval_runtime": 82.1375,
|
|
"eval_samples_per_second": 1.084,
|
|
"eval_steps_per_second": 0.146,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 0.23,
|
|
"learning_rate": 7.629596005447118e-05,
|
|
"loss": 0.189,
|
|
"step": 510
|
|
},
|
|
{
|
|
"epoch": 0.24,
|
|
"learning_rate": 7.622333182024513e-05,
|
|
"loss": 0.1882,
|
|
"step": 520
|
|
},
|
|
{
|
|
"epoch": 0.24,
|
|
"learning_rate": 7.615070358601907e-05,
|
|
"loss": 0.1683,
|
|
"step": 530
|
|
},
|
|
{
|
|
"epoch": 0.25,
|
|
"learning_rate": 7.607807535179301e-05,
|
|
"loss": 0.2009,
|
|
"step": 540
|
|
},
|
|
{
|
|
"epoch": 0.25,
|
|
"learning_rate": 7.600544711756696e-05,
|
|
"loss": 0.1955,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 0.25,
|
|
"learning_rate": 7.59328188833409e-05,
|
|
"loss": 0.159,
|
|
"step": 560
|
|
},
|
|
{
|
|
"epoch": 0.26,
|
|
"learning_rate": 7.586019064911484e-05,
|
|
"loss": 0.1591,
|
|
"step": 570
|
|
},
|
|
{
|
|
"epoch": 0.26,
|
|
"learning_rate": 7.57875624148888e-05,
|
|
"loss": 0.1813,
|
|
"step": 580
|
|
},
|
|
{
|
|
"epoch": 0.27,
|
|
"learning_rate": 7.571493418066274e-05,
|
|
"loss": 0.1742,
|
|
"step": 590
|
|
},
|
|
{
|
|
"epoch": 0.27,
|
|
"learning_rate": 7.564230594643669e-05,
|
|
"loss": 0.1768,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 0.27,
|
|
"eval_loss": 0.16358888149261475,
|
|
"eval_mean_accuracy": 0.7168573384583677,
|
|
"eval_mean_iou": 0.5947241030305628,
|
|
"eval_overall_accuracy": 0.9456647380014483,
|
|
"eval_per_category_accuracy": [
|
|
0.9883615047521485,
|
|
0.6133680920999368,
|
|
0.8897205485858919,
|
|
0.606348345089528,
|
|
0.8547445440779222,
|
|
0.6468916573558442,
|
|
0.8031315356321359,
|
|
0.7755932607529358,
|
|
0.18939472322814277,
|
|
0.6884297924297924,
|
|
0.44591670109583115,
|
|
0.912668572309663,
|
|
0.8122943006672302,
|
|
0.8121667282614434,
|
|
0.7158238489394724,
|
|
0.739613309352518,
|
|
0.8623828581831872,
|
|
0.5465817694369973
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9789898289726563,
|
|
0.5284958912900931,
|
|
0.7700365647308839,
|
|
0.5212220149253731,
|
|
0.7458776898298387,
|
|
0.5271602805638629,
|
|
0.6717823107866594,
|
|
0.5744670602921285,
|
|
0.1729007273070747,
|
|
0.40336528831020174,
|
|
0.34301128968794453,
|
|
0.8029663743656279,
|
|
0.7079338085180584,
|
|
0.7184545748212611,
|
|
0.6150804308165325,
|
|
0.5833481100196065,
|
|
0.6760090852137903,
|
|
0.3639325240985362
|
|
],
|
|
"eval_runtime": 81.4207,
|
|
"eval_samples_per_second": 1.093,
|
|
"eval_steps_per_second": 0.147,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 0.28,
|
|
"learning_rate": 7.556967771221063e-05,
|
|
"loss": 0.192,
|
|
"step": 610
|
|
},
|
|
{
|
|
"epoch": 0.28,
|
|
"learning_rate": 7.549704947798457e-05,
|
|
"loss": 0.1611,
|
|
"step": 620
|
|
},
|
|
{
|
|
"epoch": 0.29,
|
|
"learning_rate": 7.542442124375852e-05,
|
|
"loss": 0.1801,
|
|
"step": 630
|
|
},
|
|
{
|
|
"epoch": 0.29,
|
|
"learning_rate": 7.535179300953246e-05,
|
|
"loss": 0.17,
|
|
"step": 640
|
|
},
|
|
{
|
|
"epoch": 0.3,
|
|
"learning_rate": 7.52791647753064e-05,
|
|
"loss": 0.1827,
|
|
"step": 650
|
|
},
|
|
{
|
|
"epoch": 0.3,
|
|
"learning_rate": 7.520653654108035e-05,
|
|
"loss": 0.191,
|
|
"step": 660
|
|
},
|
|
{
|
|
"epoch": 0.3,
|
|
"learning_rate": 7.513390830685429e-05,
|
|
"loss": 0.1678,
|
|
"step": 670
|
|
},
|
|
{
|
|
"epoch": 0.31,
|
|
"learning_rate": 7.506128007262825e-05,
|
|
"loss": 0.2309,
|
|
"step": 680
|
|
},
|
|
{
|
|
"epoch": 0.31,
|
|
"learning_rate": 7.498865183840219e-05,
|
|
"loss": 0.1681,
|
|
"step": 690
|
|
},
|
|
{
|
|
"epoch": 0.32,
|
|
"learning_rate": 7.491602360417613e-05,
|
|
"loss": 0.1672,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 0.32,
|
|
"eval_loss": 0.1567693054676056,
|
|
"eval_mean_accuracy": 0.7141943900361636,
|
|
"eval_mean_iou": 0.596780275848511,
|
|
"eval_overall_accuracy": 0.9485598360554556,
|
|
"eval_per_category_accuracy": [
|
|
0.9900870575442013,
|
|
0.6031185738111045,
|
|
0.8562610047320348,
|
|
0.5413049376017363,
|
|
0.8634782733988677,
|
|
0.7461214720079138,
|
|
0.822484836463477,
|
|
0.7418608978075951,
|
|
0.21919296430419039,
|
|
0.47965811965811966,
|
|
0.5514470085974542,
|
|
0.927099271348539,
|
|
0.7968634785892847,
|
|
0.868974710944512,
|
|
0.7993457923779962,
|
|
0.7131680369989722,
|
|
0.8430754799073048,
|
|
0.49195710455764075
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9792376762603316,
|
|
0.5245458386142797,
|
|
0.7704556601277661,
|
|
0.48261474269819193,
|
|
0.7415891427416451,
|
|
0.6085604990922975,
|
|
0.6909641574493773,
|
|
0.6132853682648054,
|
|
0.19095047095407636,
|
|
0.3546614086785694,
|
|
0.40903154254193635,
|
|
0.7993550454399804,
|
|
0.6968145754208833,
|
|
0.7112413080349587,
|
|
0.6509687753713822,
|
|
0.6050298356992998,
|
|
0.6898486321344258,
|
|
0.2228902857489903
|
|
],
|
|
"eval_runtime": 86.481,
|
|
"eval_samples_per_second": 1.029,
|
|
"eval_steps_per_second": 0.139,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 0.32,
|
|
"learning_rate": 7.484339536995008e-05,
|
|
"loss": 0.194,
|
|
"step": 710
|
|
},
|
|
{
|
|
"epoch": 0.33,
|
|
"learning_rate": 7.477076713572402e-05,
|
|
"loss": 0.1855,
|
|
"step": 720
|
|
},
|
|
{
|
|
"epoch": 0.33,
|
|
"learning_rate": 7.469813890149796e-05,
|
|
"loss": 0.1533,
|
|
"step": 730
|
|
},
|
|
{
|
|
"epoch": 0.34,
|
|
"learning_rate": 7.462551066727191e-05,
|
|
"loss": 0.1814,
|
|
"step": 740
|
|
},
|
|
{
|
|
"epoch": 0.34,
|
|
"learning_rate": 7.455288243304585e-05,
|
|
"loss": 0.1775,
|
|
"step": 750
|
|
},
|
|
{
|
|
"epoch": 0.34,
|
|
"learning_rate": 7.44802541988198e-05,
|
|
"loss": 0.1746,
|
|
"step": 760
|
|
},
|
|
{
|
|
"epoch": 0.35,
|
|
"learning_rate": 7.440762596459374e-05,
|
|
"loss": 0.1722,
|
|
"step": 770
|
|
},
|
|
{
|
|
"epoch": 0.35,
|
|
"learning_rate": 7.43349977303677e-05,
|
|
"loss": 0.1586,
|
|
"step": 780
|
|
},
|
|
{
|
|
"epoch": 0.36,
|
|
"learning_rate": 7.426236949614164e-05,
|
|
"loss": 0.1738,
|
|
"step": 790
|
|
},
|
|
{
|
|
"epoch": 0.36,
|
|
"learning_rate": 7.418974126191558e-05,
|
|
"loss": 0.1745,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 0.36,
|
|
"eval_loss": 0.15606097877025604,
|
|
"eval_mean_accuracy": 0.7226644915074127,
|
|
"eval_mean_iou": 0.6052765335352305,
|
|
"eval_overall_accuracy": 0.9491321263688334,
|
|
"eval_per_category_accuracy": [
|
|
0.9890046209624822,
|
|
0.6175132369371266,
|
|
0.8928362633432376,
|
|
0.6507053716766142,
|
|
0.8532096746859076,
|
|
0.5752365979358679,
|
|
0.8398441713905911,
|
|
0.8413509563885382,
|
|
0.23833419555095706,
|
|
0.6422954822954823,
|
|
0.5084695744555273,
|
|
0.9224774940399236,
|
|
0.8665719475567865,
|
|
0.8431225032117277,
|
|
0.8245656578720469,
|
|
0.7727517985611511,
|
|
0.8572852412882283,
|
|
0.27238605898123325
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9807662339083965,
|
|
0.5198184445110814,
|
|
0.7770744008201076,
|
|
0.554374205477869,
|
|
0.7651092891972128,
|
|
0.5236022173180742,
|
|
0.7019472398347972,
|
|
0.5826306354990312,
|
|
0.2023276240667545,
|
|
0.4164344521849272,
|
|
0.37874236140623396,
|
|
0.8109587612741775,
|
|
0.710578901186412,
|
|
0.707150659050318,
|
|
0.6494664640616643,
|
|
0.6240312892282475,
|
|
0.7346238867852686,
|
|
0.25534053782357374
|
|
],
|
|
"eval_runtime": 83.6193,
|
|
"eval_samples_per_second": 1.064,
|
|
"eval_steps_per_second": 0.144,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 0.37,
|
|
"learning_rate": 7.411711302768953e-05,
|
|
"loss": 0.1652,
|
|
"step": 810
|
|
},
|
|
{
|
|
"epoch": 0.37,
|
|
"learning_rate": 7.404448479346347e-05,
|
|
"loss": 0.1851,
|
|
"step": 820
|
|
},
|
|
{
|
|
"epoch": 0.38,
|
|
"learning_rate": 7.397185655923741e-05,
|
|
"loss": 0.1703,
|
|
"step": 830
|
|
},
|
|
{
|
|
"epoch": 0.38,
|
|
"learning_rate": 7.389922832501136e-05,
|
|
"loss": 0.1454,
|
|
"step": 840
|
|
},
|
|
{
|
|
"epoch": 0.39,
|
|
"learning_rate": 7.38266000907853e-05,
|
|
"loss": 0.1894,
|
|
"step": 850
|
|
},
|
|
{
|
|
"epoch": 0.39,
|
|
"learning_rate": 7.375397185655924e-05,
|
|
"loss": 0.1734,
|
|
"step": 860
|
|
},
|
|
{
|
|
"epoch": 0.39,
|
|
"learning_rate": 7.368134362233319e-05,
|
|
"loss": 0.1696,
|
|
"step": 870
|
|
},
|
|
{
|
|
"epoch": 0.4,
|
|
"learning_rate": 7.360871538810713e-05,
|
|
"loss": 0.1804,
|
|
"step": 880
|
|
},
|
|
{
|
|
"epoch": 0.4,
|
|
"learning_rate": 7.353608715388109e-05,
|
|
"loss": 0.1661,
|
|
"step": 890
|
|
},
|
|
{
|
|
"epoch": 0.41,
|
|
"learning_rate": 7.346345891965503e-05,
|
|
"loss": 0.1339,
|
|
"step": 900
|
|
},
|
|
{
|
|
"epoch": 0.41,
|
|
"eval_loss": 0.15481343865394592,
|
|
"eval_mean_accuracy": 0.719634431340832,
|
|
"eval_mean_iou": 0.6087190770313099,
|
|
"eval_overall_accuracy": 0.9506247874056355,
|
|
"eval_per_category_accuracy": [
|
|
0.9919838454355556,
|
|
0.5911527064881232,
|
|
0.8761297045229449,
|
|
0.668407487791644,
|
|
0.8954285318291567,
|
|
0.6038513683596907,
|
|
0.8717180469573751,
|
|
0.737148049198599,
|
|
0.25064666321779616,
|
|
0.3186715506715507,
|
|
0.7722658324971965,
|
|
0.9177924588725077,
|
|
0.7597341870042983,
|
|
0.823957728384633,
|
|
0.7929276599413692,
|
|
0.7448484069886948,
|
|
0.8559244367781333,
|
|
0.48083109919571043
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9811278355168047,
|
|
0.5256497012454107,
|
|
0.77428604860021,
|
|
0.5607396870554765,
|
|
0.7664850267467621,
|
|
0.5611307522001325,
|
|
0.7133819452453877,
|
|
0.5870733942446437,
|
|
0.19933349790175264,
|
|
0.28642920482181583,
|
|
0.44613601104727485,
|
|
0.8111479557072823,
|
|
0.6859910442065434,
|
|
0.7038314824210309,
|
|
0.6581561679496527,
|
|
0.6242086904095431,
|
|
0.7302158936248069,
|
|
0.3416190476190476
|
|
],
|
|
"eval_runtime": 76.1804,
|
|
"eval_samples_per_second": 1.168,
|
|
"eval_steps_per_second": 0.158,
|
|
"step": 900
|
|
},
|
|
{
|
|
"epoch": 0.41,
|
|
"learning_rate": 7.339083068542897e-05,
|
|
"loss": 0.1726,
|
|
"step": 910
|
|
},
|
|
{
|
|
"epoch": 0.42,
|
|
"learning_rate": 7.331820245120292e-05,
|
|
"loss": 0.1685,
|
|
"step": 920
|
|
},
|
|
{
|
|
"epoch": 0.42,
|
|
"learning_rate": 7.324557421697686e-05,
|
|
"loss": 0.1719,
|
|
"step": 930
|
|
},
|
|
{
|
|
"epoch": 0.43,
|
|
"learning_rate": 7.31729459827508e-05,
|
|
"loss": 0.1705,
|
|
"step": 940
|
|
},
|
|
{
|
|
"epoch": 0.43,
|
|
"learning_rate": 7.310031774852475e-05,
|
|
"loss": 0.1648,
|
|
"step": 950
|
|
},
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 7.302768951429869e-05,
|
|
"loss": 0.1671,
|
|
"step": 960
|
|
},
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 7.295506128007263e-05,
|
|
"loss": 0.1721,
|
|
"step": 970
|
|
},
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 7.288243304584658e-05,
|
|
"loss": 0.1781,
|
|
"step": 980
|
|
},
|
|
{
|
|
"epoch": 0.45,
|
|
"learning_rate": 7.280980481162053e-05,
|
|
"loss": 0.157,
|
|
"step": 990
|
|
},
|
|
{
|
|
"epoch": 0.45,
|
|
"learning_rate": 7.273717657739448e-05,
|
|
"loss": 0.1525,
|
|
"step": 1000
|
|
},
|
|
{
|
|
"epoch": 0.45,
|
|
"eval_loss": 0.1561342477798462,
|
|
"eval_mean_accuracy": 0.709750323770174,
|
|
"eval_mean_iou": 0.6044507479852305,
|
|
"eval_overall_accuracy": 0.9485302614362052,
|
|
"eval_per_category_accuracy": [
|
|
0.9892216740702133,
|
|
0.615634968182775,
|
|
0.8966913585341697,
|
|
0.5729788388497016,
|
|
0.9087696138400346,
|
|
0.5404590198684855,
|
|
0.9021390765119168,
|
|
0.643297512793695,
|
|
0.20713916192446974,
|
|
0.5912087912087912,
|
|
0.6094158846327884,
|
|
0.9404664476363306,
|
|
0.8413754653525889,
|
|
0.8410942752054625,
|
|
0.7898613985169857,
|
|
0.7273830935251798,
|
|
0.8567606681210714,
|
|
0.30160857908847183
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9809378800002442,
|
|
0.5447368798085878,
|
|
0.7765388584370854,
|
|
0.5160659743433109,
|
|
0.7594886697914242,
|
|
0.49892168893273703,
|
|
0.6637198728091062,
|
|
0.5417948413676431,
|
|
0.17752948479205463,
|
|
0.4290757768026145,
|
|
0.45001343822412054,
|
|
0.7856182729113205,
|
|
0.742127889084535,
|
|
0.749399262258774,
|
|
0.6515890409740961,
|
|
0.6199102206164121,
|
|
0.7204954276995257,
|
|
0.2721499848805564
|
|
],
|
|
"eval_runtime": 76.0709,
|
|
"eval_samples_per_second": 1.17,
|
|
"eval_steps_per_second": 0.158,
|
|
"step": 1000
|
|
},
|
|
{
|
|
"epoch": 0.46,
|
|
"learning_rate": 7.266454834316842e-05,
|
|
"loss": 0.1442,
|
|
"step": 1010
|
|
},
|
|
{
|
|
"epoch": 0.46,
|
|
"learning_rate": 7.259192010894236e-05,
|
|
"loss": 0.17,
|
|
"step": 1020
|
|
},
|
|
{
|
|
"epoch": 0.47,
|
|
"learning_rate": 7.25192918747163e-05,
|
|
"loss": 0.1451,
|
|
"step": 1030
|
|
},
|
|
{
|
|
"epoch": 0.47,
|
|
"learning_rate": 7.244666364049025e-05,
|
|
"loss": 0.1649,
|
|
"step": 1040
|
|
},
|
|
{
|
|
"epoch": 0.48,
|
|
"learning_rate": 7.237403540626419e-05,
|
|
"loss": 0.1765,
|
|
"step": 1050
|
|
},
|
|
{
|
|
"epoch": 0.48,
|
|
"learning_rate": 7.230140717203814e-05,
|
|
"loss": 0.1611,
|
|
"step": 1060
|
|
},
|
|
{
|
|
"epoch": 0.49,
|
|
"learning_rate": 7.222877893781208e-05,
|
|
"loss": 0.1544,
|
|
"step": 1070
|
|
},
|
|
{
|
|
"epoch": 0.49,
|
|
"learning_rate": 7.215615070358602e-05,
|
|
"loss": 0.1671,
|
|
"step": 1080
|
|
},
|
|
{
|
|
"epoch": 0.49,
|
|
"learning_rate": 7.208352246935997e-05,
|
|
"loss": 0.1831,
|
|
"step": 1090
|
|
},
|
|
{
|
|
"epoch": 0.5,
|
|
"learning_rate": 7.201089423513392e-05,
|
|
"loss": 0.1444,
|
|
"step": 1100
|
|
},
|
|
{
|
|
"epoch": 0.5,
|
|
"eval_loss": 0.1541578769683838,
|
|
"eval_mean_accuracy": 0.7261158297983399,
|
|
"eval_mean_iou": 0.6067205446661403,
|
|
"eval_overall_accuracy": 0.9484547818644663,
|
|
"eval_per_category_accuracy": [
|
|
0.9907627899639259,
|
|
0.6337861688175003,
|
|
0.8650957411687026,
|
|
0.5448996201844818,
|
|
0.8902592399206012,
|
|
0.6789908098869895,
|
|
0.7145583870103387,
|
|
0.7030056377157948,
|
|
0.2124159337816865,
|
|
0.5363516483516484,
|
|
0.5988805603100592,
|
|
0.9369991262508253,
|
|
0.8605920543855547,
|
|
0.8890370096615807,
|
|
0.8072943610967408,
|
|
0.7692381808838643,
|
|
0.8745128412425596,
|
|
0.5634048257372655
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.981575441938742,
|
|
0.5338588906013448,
|
|
0.775538517031844,
|
|
0.4985417313062364,
|
|
0.7595387897122733,
|
|
0.5114676877169579,
|
|
0.6549842187273229,
|
|
0.5485690202468412,
|
|
0.1797644586489208,
|
|
0.4145013965426134,
|
|
0.44945222064729506,
|
|
0.8091983064736411,
|
|
0.7258117726467789,
|
|
0.7419751635810855,
|
|
0.6691561400060748,
|
|
0.6295969170754275,
|
|
0.7225487021678552,
|
|
0.31489042891927327
|
|
],
|
|
"eval_runtime": 72.9643,
|
|
"eval_samples_per_second": 1.22,
|
|
"eval_steps_per_second": 0.164,
|
|
"step": 1100
|
|
},
|
|
{
|
|
"epoch": 0.5,
|
|
"learning_rate": 7.193826600090787e-05,
|
|
"loss": 0.144,
|
|
"step": 1110
|
|
},
|
|
{
|
|
"epoch": 0.51,
|
|
"learning_rate": 7.186563776668181e-05,
|
|
"loss": 0.1601,
|
|
"step": 1120
|
|
},
|
|
{
|
|
"epoch": 0.51,
|
|
"learning_rate": 7.179300953245575e-05,
|
|
"loss": 0.1434,
|
|
"step": 1130
|
|
},
|
|
{
|
|
"epoch": 0.52,
|
|
"learning_rate": 7.17203812982297e-05,
|
|
"loss": 0.1475,
|
|
"step": 1140
|
|
},
|
|
{
|
|
"epoch": 0.52,
|
|
"learning_rate": 7.164775306400364e-05,
|
|
"loss": 0.1512,
|
|
"step": 1150
|
|
},
|
|
{
|
|
"epoch": 0.53,
|
|
"learning_rate": 7.157512482977758e-05,
|
|
"loss": 0.1462,
|
|
"step": 1160
|
|
},
|
|
{
|
|
"epoch": 0.53,
|
|
"learning_rate": 7.150249659555153e-05,
|
|
"loss": 0.1585,
|
|
"step": 1170
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"learning_rate": 7.142986836132547e-05,
|
|
"loss": 0.1654,
|
|
"step": 1180
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"learning_rate": 7.135724012709941e-05,
|
|
"loss": 0.1471,
|
|
"step": 1190
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"learning_rate": 7.128461189287337e-05,
|
|
"loss": 0.1752,
|
|
"step": 1200
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"eval_loss": 0.15197259187698364,
|
|
"eval_mean_accuracy": 0.7405501900221605,
|
|
"eval_mean_iou": 0.6187261085826632,
|
|
"eval_overall_accuracy": 0.9496975587994865,
|
|
"eval_per_category_accuracy": [
|
|
0.9902535777152744,
|
|
0.6046406191810101,
|
|
0.8918200451194014,
|
|
0.7129001627780792,
|
|
0.8910963166565234,
|
|
0.6483997897888615,
|
|
0.88835182884604,
|
|
0.6013659854803077,
|
|
0.3084324883600621,
|
|
0.6496898656898656,
|
|
0.5727439060378917,
|
|
0.9107629293132964,
|
|
0.8787745265543226,
|
|
0.8375877725568872,
|
|
0.7801830056906364,
|
|
0.7421184480986639,
|
|
0.8873371123172865,
|
|
0.5334450402144773
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9817709454707259,
|
|
0.5506045414332056,
|
|
0.7855775347611402,
|
|
0.5654115115653577,
|
|
0.7486948450993278,
|
|
0.559655111294076,
|
|
0.716737769654396,
|
|
0.4881858588110045,
|
|
0.24148406172789502,
|
|
0.4524287298229384,
|
|
0.4322109388918582,
|
|
0.817335769660629,
|
|
0.7500805919285265,
|
|
0.7571950171821306,
|
|
0.6612331582553095,
|
|
0.6318837884696372,
|
|
0.7427932302859976,
|
|
0.2537865501737827
|
|
],
|
|
"eval_runtime": 75.6612,
|
|
"eval_samples_per_second": 1.176,
|
|
"eval_steps_per_second": 0.159,
|
|
"step": 1200
|
|
},
|
|
{
|
|
"epoch": 0.55,
|
|
"learning_rate": 7.121198365864731e-05,
|
|
"loss": 0.1617,
|
|
"step": 1210
|
|
},
|
|
{
|
|
"epoch": 0.55,
|
|
"learning_rate": 7.113935542442126e-05,
|
|
"loss": 0.1982,
|
|
"step": 1220
|
|
},
|
|
{
|
|
"epoch": 0.56,
|
|
"learning_rate": 7.10667271901952e-05,
|
|
"loss": 0.1574,
|
|
"step": 1230
|
|
},
|
|
{
|
|
"epoch": 0.56,
|
|
"learning_rate": 7.099409895596914e-05,
|
|
"loss": 0.1615,
|
|
"step": 1240
|
|
},
|
|
{
|
|
"epoch": 0.57,
|
|
"learning_rate": 7.092147072174309e-05,
|
|
"loss": 0.1692,
|
|
"step": 1250
|
|
},
|
|
{
|
|
"epoch": 0.57,
|
|
"learning_rate": 7.084884248751703e-05,
|
|
"loss": 0.1685,
|
|
"step": 1260
|
|
},
|
|
{
|
|
"epoch": 0.58,
|
|
"learning_rate": 7.077621425329097e-05,
|
|
"loss": 0.1412,
|
|
"step": 1270
|
|
},
|
|
{
|
|
"epoch": 0.58,
|
|
"learning_rate": 7.070358601906492e-05,
|
|
"loss": 0.144,
|
|
"step": 1280
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"learning_rate": 7.063095778483886e-05,
|
|
"loss": 0.1503,
|
|
"step": 1290
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"learning_rate": 7.055832955061282e-05,
|
|
"loss": 0.1718,
|
|
"step": 1300
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"eval_loss": 0.15203571319580078,
|
|
"eval_mean_accuracy": 0.7414586755906867,
|
|
"eval_mean_iou": 0.6240856832008199,
|
|
"eval_overall_accuracy": 0.9509381497843882,
|
|
"eval_per_category_accuracy": [
|
|
0.9905460696746577,
|
|
0.6256416068913033,
|
|
0.9004415648728954,
|
|
0.6867878459034183,
|
|
0.8565435729316829,
|
|
0.6157419879084434,
|
|
0.8422365824885947,
|
|
0.819614771140531,
|
|
0.2132436627004656,
|
|
0.759931623931624,
|
|
0.5529815656416612,
|
|
0.9297837767313873,
|
|
0.7906048234807475,
|
|
0.8327965788501135,
|
|
0.7729942662528022,
|
|
0.7829522096608428,
|
|
0.8637621770403582,
|
|
0.5096514745308312
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9817290567284135,
|
|
0.5610588371957949,
|
|
0.7816651963835514,
|
|
0.5702861004730795,
|
|
0.765167781080122,
|
|
0.5597764555681353,
|
|
0.7079144644669532,
|
|
0.5854913989220103,
|
|
0.1928872250818905,
|
|
0.43874104026031885,
|
|
0.38286033413017867,
|
|
0.8100921792328042,
|
|
0.7233445636032301,
|
|
0.7350904649654567,
|
|
0.6550626552681572,
|
|
0.6321275768183586,
|
|
0.7493401292436516,
|
|
0.4009068381926504
|
|
],
|
|
"eval_runtime": 73.6261,
|
|
"eval_samples_per_second": 1.209,
|
|
"eval_steps_per_second": 0.163,
|
|
"step": 1300
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"learning_rate": 7.048570131638676e-05,
|
|
"loss": 0.1693,
|
|
"step": 1310
|
|
},
|
|
{
|
|
"epoch": 0.6,
|
|
"learning_rate": 7.04130730821607e-05,
|
|
"loss": 0.1236,
|
|
"step": 1320
|
|
},
|
|
{
|
|
"epoch": 0.6,
|
|
"learning_rate": 7.034044484793465e-05,
|
|
"loss": 0.1621,
|
|
"step": 1330
|
|
},
|
|
{
|
|
"epoch": 0.61,
|
|
"learning_rate": 7.026781661370859e-05,
|
|
"loss": 0.1611,
|
|
"step": 1340
|
|
},
|
|
{
|
|
"epoch": 0.61,
|
|
"learning_rate": 7.019518837948253e-05,
|
|
"loss": 0.1957,
|
|
"step": 1350
|
|
},
|
|
{
|
|
"epoch": 0.62,
|
|
"learning_rate": 7.012256014525648e-05,
|
|
"loss": 0.168,
|
|
"step": 1360
|
|
},
|
|
{
|
|
"epoch": 0.62,
|
|
"learning_rate": 7.004993191103042e-05,
|
|
"loss": 0.1436,
|
|
"step": 1370
|
|
},
|
|
{
|
|
"epoch": 0.63,
|
|
"learning_rate": 6.997730367680436e-05,
|
|
"loss": 0.1626,
|
|
"step": 1380
|
|
},
|
|
{
|
|
"epoch": 0.63,
|
|
"learning_rate": 6.99046754425783e-05,
|
|
"loss": 0.1659,
|
|
"step": 1390
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"learning_rate": 6.983204720835225e-05,
|
|
"loss": 0.1539,
|
|
"step": 1400
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"eval_loss": 0.14264430105686188,
|
|
"eval_mean_accuracy": 0.7458751803493104,
|
|
"eval_mean_iou": 0.6292022953035405,
|
|
"eval_overall_accuracy": 0.9526165308577291,
|
|
"eval_per_category_accuracy": [
|
|
0.9905214411083831,
|
|
0.6409592124224809,
|
|
0.8682527236711786,
|
|
0.665491047205643,
|
|
0.9027220774872329,
|
|
0.6931469402355601,
|
|
0.900312144400714,
|
|
0.6993621667904037,
|
|
0.22519399896533884,
|
|
0.5980268620268621,
|
|
0.7019024572586515,
|
|
0.9121901847524838,
|
|
0.8336016042299876,
|
|
0.8247628601094628,
|
|
0.7937198223831695,
|
|
0.7335495889003083,
|
|
0.8290354333745792,
|
|
0.6130026809651474
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9813920975948569,
|
|
0.5516996278797508,
|
|
0.7827110010897079,
|
|
0.5675941458899751,
|
|
0.7729552750191524,
|
|
0.597299971458472,
|
|
0.7236993928611889,
|
|
0.5609767053003881,
|
|
0.20242745535714285,
|
|
0.4380612339813536,
|
|
0.48269890341827726,
|
|
0.8207891170014835,
|
|
0.7432858798013126,
|
|
0.7517122463710001,
|
|
0.6698090469628962,
|
|
0.6459694434545526,
|
|
0.7570778443113773,
|
|
0.2754819277108434
|
|
],
|
|
"eval_runtime": 74.8707,
|
|
"eval_samples_per_second": 1.189,
|
|
"eval_steps_per_second": 0.16,
|
|
"step": 1400
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"learning_rate": 6.97594189741262e-05,
|
|
"loss": 0.1324,
|
|
"step": 1410
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"learning_rate": 6.968679073990015e-05,
|
|
"loss": 0.1607,
|
|
"step": 1420
|
|
},
|
|
{
|
|
"epoch": 0.65,
|
|
"learning_rate": 6.96141625056741e-05,
|
|
"loss": 0.1534,
|
|
"step": 1430
|
|
},
|
|
{
|
|
"epoch": 0.65,
|
|
"learning_rate": 6.954153427144804e-05,
|
|
"loss": 0.1372,
|
|
"step": 1440
|
|
},
|
|
{
|
|
"epoch": 0.66,
|
|
"learning_rate": 6.946890603722198e-05,
|
|
"loss": 0.1451,
|
|
"step": 1450
|
|
},
|
|
{
|
|
"epoch": 0.66,
|
|
"learning_rate": 6.939627780299592e-05,
|
|
"loss": 0.1243,
|
|
"step": 1460
|
|
},
|
|
{
|
|
"epoch": 0.67,
|
|
"learning_rate": 6.932364956876987e-05,
|
|
"loss": 0.1668,
|
|
"step": 1470
|
|
},
|
|
{
|
|
"epoch": 0.67,
|
|
"learning_rate": 6.925102133454381e-05,
|
|
"loss": 0.158,
|
|
"step": 1480
|
|
},
|
|
{
|
|
"epoch": 0.68,
|
|
"learning_rate": 6.917839310031775e-05,
|
|
"loss": 0.1576,
|
|
"step": 1490
|
|
},
|
|
{
|
|
"epoch": 0.68,
|
|
"learning_rate": 6.91057648660917e-05,
|
|
"loss": 0.1409,
|
|
"step": 1500
|
|
},
|
|
{
|
|
"epoch": 0.68,
|
|
"eval_loss": 0.1411726325750351,
|
|
"eval_mean_accuracy": 0.743708398966615,
|
|
"eval_mean_iou": 0.6416885840188883,
|
|
"eval_overall_accuracy": 0.9534817384869865,
|
|
"eval_per_category_accuracy": [
|
|
0.9921596290538537,
|
|
0.6701695299470523,
|
|
0.8661704220314735,
|
|
0.6604720564297342,
|
|
0.8799426122145962,
|
|
0.760381383065638,
|
|
0.8176756551726538,
|
|
0.7018658111645197,
|
|
0.2421107087428867,
|
|
0.5839120879120879,
|
|
0.5768852426764248,
|
|
0.9228965773544773,
|
|
0.8362993003974606,
|
|
0.8946421343470073,
|
|
0.8340877737540955,
|
|
0.7522160842754368,
|
|
0.9064593471224076,
|
|
0.4884048257372654
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9811417721668967,
|
|
0.5928042509918503,
|
|
0.7835484031532512,
|
|
0.5689413414349147,
|
|
0.7742403844412381,
|
|
0.6280709561867798,
|
|
0.7250641832679158,
|
|
0.5495596286935109,
|
|
0.20547043069763357,
|
|
0.46297902661172896,
|
|
0.47025475306513564,
|
|
0.8225947408853524,
|
|
0.746202851572216,
|
|
0.7547453437358493,
|
|
0.6814659575217499,
|
|
0.6435401439797769,
|
|
0.7531297189341907,
|
|
0.406640625
|
|
],
|
|
"eval_runtime": 75.8092,
|
|
"eval_samples_per_second": 1.174,
|
|
"eval_steps_per_second": 0.158,
|
|
"step": 1500
|
|
},
|
|
{
|
|
"epoch": 0.69,
|
|
"learning_rate": 6.903313663186565e-05,
|
|
"loss": 0.149,
|
|
"step": 1510
|
|
},
|
|
{
|
|
"epoch": 0.69,
|
|
"learning_rate": 6.89605083976396e-05,
|
|
"loss": 0.1326,
|
|
"step": 1520
|
|
},
|
|
{
|
|
"epoch": 0.69,
|
|
"learning_rate": 6.888788016341354e-05,
|
|
"loss": 0.1549,
|
|
"step": 1530
|
|
},
|
|
{
|
|
"epoch": 0.7,
|
|
"learning_rate": 6.881525192918748e-05,
|
|
"loss": 0.1538,
|
|
"step": 1540
|
|
},
|
|
{
|
|
"epoch": 0.7,
|
|
"learning_rate": 6.874262369496143e-05,
|
|
"loss": 0.1435,
|
|
"step": 1550
|
|
},
|
|
{
|
|
"epoch": 0.71,
|
|
"learning_rate": 6.866999546073537e-05,
|
|
"loss": 0.1432,
|
|
"step": 1560
|
|
},
|
|
{
|
|
"epoch": 0.71,
|
|
"learning_rate": 6.859736722650931e-05,
|
|
"loss": 0.1504,
|
|
"step": 1570
|
|
},
|
|
{
|
|
"epoch": 0.72,
|
|
"learning_rate": 6.852473899228326e-05,
|
|
"loss": 0.1373,
|
|
"step": 1580
|
|
},
|
|
{
|
|
"epoch": 0.72,
|
|
"learning_rate": 6.84521107580572e-05,
|
|
"loss": 0.1476,
|
|
"step": 1590
|
|
},
|
|
{
|
|
"epoch": 0.73,
|
|
"learning_rate": 6.837948252383114e-05,
|
|
"loss": 0.137,
|
|
"step": 1600
|
|
},
|
|
{
|
|
"epoch": 0.73,
|
|
"eval_loss": 0.13676507771015167,
|
|
"eval_mean_accuracy": 0.7591539253549602,
|
|
"eval_mean_iou": 0.6450359568354115,
|
|
"eval_overall_accuracy": 0.9548248548186227,
|
|
"eval_per_category_accuracy": [
|
|
0.9906033144503231,
|
|
0.7076863291180233,
|
|
0.8855043606250688,
|
|
0.7053038524145415,
|
|
0.9115360838394102,
|
|
0.713271565410552,
|
|
0.8991724867503924,
|
|
0.6526473420003505,
|
|
0.26078634247284016,
|
|
0.5975970695970696,
|
|
0.674821460190049,
|
|
0.9255099176461751,
|
|
0.8615002787619374,
|
|
0.835172377382398,
|
|
0.7989739610277634,
|
|
0.742587358684481,
|
|
0.9155684058838595,
|
|
0.5865281501340482
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9825673729388702,
|
|
0.6056818181818182,
|
|
0.788577002408694,
|
|
0.5720336652181088,
|
|
0.7769033245790549,
|
|
0.6014769492024065,
|
|
0.7574049918658674,
|
|
0.5589727801276676,
|
|
0.21814955859442617,
|
|
0.45864069809283914,
|
|
0.5034824921286137,
|
|
0.8186695203922474,
|
|
0.7627312280813493,
|
|
0.7667370819011305,
|
|
0.6650399210549923,
|
|
0.6483318658762962,
|
|
0.7662021076878184,
|
|
0.3590448447052066
|
|
],
|
|
"eval_runtime": 77.081,
|
|
"eval_samples_per_second": 1.155,
|
|
"eval_steps_per_second": 0.156,
|
|
"step": 1600
|
|
},
|
|
{
|
|
"epoch": 0.73,
|
|
"learning_rate": 6.830685428960509e-05,
|
|
"loss": 0.1419,
|
|
"step": 1610
|
|
},
|
|
{
|
|
"epoch": 0.74,
|
|
"learning_rate": 6.823422605537904e-05,
|
|
"loss": 0.1592,
|
|
"step": 1620
|
|
},
|
|
{
|
|
"epoch": 0.74,
|
|
"learning_rate": 6.816159782115299e-05,
|
|
"loss": 0.1275,
|
|
"step": 1630
|
|
},
|
|
{
|
|
"epoch": 0.74,
|
|
"learning_rate": 6.808896958692693e-05,
|
|
"loss": 0.146,
|
|
"step": 1640
|
|
},
|
|
{
|
|
"epoch": 0.75,
|
|
"learning_rate": 6.801634135270087e-05,
|
|
"loss": 0.1578,
|
|
"step": 1650
|
|
},
|
|
{
|
|
"epoch": 0.75,
|
|
"learning_rate": 6.794371311847482e-05,
|
|
"loss": 0.1545,
|
|
"step": 1660
|
|
},
|
|
{
|
|
"epoch": 0.76,
|
|
"learning_rate": 6.787108488424876e-05,
|
|
"loss": 0.1419,
|
|
"step": 1670
|
|
},
|
|
{
|
|
"epoch": 0.76,
|
|
"learning_rate": 6.77984566500227e-05,
|
|
"loss": 0.139,
|
|
"step": 1680
|
|
},
|
|
{
|
|
"epoch": 0.77,
|
|
"learning_rate": 6.772582841579665e-05,
|
|
"loss": 0.1429,
|
|
"step": 1690
|
|
},
|
|
{
|
|
"epoch": 0.77,
|
|
"learning_rate": 6.765320018157059e-05,
|
|
"loss": 0.1465,
|
|
"step": 1700
|
|
},
|
|
{
|
|
"epoch": 0.77,
|
|
"eval_loss": 0.14493517577648163,
|
|
"eval_mean_accuracy": 0.7614594499739448,
|
|
"eval_mean_iou": 0.6397058212331452,
|
|
"eval_overall_accuracy": 0.9530878817097524,
|
|
"eval_per_category_accuracy": [
|
|
0.9910301541293397,
|
|
0.6895027445392574,
|
|
0.8786986216573126,
|
|
0.7249728703201302,
|
|
0.8644518897007314,
|
|
0.7707660782815681,
|
|
0.7903360511123406,
|
|
0.7853730086905001,
|
|
0.21132953957578893,
|
|
0.6756825396825397,
|
|
0.7357905919848905,
|
|
0.8946479897838558,
|
|
0.8488390914159308,
|
|
0.8385732889850941,
|
|
0.7876196326952923,
|
|
0.7315904419321686,
|
|
0.8796258867600818,
|
|
0.6074396782841823
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9826954864563329,
|
|
0.6211237200618455,
|
|
0.7859897103043074,
|
|
0.5849614184862912,
|
|
0.7605278248954189,
|
|
0.6214059754580039,
|
|
0.7406254585772285,
|
|
0.5496833590615948,
|
|
0.19048729307530893,
|
|
0.48219974486417155,
|
|
0.5046995715394218,
|
|
0.819995651543702,
|
|
0.7667894610720073,
|
|
0.7709211367184656,
|
|
0.6724932822910148,
|
|
0.6410459900151405,
|
|
0.7675053511570615,
|
|
0.2515543466192961
|
|
],
|
|
"eval_runtime": 75.6776,
|
|
"eval_samples_per_second": 1.176,
|
|
"eval_steps_per_second": 0.159,
|
|
"step": 1700
|
|
},
|
|
{
|
|
"epoch": 0.78,
|
|
"learning_rate": 6.758057194734453e-05,
|
|
"loss": 0.1868,
|
|
"step": 1710
|
|
},
|
|
{
|
|
"epoch": 0.78,
|
|
"learning_rate": 6.750794371311849e-05,
|
|
"loss": 0.1528,
|
|
"step": 1720
|
|
},
|
|
{
|
|
"epoch": 0.79,
|
|
"learning_rate": 6.743531547889243e-05,
|
|
"loss": 0.1388,
|
|
"step": 1730
|
|
},
|
|
{
|
|
"epoch": 0.79,
|
|
"learning_rate": 6.736268724466638e-05,
|
|
"loss": 0.1397,
|
|
"step": 1740
|
|
},
|
|
{
|
|
"epoch": 0.79,
|
|
"learning_rate": 6.729005901044032e-05,
|
|
"loss": 0.1365,
|
|
"step": 1750
|
|
},
|
|
{
|
|
"epoch": 0.8,
|
|
"learning_rate": 6.721743077621426e-05,
|
|
"loss": 0.1386,
|
|
"step": 1760
|
|
},
|
|
{
|
|
"epoch": 0.8,
|
|
"learning_rate": 6.714480254198821e-05,
|
|
"loss": 0.1486,
|
|
"step": 1770
|
|
},
|
|
{
|
|
"epoch": 0.81,
|
|
"learning_rate": 6.707217430776215e-05,
|
|
"loss": 0.1505,
|
|
"step": 1780
|
|
},
|
|
{
|
|
"epoch": 0.81,
|
|
"learning_rate": 6.69995460735361e-05,
|
|
"loss": 0.129,
|
|
"step": 1790
|
|
},
|
|
{
|
|
"epoch": 0.82,
|
|
"learning_rate": 6.692691783931004e-05,
|
|
"loss": 0.1407,
|
|
"step": 1800
|
|
},
|
|
{
|
|
"epoch": 0.82,
|
|
"eval_loss": 0.13847893476486206,
|
|
"eval_mean_accuracy": 0.7597777775507725,
|
|
"eval_mean_iou": 0.6430529149418526,
|
|
"eval_overall_accuracy": 0.9541733988215414,
|
|
"eval_per_category_accuracy": [
|
|
0.989219455280459,
|
|
0.7124629608640036,
|
|
0.9102512517882689,
|
|
0.6443298969072165,
|
|
0.8693117376683845,
|
|
0.7514032476737664,
|
|
0.8841255258954577,
|
|
0.737017989750593,
|
|
0.1948784273150543,
|
|
0.6101587301587301,
|
|
0.6794448051309292,
|
|
0.9216195593299411,
|
|
0.8343704476377174,
|
|
0.8683631628037942,
|
|
0.8187241334712881,
|
|
0.7679085303186023,
|
|
0.895680911399592,
|
|
0.5867292225201073
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.982356198961782,
|
|
0.5846609707809033,
|
|
0.7888197647128946,
|
|
0.5730831875490137,
|
|
0.7745111703668928,
|
|
0.6120406691378941,
|
|
0.7480644634158191,
|
|
0.5691702692485286,
|
|
0.172056271124509,
|
|
0.4593757813763991,
|
|
0.5119479980432559,
|
|
0.8253918794990458,
|
|
0.758100887298397,
|
|
0.758420079849063,
|
|
0.6812195942158951,
|
|
0.6619563893287855,
|
|
0.7492243600002065,
|
|
0.36455253404405946
|
|
],
|
|
"eval_runtime": 75.7862,
|
|
"eval_samples_per_second": 1.174,
|
|
"eval_steps_per_second": 0.158,
|
|
"step": 1800
|
|
},
|
|
{
|
|
"epoch": 0.82,
|
|
"learning_rate": 6.685428960508398e-05,
|
|
"loss": 0.1381,
|
|
"step": 1810
|
|
},
|
|
{
|
|
"epoch": 0.83,
|
|
"learning_rate": 6.678166137085794e-05,
|
|
"loss": 0.1347,
|
|
"step": 1820
|
|
},
|
|
{
|
|
"epoch": 0.83,
|
|
"learning_rate": 6.670903313663188e-05,
|
|
"loss": 0.1607,
|
|
"step": 1830
|
|
},
|
|
{
|
|
"epoch": 0.84,
|
|
"learning_rate": 6.663640490240581e-05,
|
|
"loss": 0.1324,
|
|
"step": 1840
|
|
},
|
|
{
|
|
"epoch": 0.84,
|
|
"learning_rate": 6.656377666817975e-05,
|
|
"loss": 0.1427,
|
|
"step": 1850
|
|
},
|
|
{
|
|
"epoch": 0.84,
|
|
"learning_rate": 6.64911484339537e-05,
|
|
"loss": 0.1455,
|
|
"step": 1860
|
|
},
|
|
{
|
|
"epoch": 0.85,
|
|
"learning_rate": 6.641852019972764e-05,
|
|
"loss": 0.1442,
|
|
"step": 1870
|
|
},
|
|
{
|
|
"epoch": 0.85,
|
|
"learning_rate": 6.634589196550158e-05,
|
|
"loss": 0.1352,
|
|
"step": 1880
|
|
},
|
|
{
|
|
"epoch": 0.86,
|
|
"learning_rate": 6.627326373127554e-05,
|
|
"loss": 0.1609,
|
|
"step": 1890
|
|
},
|
|
{
|
|
"epoch": 0.86,
|
|
"learning_rate": 6.620063549704948e-05,
|
|
"loss": 0.1539,
|
|
"step": 1900
|
|
},
|
|
{
|
|
"epoch": 0.86,
|
|
"eval_loss": 0.14046506583690643,
|
|
"eval_mean_accuracy": 0.7531720155067204,
|
|
"eval_mean_iou": 0.6426226055599306,
|
|
"eval_overall_accuracy": 0.9543066131934691,
|
|
"eval_per_category_accuracy": [
|
|
0.9924802996431188,
|
|
0.6779578684888032,
|
|
0.9062396830637174,
|
|
0.6718665219750407,
|
|
0.8569332939038171,
|
|
0.7683724976704543,
|
|
0.8697345206652121,
|
|
0.7159754548919514,
|
|
0.20801862390067252,
|
|
0.5993455433455434,
|
|
0.6479568750122962,
|
|
0.9047059893331436,
|
|
0.8610956243368164,
|
|
0.871962057617514,
|
|
0.8265810915675117,
|
|
0.7592433196300102,
|
|
0.8344447084453194,
|
|
0.5841823056300268
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9821772119045775,
|
|
0.6070843422407168,
|
|
0.7870250768672508,
|
|
0.5831518219815153,
|
|
0.7685410986173595,
|
|
0.6149549094212183,
|
|
0.7837605189258348,
|
|
0.5437053152555021,
|
|
0.17449984811005512,
|
|
0.47517192243355433,
|
|
0.5019584533552802,
|
|
0.8281520299369558,
|
|
0.763911657831652,
|
|
0.7773681113943911,
|
|
0.6905063091695666,
|
|
0.659743579725272,
|
|
0.7300270499370994,
|
|
0.2954676429709482
|
|
],
|
|
"eval_runtime": 79.4887,
|
|
"eval_samples_per_second": 1.12,
|
|
"eval_steps_per_second": 0.151,
|
|
"step": 1900
|
|
},
|
|
{
|
|
"epoch": 0.87,
|
|
"learning_rate": 6.612800726282343e-05,
|
|
"loss": 0.1351,
|
|
"step": 1910
|
|
},
|
|
{
|
|
"epoch": 0.87,
|
|
"learning_rate": 6.605537902859737e-05,
|
|
"loss": 0.1178,
|
|
"step": 1920
|
|
},
|
|
{
|
|
"epoch": 0.88,
|
|
"learning_rate": 6.598275079437131e-05,
|
|
"loss": 0.1608,
|
|
"step": 1930
|
|
},
|
|
{
|
|
"epoch": 0.88,
|
|
"learning_rate": 6.591012256014526e-05,
|
|
"loss": 0.1409,
|
|
"step": 1940
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 6.58374943259192e-05,
|
|
"loss": 0.1614,
|
|
"step": 1950
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 6.576486609169314e-05,
|
|
"loss": 0.1299,
|
|
"step": 1960
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 6.569223785746709e-05,
|
|
"loss": 0.1488,
|
|
"step": 1970
|
|
},
|
|
{
|
|
"epoch": 0.9,
|
|
"learning_rate": 6.561960962324103e-05,
|
|
"loss": 0.1324,
|
|
"step": 1980
|
|
},
|
|
{
|
|
"epoch": 0.9,
|
|
"learning_rate": 6.554698138901497e-05,
|
|
"loss": 0.1705,
|
|
"step": 1990
|
|
},
|
|
{
|
|
"epoch": 0.91,
|
|
"learning_rate": 6.547435315478893e-05,
|
|
"loss": 0.1715,
|
|
"step": 2000
|
|
},
|
|
{
|
|
"epoch": 0.91,
|
|
"eval_loss": 0.15044108033180237,
|
|
"eval_mean_accuracy": 0.7669366594749936,
|
|
"eval_mean_iou": 0.638605446391735,
|
|
"eval_overall_accuracy": 0.9518973532687413,
|
|
"eval_per_category_accuracy": [
|
|
0.9927349057674444,
|
|
0.6858757428067165,
|
|
0.9015540745020358,
|
|
0.7554937601736299,
|
|
0.7849426362290928,
|
|
0.6700126744951179,
|
|
0.8660354181238625,
|
|
0.8764435243942752,
|
|
0.2963269529229177,
|
|
0.7319365079365079,
|
|
0.6655550965000295,
|
|
0.8966643340331234,
|
|
0.8368163588295595,
|
|
0.8751077908593351,
|
|
0.7769119675806173,
|
|
0.7467818602261048,
|
|
0.845331144526079,
|
|
0.6003351206434316
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9816996750938647,
|
|
0.6128061571401704,
|
|
0.7911633072486774,
|
|
0.5907403478998727,
|
|
0.7348832458745334,
|
|
0.5780840395049686,
|
|
0.7868250201946592,
|
|
0.5639012019683365,
|
|
0.2508759635599159,
|
|
0.5183775968343353,
|
|
0.5151988182081233,
|
|
0.8226874204231764,
|
|
0.7582076905214873,
|
|
0.7750139296387645,
|
|
0.6662846209659903,
|
|
0.6491362270935466,
|
|
0.7512367549305662,
|
|
0.14777601795024087
|
|
],
|
|
"eval_runtime": 88.822,
|
|
"eval_samples_per_second": 1.002,
|
|
"eval_steps_per_second": 0.135,
|
|
"step": 2000
|
|
},
|
|
{
|
|
"epoch": 0.91,
|
|
"learning_rate": 6.540172492056287e-05,
|
|
"loss": 0.1778,
|
|
"step": 2010
|
|
},
|
|
{
|
|
"epoch": 0.92,
|
|
"learning_rate": 6.532909668633682e-05,
|
|
"loss": 0.1385,
|
|
"step": 2020
|
|
},
|
|
{
|
|
"epoch": 0.92,
|
|
"learning_rate": 6.525646845211076e-05,
|
|
"loss": 0.1401,
|
|
"step": 2030
|
|
},
|
|
{
|
|
"epoch": 0.93,
|
|
"learning_rate": 6.51838402178847e-05,
|
|
"loss": 0.1419,
|
|
"step": 2040
|
|
},
|
|
{
|
|
"epoch": 0.93,
|
|
"learning_rate": 6.511121198365865e-05,
|
|
"loss": 0.1546,
|
|
"step": 2050
|
|
},
|
|
{
|
|
"epoch": 0.94,
|
|
"learning_rate": 6.503858374943259e-05,
|
|
"loss": 0.1494,
|
|
"step": 2060
|
|
},
|
|
{
|
|
"epoch": 0.94,
|
|
"learning_rate": 6.496595551520653e-05,
|
|
"loss": 0.1306,
|
|
"step": 2070
|
|
},
|
|
{
|
|
"epoch": 0.94,
|
|
"learning_rate": 6.489332728098048e-05,
|
|
"loss": 0.1382,
|
|
"step": 2080
|
|
},
|
|
{
|
|
"epoch": 0.95,
|
|
"learning_rate": 6.482069904675442e-05,
|
|
"loss": 0.1182,
|
|
"step": 2090
|
|
},
|
|
{
|
|
"epoch": 0.95,
|
|
"learning_rate": 6.474807081252838e-05,
|
|
"loss": 0.132,
|
|
"step": 2100
|
|
},
|
|
{
|
|
"epoch": 0.95,
|
|
"eval_loss": 0.13885533809661865,
|
|
"eval_mean_accuracy": 0.7716184264734858,
|
|
"eval_mean_iou": 0.6614885882734279,
|
|
"eval_overall_accuracy": 0.957003561298499,
|
|
"eval_per_category_accuracy": [
|
|
0.9907188024570434,
|
|
0.7016467235544617,
|
|
0.8972966187960824,
|
|
0.7398941942485079,
|
|
0.8977901860300126,
|
|
0.7837386669257504,
|
|
0.8968322737127088,
|
|
0.7030237015280179,
|
|
0.26797723745473356,
|
|
0.7476336996336996,
|
|
0.6871471010643531,
|
|
0.9032036151866305,
|
|
0.8693505746092837,
|
|
0.8497351424599194,
|
|
0.7875549663735126,
|
|
0.7708954265159301,
|
|
0.8718375180900599,
|
|
0.5228552278820375
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9831048553381242,
|
|
0.5972434704706774,
|
|
0.7928731831980577,
|
|
0.5866946326772077,
|
|
0.7869940498207969,
|
|
0.64052506965799,
|
|
0.7954271887210573,
|
|
0.5656907305204008,
|
|
0.2305706400783406,
|
|
0.5128138127876827,
|
|
0.5270090835018257,
|
|
0.8293610888245908,
|
|
0.7775557065818417,
|
|
0.785420326631531,
|
|
0.6790745869189451,
|
|
0.6645274890780126,
|
|
0.7540405657859621,
|
|
0.39786810832865815
|
|
],
|
|
"eval_runtime": 84.5809,
|
|
"eval_samples_per_second": 1.052,
|
|
"eval_steps_per_second": 0.142,
|
|
"step": 2100
|
|
},
|
|
{
|
|
"epoch": 0.96,
|
|
"learning_rate": 6.467544257830232e-05,
|
|
"loss": 0.1406,
|
|
"step": 2110
|
|
},
|
|
{
|
|
"epoch": 0.96,
|
|
"learning_rate": 6.460281434407626e-05,
|
|
"loss": 0.1566,
|
|
"step": 2120
|
|
},
|
|
{
|
|
"epoch": 0.97,
|
|
"learning_rate": 6.453018610985021e-05,
|
|
"loss": 0.1213,
|
|
"step": 2130
|
|
},
|
|
{
|
|
"epoch": 0.97,
|
|
"learning_rate": 6.445755787562415e-05,
|
|
"loss": 0.1348,
|
|
"step": 2140
|
|
},
|
|
{
|
|
"epoch": 0.98,
|
|
"learning_rate": 6.43849296413981e-05,
|
|
"loss": 0.1395,
|
|
"step": 2150
|
|
},
|
|
{
|
|
"epoch": 0.98,
|
|
"learning_rate": 6.431230140717204e-05,
|
|
"loss": 0.1235,
|
|
"step": 2160
|
|
},
|
|
{
|
|
"epoch": 0.99,
|
|
"learning_rate": 6.423967317294598e-05,
|
|
"loss": 0.1551,
|
|
"step": 2170
|
|
},
|
|
{
|
|
"epoch": 0.99,
|
|
"learning_rate": 6.416704493871992e-05,
|
|
"loss": 0.146,
|
|
"step": 2180
|
|
},
|
|
{
|
|
"epoch": 0.99,
|
|
"learning_rate": 6.409441670449387e-05,
|
|
"loss": 0.1482,
|
|
"step": 2190
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"learning_rate": 6.402178847026782e-05,
|
|
"loss": 0.1399,
|
|
"step": 2200
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"eval_loss": 0.13540491461753845,
|
|
"eval_mean_accuracy": 0.7651852022497144,
|
|
"eval_mean_iou": 0.6567450630698053,
|
|
"eval_overall_accuracy": 0.9556758752029934,
|
|
"eval_per_category_accuracy": [
|
|
0.9901550079804321,
|
|
0.6457844200845221,
|
|
0.8885581737647188,
|
|
0.7129679869777537,
|
|
0.9021107370186209,
|
|
0.748916926854473,
|
|
0.8618665330396322,
|
|
0.7203270272564862,
|
|
0.24423176409725814,
|
|
0.8005079365079365,
|
|
0.6490487713706742,
|
|
0.9108973522632475,
|
|
0.8443114580148553,
|
|
0.886845995459585,
|
|
0.8161913692015865,
|
|
0.7391957862281603,
|
|
0.8769659922301457,
|
|
0.5344504021447721
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9832376527547144,
|
|
0.556791846991484,
|
|
0.7927362260590767,
|
|
0.5874923154306153,
|
|
0.7778940044397533,
|
|
0.6219762445741573,
|
|
0.7687129198409016,
|
|
0.5668561551047446,
|
|
0.2103457494207806,
|
|
0.5276944276313248,
|
|
0.5096671533072247,
|
|
0.8256855746211959,
|
|
0.7678766392286271,
|
|
0.7780968814053941,
|
|
0.6865619546247819,
|
|
0.6525027783447869,
|
|
0.7450485380898826,
|
|
0.46223407338705
|
|
],
|
|
"eval_runtime": 84.7624,
|
|
"eval_samples_per_second": 1.05,
|
|
"eval_steps_per_second": 0.142,
|
|
"step": 2200
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"learning_rate": 6.394916023604177e-05,
|
|
"loss": 0.1206,
|
|
"step": 2210
|
|
},
|
|
{
|
|
"epoch": 1.01,
|
|
"learning_rate": 6.387653200181571e-05,
|
|
"loss": 0.1331,
|
|
"step": 2220
|
|
},
|
|
{
|
|
"epoch": 1.01,
|
|
"learning_rate": 6.380390376758965e-05,
|
|
"loss": 0.138,
|
|
"step": 2230
|
|
},
|
|
{
|
|
"epoch": 1.02,
|
|
"learning_rate": 6.37312755333636e-05,
|
|
"loss": 0.136,
|
|
"step": 2240
|
|
},
|
|
{
|
|
"epoch": 1.02,
|
|
"learning_rate": 6.365864729913754e-05,
|
|
"loss": 0.1374,
|
|
"step": 2250
|
|
},
|
|
{
|
|
"epoch": 1.03,
|
|
"learning_rate": 6.358601906491148e-05,
|
|
"loss": 0.1356,
|
|
"step": 2260
|
|
},
|
|
{
|
|
"epoch": 1.03,
|
|
"learning_rate": 6.351339083068543e-05,
|
|
"loss": 0.1176,
|
|
"step": 2270
|
|
},
|
|
{
|
|
"epoch": 1.03,
|
|
"learning_rate": 6.344076259645937e-05,
|
|
"loss": 0.1337,
|
|
"step": 2280
|
|
},
|
|
{
|
|
"epoch": 1.04,
|
|
"learning_rate": 6.336813436223331e-05,
|
|
"loss": 0.1173,
|
|
"step": 2290
|
|
},
|
|
{
|
|
"epoch": 1.04,
|
|
"learning_rate": 6.329550612800726e-05,
|
|
"loss": 0.1175,
|
|
"step": 2300
|
|
},
|
|
{
|
|
"epoch": 1.04,
|
|
"eval_loss": 0.13777972757816315,
|
|
"eval_mean_accuracy": 0.7627759347632277,
|
|
"eval_mean_iou": 0.6573236880408356,
|
|
"eval_overall_accuracy": 0.9549025632022472,
|
|
"eval_per_category_accuracy": [
|
|
0.9917725057114422,
|
|
0.6548195404718341,
|
|
0.903699997248817,
|
|
0.6839392295170917,
|
|
0.8730676049244813,
|
|
0.5990509585362945,
|
|
0.8252635132078492,
|
|
0.8450377804632645,
|
|
0.3003621314019659,
|
|
0.7047032967032967,
|
|
0.6207381612858801,
|
|
0.9128029952596142,
|
|
0.885446828408539,
|
|
0.8925215141755979,
|
|
0.8265110363855838,
|
|
0.7685508735868448,
|
|
0.8763303329805322,
|
|
0.5653485254691689
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9829498718447592,
|
|
0.601791640005357,
|
|
0.7953290799803878,
|
|
0.587679934728131,
|
|
0.7885745909492605,
|
|
0.5241967894765317,
|
|
0.7877833104680789,
|
|
0.5643395347827241,
|
|
0.2465602174282317,
|
|
0.500346769495381,
|
|
0.48634671558162296,
|
|
0.8297144417850802,
|
|
0.7721932933906334,
|
|
0.7827742814256885,
|
|
0.6904011667686989,
|
|
0.6651360591489007,
|
|
0.7551893590882257,
|
|
0.47051932838734867
|
|
],
|
|
"eval_runtime": 87.5329,
|
|
"eval_samples_per_second": 1.017,
|
|
"eval_steps_per_second": 0.137,
|
|
"step": 2300
|
|
},
|
|
{
|
|
"epoch": 1.05,
|
|
"learning_rate": 6.322287789378122e-05,
|
|
"loss": 0.112,
|
|
"step": 2310
|
|
},
|
|
{
|
|
"epoch": 1.05,
|
|
"learning_rate": 6.315024965955516e-05,
|
|
"loss": 0.1096,
|
|
"step": 2320
|
|
},
|
|
{
|
|
"epoch": 1.06,
|
|
"learning_rate": 6.30776214253291e-05,
|
|
"loss": 0.1316,
|
|
"step": 2330
|
|
},
|
|
{
|
|
"epoch": 1.06,
|
|
"learning_rate": 6.300499319110305e-05,
|
|
"loss": 0.122,
|
|
"step": 2340
|
|
},
|
|
{
|
|
"epoch": 1.07,
|
|
"learning_rate": 6.293236495687699e-05,
|
|
"loss": 0.1171,
|
|
"step": 2350
|
|
},
|
|
{
|
|
"epoch": 1.07,
|
|
"learning_rate": 6.285973672265093e-05,
|
|
"loss": 0.1172,
|
|
"step": 2360
|
|
},
|
|
{
|
|
"epoch": 1.08,
|
|
"learning_rate": 6.278710848842488e-05,
|
|
"loss": 0.1295,
|
|
"step": 2370
|
|
},
|
|
{
|
|
"epoch": 1.08,
|
|
"learning_rate": 6.271448025419882e-05,
|
|
"loss": 0.1036,
|
|
"step": 2380
|
|
},
|
|
{
|
|
"epoch": 1.08,
|
|
"learning_rate": 6.264185201997276e-05,
|
|
"loss": 0.1333,
|
|
"step": 2390
|
|
},
|
|
{
|
|
"epoch": 1.09,
|
|
"learning_rate": 6.25692237857467e-05,
|
|
"loss": 0.1166,
|
|
"step": 2400
|
|
},
|
|
{
|
|
"epoch": 1.09,
|
|
"eval_loss": 0.13342882692813873,
|
|
"eval_mean_accuracy": 0.7821907896480783,
|
|
"eval_mean_iou": 0.6697036848077302,
|
|
"eval_overall_accuracy": 0.958539555581768,
|
|
"eval_per_category_accuracy": [
|
|
0.9915974986695582,
|
|
0.7106008840816723,
|
|
0.9216514663805436,
|
|
0.6840070537167662,
|
|
0.862522496437278,
|
|
0.8110374096334996,
|
|
0.8533460696600514,
|
|
0.7866519265958927,
|
|
0.29689601655457837,
|
|
0.7033455433455433,
|
|
0.6781364968817014,
|
|
0.9052950781432237,
|
|
0.8650432530618851,
|
|
0.901985111662531,
|
|
0.8428231160544921,
|
|
0.7853995375128469,
|
|
0.8831837271232099,
|
|
0.595911528150134
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9838216503792198,
|
|
0.6230886090326977,
|
|
0.7986503480665618,
|
|
0.5904566744730679,
|
|
0.7901460381632656,
|
|
0.6592556695294761,
|
|
0.7894455577822311,
|
|
0.5933676785376083,
|
|
0.2402863841902529,
|
|
0.5234176801122362,
|
|
0.5514907642216587,
|
|
0.8321425456448425,
|
|
0.7751464509318873,
|
|
0.7992249996101608,
|
|
0.7001190731986822,
|
|
0.6766631432729929,
|
|
0.7727151237975934,
|
|
0.3552279355947101
|
|
],
|
|
"eval_runtime": 82.9803,
|
|
"eval_samples_per_second": 1.073,
|
|
"eval_steps_per_second": 0.145,
|
|
"step": 2400
|
|
},
|
|
{
|
|
"epoch": 1.09,
|
|
"learning_rate": 6.249659555152066e-05,
|
|
"loss": 0.1159,
|
|
"step": 2410
|
|
},
|
|
{
|
|
"epoch": 1.1,
|
|
"learning_rate": 6.24239673172946e-05,
|
|
"loss": 0.1178,
|
|
"step": 2420
|
|
},
|
|
{
|
|
"epoch": 1.1,
|
|
"learning_rate": 6.235133908306855e-05,
|
|
"loss": 0.1391,
|
|
"step": 2430
|
|
},
|
|
{
|
|
"epoch": 1.11,
|
|
"learning_rate": 6.227871084884249e-05,
|
|
"loss": 0.1283,
|
|
"step": 2440
|
|
},
|
|
{
|
|
"epoch": 1.11,
|
|
"learning_rate": 6.220608261461644e-05,
|
|
"loss": 0.103,
|
|
"step": 2450
|
|
},
|
|
{
|
|
"epoch": 1.12,
|
|
"learning_rate": 6.213345438039038e-05,
|
|
"loss": 0.1168,
|
|
"step": 2460
|
|
},
|
|
{
|
|
"epoch": 1.12,
|
|
"learning_rate": 6.206082614616432e-05,
|
|
"loss": 0.114,
|
|
"step": 2470
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"learning_rate": 6.198819791193827e-05,
|
|
"loss": 0.1299,
|
|
"step": 2480
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"learning_rate": 6.191556967771221e-05,
|
|
"loss": 0.1257,
|
|
"step": 2490
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"learning_rate": 6.184294144348615e-05,
|
|
"loss": 0.1227,
|
|
"step": 2500
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"eval_loss": 0.13308031857013702,
|
|
"eval_mean_accuracy": 0.7715319784801719,
|
|
"eval_mean_iou": 0.6651590960897119,
|
|
"eval_overall_accuracy": 0.9574322218305609,
|
|
"eval_per_category_accuracy": [
|
|
0.9910145671313145,
|
|
0.7075729853138814,
|
|
0.8978554528447232,
|
|
0.653893109061313,
|
|
0.8888540488098072,
|
|
0.690135091569915,
|
|
0.8967296175274126,
|
|
0.7883589568509718,
|
|
0.26006207966890843,
|
|
0.7302075702075702,
|
|
0.6023431505636546,
|
|
0.9364337591377954,
|
|
0.8841834073701059,
|
|
0.8806689192756454,
|
|
0.8377575875150888,
|
|
0.7643820657759507,
|
|
0.858155415600806,
|
|
0.6189678284182306
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9829147742408483,
|
|
0.6474789231156747,
|
|
0.7973113698759822,
|
|
0.5718607272080195,
|
|
0.7868248176069152,
|
|
0.6189919770898483,
|
|
0.7855232213431205,
|
|
0.5875494919877167,
|
|
0.22661497543163683,
|
|
0.5332215842219765,
|
|
0.521154091663475,
|
|
0.8224189056868451,
|
|
0.7817246574416746,
|
|
0.7985454806576056,
|
|
0.7001360090792815,
|
|
0.6653973685829153,
|
|
0.766913291545905,
|
|
0.3782820628353746
|
|
],
|
|
"eval_runtime": 84.6288,
|
|
"eval_samples_per_second": 1.052,
|
|
"eval_steps_per_second": 0.142,
|
|
"step": 2500
|
|
},
|
|
{
|
|
"epoch": 1.14,
|
|
"learning_rate": 6.17703132092601e-05,
|
|
"loss": 0.1234,
|
|
"step": 2510
|
|
},
|
|
{
|
|
"epoch": 1.14,
|
|
"learning_rate": 6.169768497503405e-05,
|
|
"loss": 0.1139,
|
|
"step": 2520
|
|
},
|
|
{
|
|
"epoch": 1.15,
|
|
"learning_rate": 6.1625056740808e-05,
|
|
"loss": 0.1306,
|
|
"step": 2530
|
|
},
|
|
{
|
|
"epoch": 1.15,
|
|
"learning_rate": 6.155242850658194e-05,
|
|
"loss": 0.1122,
|
|
"step": 2540
|
|
},
|
|
{
|
|
"epoch": 1.16,
|
|
"learning_rate": 6.147980027235588e-05,
|
|
"loss": 0.1281,
|
|
"step": 2550
|
|
},
|
|
{
|
|
"epoch": 1.16,
|
|
"learning_rate": 6.140717203812983e-05,
|
|
"loss": 0.1223,
|
|
"step": 2560
|
|
},
|
|
{
|
|
"epoch": 1.17,
|
|
"learning_rate": 6.133454380390377e-05,
|
|
"loss": 0.1115,
|
|
"step": 2570
|
|
},
|
|
{
|
|
"epoch": 1.17,
|
|
"learning_rate": 6.126191556967771e-05,
|
|
"loss": 0.1315,
|
|
"step": 2580
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"learning_rate": 6.118928733545166e-05,
|
|
"loss": 0.1266,
|
|
"step": 2590
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"learning_rate": 6.11166591012256e-05,
|
|
"loss": 0.1085,
|
|
"step": 2600
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"eval_loss": 0.15071263909339905,
|
|
"eval_mean_accuracy": 0.7684598270876742,
|
|
"eval_mean_iou": 0.6511462088637711,
|
|
"eval_overall_accuracy": 0.9541516679056575,
|
|
"eval_per_category_accuracy": [
|
|
0.9908049469692608,
|
|
0.7208180184264642,
|
|
0.9032684054143282,
|
|
0.7053038524145415,
|
|
0.8603680530349864,
|
|
0.5475006513895575,
|
|
0.9093806874136557,
|
|
0.8241234986714067,
|
|
0.22684945680289706,
|
|
0.7364590964590965,
|
|
0.6194888744614295,
|
|
0.9100670928664903,
|
|
0.898566624103016,
|
|
0.8912676205057811,
|
|
0.8460779875840663,
|
|
0.7418293936279547,
|
|
0.8662554424466092,
|
|
0.6338471849865952
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9835189991166429,
|
|
0.642371682948298,
|
|
0.7972390934173503,
|
|
0.5932793245093565,
|
|
0.7729924294028407,
|
|
0.5067484436887132,
|
|
0.7685437207284704,
|
|
0.5522156820737395,
|
|
0.1984432275874553,
|
|
0.5278800778569728,
|
|
0.5263835371408989,
|
|
0.8333098023031449,
|
|
0.7733761585047307,
|
|
0.7850704159852425,
|
|
0.705621421444815,
|
|
0.6715746135864066,
|
|
0.7473020673059005,
|
|
0.33476106194690264
|
|
],
|
|
"eval_runtime": 79.5578,
|
|
"eval_samples_per_second": 1.119,
|
|
"eval_steps_per_second": 0.151,
|
|
"step": 2600
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"learning_rate": 6.104403086699954e-05,
|
|
"loss": 0.149,
|
|
"step": 2610
|
|
},
|
|
{
|
|
"epoch": 1.19,
|
|
"learning_rate": 6.097140263277349e-05,
|
|
"loss": 0.133,
|
|
"step": 2620
|
|
},
|
|
{
|
|
"epoch": 1.19,
|
|
"learning_rate": 6.0898774398547436e-05,
|
|
"loss": 0.1239,
|
|
"step": 2630
|
|
},
|
|
{
|
|
"epoch": 1.2,
|
|
"learning_rate": 6.082614616432138e-05,
|
|
"loss": 0.1032,
|
|
"step": 2640
|
|
},
|
|
{
|
|
"epoch": 1.2,
|
|
"learning_rate": 6.075351793009533e-05,
|
|
"loss": 0.1371,
|
|
"step": 2650
|
|
},
|
|
{
|
|
"epoch": 1.21,
|
|
"learning_rate": 6.068088969586927e-05,
|
|
"loss": 0.1185,
|
|
"step": 2660
|
|
},
|
|
{
|
|
"epoch": 1.21,
|
|
"learning_rate": 6.0608261461643216e-05,
|
|
"loss": 0.1206,
|
|
"step": 2670
|
|
},
|
|
{
|
|
"epoch": 1.22,
|
|
"learning_rate": 6.053563322741716e-05,
|
|
"loss": 0.1166,
|
|
"step": 2680
|
|
},
|
|
{
|
|
"epoch": 1.22,
|
|
"learning_rate": 6.04630049931911e-05,
|
|
"loss": 0.1285,
|
|
"step": 2690
|
|
},
|
|
{
|
|
"epoch": 1.23,
|
|
"learning_rate": 6.039037675896505e-05,
|
|
"loss": 0.1229,
|
|
"step": 2700
|
|
},
|
|
{
|
|
"epoch": 1.23,
|
|
"eval_loss": 0.1449834704399109,
|
|
"eval_mean_accuracy": 0.7712138167360404,
|
|
"eval_mean_iou": 0.6459629394431743,
|
|
"eval_overall_accuracy": 0.9520613852511631,
|
|
"eval_per_category_accuracy": [
|
|
0.9912854258905917,
|
|
0.7057271005035703,
|
|
0.9015093677781446,
|
|
0.7000813890396094,
|
|
0.8255847701435313,
|
|
0.6761776902388723,
|
|
0.8554566112323266,
|
|
0.7448269757746214,
|
|
0.2387997930677703,
|
|
0.7486007326007326,
|
|
0.7296130161915442,
|
|
0.9225881952928245,
|
|
0.8640990594032696,
|
|
0.8880866902486669,
|
|
0.8219628384204173,
|
|
0.7897738951695786,
|
|
0.8721121475716891,
|
|
0.6055630026809652
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9831245174321325,
|
|
0.6157204006385353,
|
|
0.8005328854448983,
|
|
0.590976754837971,
|
|
0.7490012287780567,
|
|
0.5485306287683649,
|
|
0.7679823554312188,
|
|
0.49237726422180267,
|
|
0.19991338241663056,
|
|
0.5389678889404617,
|
|
0.5608011553089015,
|
|
0.8298057706942044,
|
|
0.7745812442567186,
|
|
0.7898227491489611,
|
|
0.6985253709470599,
|
|
0.6807784945045818,
|
|
0.77171425918948,
|
|
0.23417655901715825
|
|
],
|
|
"eval_runtime": 85.4956,
|
|
"eval_samples_per_second": 1.041,
|
|
"eval_steps_per_second": 0.14,
|
|
"step": 2700
|
|
},
|
|
{
|
|
"epoch": 1.23,
|
|
"learning_rate": 6.0317748524738996e-05,
|
|
"loss": 0.1234,
|
|
"step": 2710
|
|
},
|
|
{
|
|
"epoch": 1.23,
|
|
"learning_rate": 6.024512029051294e-05,
|
|
"loss": 0.1113,
|
|
"step": 2720
|
|
},
|
|
{
|
|
"epoch": 1.24,
|
|
"learning_rate": 6.017249205628688e-05,
|
|
"loss": 0.1159,
|
|
"step": 2730
|
|
},
|
|
{
|
|
"epoch": 1.24,
|
|
"learning_rate": 6.0099863822060826e-05,
|
|
"loss": 0.1205,
|
|
"step": 2740
|
|
},
|
|
{
|
|
"epoch": 1.25,
|
|
"learning_rate": 6.0027235587834776e-05,
|
|
"loss": 0.1389,
|
|
"step": 2750
|
|
},
|
|
{
|
|
"epoch": 1.25,
|
|
"learning_rate": 5.995460735360872e-05,
|
|
"loss": 0.1215,
|
|
"step": 2760
|
|
},
|
|
{
|
|
"epoch": 1.26,
|
|
"learning_rate": 5.988197911938266e-05,
|
|
"loss": 0.1106,
|
|
"step": 2770
|
|
},
|
|
{
|
|
"epoch": 1.26,
|
|
"learning_rate": 5.9809350885156606e-05,
|
|
"loss": 0.1076,
|
|
"step": 2780
|
|
},
|
|
{
|
|
"epoch": 1.27,
|
|
"learning_rate": 5.973672265093055e-05,
|
|
"loss": 0.1136,
|
|
"step": 2790
|
|
},
|
|
{
|
|
"epoch": 1.27,
|
|
"learning_rate": 5.96640944167045e-05,
|
|
"loss": 0.1403,
|
|
"step": 2800
|
|
},
|
|
{
|
|
"epoch": 1.27,
|
|
"eval_loss": 0.14680607616901398,
|
|
"eval_mean_accuracy": 0.770428266757619,
|
|
"eval_mean_iou": 0.6534443453914569,
|
|
"eval_overall_accuracy": 0.9538569503955627,
|
|
"eval_per_category_accuracy": [
|
|
0.9909573223556492,
|
|
0.700416133680921,
|
|
0.9043791955540883,
|
|
0.652197504069452,
|
|
0.8206913018807468,
|
|
0.6879667371786662,
|
|
0.8194730083830085,
|
|
0.8910932761071763,
|
|
0.24345576823590273,
|
|
0.713025641025641,
|
|
0.7126345196639714,
|
|
0.932839922034689,
|
|
0.8290424977069583,
|
|
0.900564032170072,
|
|
0.8147417658216934,
|
|
0.8066546762589928,
|
|
0.857495070555091,
|
|
0.5900804289544236
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9831513783900067,
|
|
0.6420715144498375,
|
|
0.7962268549706841,
|
|
0.5800808348917175,
|
|
0.7652580537636267,
|
|
0.5741292193840133,
|
|
0.749765986169305,
|
|
0.5667862735838759,
|
|
0.20660286241109843,
|
|
0.544457787291808,
|
|
0.5596153103395002,
|
|
0.8304156548058987,
|
|
0.7615320758769086,
|
|
0.7683942219619503,
|
|
0.693968227739451,
|
|
0.6743998711132593,
|
|
0.7352683591527868,
|
|
0.3298737307504965
|
|
],
|
|
"eval_runtime": 81.9751,
|
|
"eval_samples_per_second": 1.086,
|
|
"eval_steps_per_second": 0.146,
|
|
"step": 2800
|
|
},
|
|
{
|
|
"epoch": 1.28,
|
|
"learning_rate": 5.959146618247844e-05,
|
|
"loss": 0.1679,
|
|
"step": 2810
|
|
},
|
|
{
|
|
"epoch": 1.28,
|
|
"learning_rate": 5.9518837948252386e-05,
|
|
"loss": 0.0953,
|
|
"step": 2820
|
|
},
|
|
{
|
|
"epoch": 1.28,
|
|
"learning_rate": 5.944620971402633e-05,
|
|
"loss": 0.1176,
|
|
"step": 2830
|
|
},
|
|
{
|
|
"epoch": 1.29,
|
|
"learning_rate": 5.937358147980027e-05,
|
|
"loss": 0.1342,
|
|
"step": 2840
|
|
},
|
|
{
|
|
"epoch": 1.29,
|
|
"learning_rate": 5.930095324557422e-05,
|
|
"loss": 0.1101,
|
|
"step": 2850
|
|
},
|
|
{
|
|
"epoch": 1.3,
|
|
"learning_rate": 5.9228325011348166e-05,
|
|
"loss": 0.1171,
|
|
"step": 2860
|
|
},
|
|
{
|
|
"epoch": 1.3,
|
|
"learning_rate": 5.915569677712211e-05,
|
|
"loss": 0.1246,
|
|
"step": 2870
|
|
},
|
|
{
|
|
"epoch": 1.31,
|
|
"learning_rate": 5.908306854289605e-05,
|
|
"loss": 0.1304,
|
|
"step": 2880
|
|
},
|
|
{
|
|
"epoch": 1.31,
|
|
"learning_rate": 5.9010440308669996e-05,
|
|
"loss": 0.1311,
|
|
"step": 2890
|
|
},
|
|
{
|
|
"epoch": 1.32,
|
|
"learning_rate": 5.893781207444394e-05,
|
|
"loss": 0.1391,
|
|
"step": 2900
|
|
},
|
|
{
|
|
"epoch": 1.32,
|
|
"eval_loss": 0.13498260080814362,
|
|
"eval_mean_accuracy": 0.7738483377686847,
|
|
"eval_mean_iou": 0.6692940000228469,
|
|
"eval_overall_accuracy": 0.9572275140312281,
|
|
"eval_per_category_accuracy": [
|
|
0.9919934417012437,
|
|
0.7063909713564015,
|
|
0.90625,
|
|
0.7253119913185024,
|
|
0.8756254918683484,
|
|
0.7003188496681225,
|
|
0.865045394913125,
|
|
0.8060849757854597,
|
|
0.26539058458354886,
|
|
0.7178608058608058,
|
|
0.700210509748372,
|
|
0.9006733008346084,
|
|
0.843268348830099,
|
|
0.8895033701142143,
|
|
0.8045298758406622,
|
|
0.782277749229188,
|
|
0.8773547935187442,
|
|
0.5711796246648794
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9824737521334264,
|
|
0.6113851673300073,
|
|
0.805191273527255,
|
|
0.5949043168669337,
|
|
0.7938930015552099,
|
|
0.5960402924207401,
|
|
0.7727594187490577,
|
|
0.5946204163540835,
|
|
0.22388059701492538,
|
|
0.5397400117508813,
|
|
0.5652370705057451,
|
|
0.8345147169258385,
|
|
0.7745163221903326,
|
|
0.792605428122268,
|
|
0.688049294184337,
|
|
0.6662818752290968,
|
|
0.7580698915127164,
|
|
0.45312915403838994
|
|
],
|
|
"eval_runtime": 83.7881,
|
|
"eval_samples_per_second": 1.062,
|
|
"eval_steps_per_second": 0.143,
|
|
"step": 2900
|
|
},
|
|
{
|
|
"epoch": 1.32,
|
|
"learning_rate": 5.886518384021789e-05,
|
|
"loss": 0.1169,
|
|
"step": 2910
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 5.879255560599183e-05,
|
|
"loss": 0.1069,
|
|
"step": 2920
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 5.8719927371765777e-05,
|
|
"loss": 0.1102,
|
|
"step": 2930
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 5.864729913753972e-05,
|
|
"loss": 0.1272,
|
|
"step": 2940
|
|
},
|
|
{
|
|
"epoch": 1.34,
|
|
"learning_rate": 5.857467090331366e-05,
|
|
"loss": 0.1301,
|
|
"step": 2950
|
|
},
|
|
{
|
|
"epoch": 1.34,
|
|
"learning_rate": 5.850204266908761e-05,
|
|
"loss": 0.1067,
|
|
"step": 2960
|
|
},
|
|
{
|
|
"epoch": 1.35,
|
|
"learning_rate": 5.842941443486156e-05,
|
|
"loss": 0.1238,
|
|
"step": 2970
|
|
},
|
|
{
|
|
"epoch": 1.35,
|
|
"learning_rate": 5.83567862006355e-05,
|
|
"loss": 0.1291,
|
|
"step": 2980
|
|
},
|
|
{
|
|
"epoch": 1.36,
|
|
"learning_rate": 5.828415796640944e-05,
|
|
"loss": 0.1353,
|
|
"step": 2990
|
|
},
|
|
{
|
|
"epoch": 1.36,
|
|
"learning_rate": 5.821152973218339e-05,
|
|
"loss": 0.1166,
|
|
"step": 3000
|
|
},
|
|
{
|
|
"epoch": 1.36,
|
|
"eval_loss": 0.13106530904769897,
|
|
"eval_mean_accuracy": 0.7774974792072407,
|
|
"eval_mean_iou": 0.6582082239520443,
|
|
"eval_overall_accuracy": 0.9579072587945489,
|
|
"eval_per_category_accuracy": [
|
|
0.9907893599712356,
|
|
0.7040917113295229,
|
|
0.8964609469571916,
|
|
0.7219886055344547,
|
|
0.89360754686429,
|
|
0.6154107728792302,
|
|
0.8968757720963089,
|
|
0.8536018338382169,
|
|
0.2566994309363683,
|
|
0.6869352869352869,
|
|
0.7524051230596707,
|
|
0.9118383129129057,
|
|
0.8560014747405715,
|
|
0.8767092550551713,
|
|
0.8214940075875151,
|
|
0.7728545734840699,
|
|
0.8906758662400138,
|
|
0.596514745308311
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9834721187016149,
|
|
0.6089343229239602,
|
|
0.7993526714186493,
|
|
0.6008692707157371,
|
|
0.7995465685070772,
|
|
0.5476689690407648,
|
|
0.7929279255720083,
|
|
0.6310686613088036,
|
|
0.21930522407849376,
|
|
0.533351029540025,
|
|
0.5590533340155097,
|
|
0.8322021238593187,
|
|
0.7819552886767732,
|
|
0.7960570469798658,
|
|
0.6906342166185685,
|
|
0.6753671022497643,
|
|
0.7782930854104425,
|
|
0.2176890715194208
|
|
],
|
|
"eval_runtime": 83.8227,
|
|
"eval_samples_per_second": 1.062,
|
|
"eval_steps_per_second": 0.143,
|
|
"step": 3000
|
|
},
|
|
{
|
|
"epoch": 1.37,
|
|
"learning_rate": 5.813890149795734e-05,
|
|
"loss": 0.1305,
|
|
"step": 3010
|
|
},
|
|
{
|
|
"epoch": 1.37,
|
|
"learning_rate": 5.806627326373128e-05,
|
|
"loss": 0.1173,
|
|
"step": 3020
|
|
},
|
|
{
|
|
"epoch": 1.38,
|
|
"learning_rate": 5.7993645029505223e-05,
|
|
"loss": 0.114,
|
|
"step": 3030
|
|
},
|
|
{
|
|
"epoch": 1.38,
|
|
"learning_rate": 5.792101679527917e-05,
|
|
"loss": 0.1251,
|
|
"step": 3040
|
|
},
|
|
{
|
|
"epoch": 1.38,
|
|
"learning_rate": 5.784838856105311e-05,
|
|
"loss": 0.1228,
|
|
"step": 3050
|
|
},
|
|
{
|
|
"epoch": 1.39,
|
|
"learning_rate": 5.777576032682706e-05,
|
|
"loss": 0.1148,
|
|
"step": 3060
|
|
},
|
|
{
|
|
"epoch": 1.39,
|
|
"learning_rate": 5.7703132092601004e-05,
|
|
"loss": 0.1097,
|
|
"step": 3070
|
|
},
|
|
{
|
|
"epoch": 1.4,
|
|
"learning_rate": 5.763050385837495e-05,
|
|
"loss": 0.1262,
|
|
"step": 3080
|
|
},
|
|
{
|
|
"epoch": 1.4,
|
|
"learning_rate": 5.755787562414889e-05,
|
|
"loss": 0.1187,
|
|
"step": 3090
|
|
},
|
|
{
|
|
"epoch": 1.41,
|
|
"learning_rate": 5.7485247389922834e-05,
|
|
"loss": 0.1166,
|
|
"step": 3100
|
|
},
|
|
{
|
|
"epoch": 1.41,
|
|
"eval_loss": 0.13692022860050201,
|
|
"eval_mean_accuracy": 0.7756804972624743,
|
|
"eval_mean_iou": 0.6519453203573877,
|
|
"eval_overall_accuracy": 0.9548195399595111,
|
|
"eval_per_category_accuracy": [
|
|
0.9909551035658947,
|
|
0.6856328632264124,
|
|
0.9032151012435348,
|
|
0.6671188279978296,
|
|
0.8419276642197038,
|
|
0.6501706861450545,
|
|
0.9188302762669339,
|
|
0.7870113964591315,
|
|
0.3085876875323332,
|
|
0.6792380952380952,
|
|
0.694593637490409,
|
|
0.9245373280671166,
|
|
0.8602863154865744,
|
|
0.8926403041022122,
|
|
0.8217041731332988,
|
|
0.7894784172661871,
|
|
0.8945114218092837,
|
|
0.6518096514745308
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9840541655459478,
|
|
0.6288930804532831,
|
|
0.8037017826514703,
|
|
0.5912478961288772,
|
|
0.7658389171268511,
|
|
0.5715827582525308,
|
|
0.7699181361578668,
|
|
0.546816874256367,
|
|
0.25362472894255705,
|
|
0.5236574767868304,
|
|
0.5561234937386784,
|
|
0.8321595951759896,
|
|
0.7793014967925873,
|
|
0.7799771646490314,
|
|
0.7007831314226888,
|
|
0.686008673762705,
|
|
0.7744652770727694,
|
|
0.18686111751594806
|
|
],
|
|
"eval_runtime": 78.5808,
|
|
"eval_samples_per_second": 1.133,
|
|
"eval_steps_per_second": 0.153,
|
|
"step": 3100
|
|
},
|
|
{
|
|
"epoch": 1.41,
|
|
"learning_rate": 5.7412619155696784e-05,
|
|
"loss": 0.1106,
|
|
"step": 3110
|
|
},
|
|
{
|
|
"epoch": 1.42,
|
|
"learning_rate": 5.733999092147073e-05,
|
|
"loss": 0.1186,
|
|
"step": 3120
|
|
},
|
|
{
|
|
"epoch": 1.42,
|
|
"learning_rate": 5.726736268724467e-05,
|
|
"loss": 0.113,
|
|
"step": 3130
|
|
},
|
|
{
|
|
"epoch": 1.43,
|
|
"learning_rate": 5.7194734453018614e-05,
|
|
"loss": 0.1327,
|
|
"step": 3140
|
|
},
|
|
{
|
|
"epoch": 1.43,
|
|
"learning_rate": 5.712210621879256e-05,
|
|
"loss": 0.1111,
|
|
"step": 3150
|
|
},
|
|
{
|
|
"epoch": 1.43,
|
|
"learning_rate": 5.70494779845665e-05,
|
|
"loss": 0.1182,
|
|
"step": 3160
|
|
},
|
|
{
|
|
"epoch": 1.44,
|
|
"learning_rate": 5.697684975034045e-05,
|
|
"loss": 0.1322,
|
|
"step": 3170
|
|
},
|
|
{
|
|
"epoch": 1.44,
|
|
"learning_rate": 5.6904221516114394e-05,
|
|
"loss": 0.115,
|
|
"step": 3180
|
|
},
|
|
{
|
|
"epoch": 1.45,
|
|
"learning_rate": 5.683159328188834e-05,
|
|
"loss": 0.1225,
|
|
"step": 3190
|
|
},
|
|
{
|
|
"epoch": 1.45,
|
|
"learning_rate": 5.675896504766228e-05,
|
|
"loss": 0.114,
|
|
"step": 3200
|
|
},
|
|
{
|
|
"epoch": 1.45,
|
|
"eval_loss": 0.13359396159648895,
|
|
"eval_mean_accuracy": 0.771436184706532,
|
|
"eval_mean_iou": 0.6682175047820916,
|
|
"eval_overall_accuracy": 0.9564487157243021,
|
|
"eval_per_category_accuracy": [
|
|
0.9918683019590915,
|
|
0.7046098544341716,
|
|
0.9019409596126334,
|
|
0.729381443298969,
|
|
0.8858042077514678,
|
|
0.6629136323689824,
|
|
0.9221465930326029,
|
|
0.70561043943836,
|
|
0.3356958096223487,
|
|
0.7674627594627594,
|
|
0.6211119636428023,
|
|
0.9178043197210328,
|
|
0.893944571336079,
|
|
0.8782271263396864,
|
|
0.8123329453354027,
|
|
0.7683260534429599,
|
|
0.8830510409691643,
|
|
0.5036193029490617
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9831915677087678,
|
|
0.6390765434999707,
|
|
0.8028579256467909,
|
|
0.6069191263615328,
|
|
0.7893015970448374,
|
|
0.556611521982168,
|
|
0.7495947861064673,
|
|
0.5759256556220503,
|
|
0.2610112223965247,
|
|
0.542952310530935,
|
|
0.5284296331012319,
|
|
0.8291532131325542,
|
|
0.7835274181789521,
|
|
0.7963091505301707,
|
|
0.7004167847634271,
|
|
0.673731820793295,
|
|
0.7578031697273823,
|
|
0.45110163895059135
|
|
],
|
|
"eval_runtime": 76.057,
|
|
"eval_samples_per_second": 1.17,
|
|
"eval_steps_per_second": 0.158,
|
|
"step": 3200
|
|
},
|
|
{
|
|
"epoch": 1.46,
|
|
"learning_rate": 5.6686336813436224e-05,
|
|
"loss": 0.1078,
|
|
"step": 3210
|
|
},
|
|
{
|
|
"epoch": 1.46,
|
|
"learning_rate": 5.6613708579210174e-05,
|
|
"loss": 0.1245,
|
|
"step": 3220
|
|
},
|
|
{
|
|
"epoch": 1.47,
|
|
"learning_rate": 5.654108034498412e-05,
|
|
"loss": 0.1169,
|
|
"step": 3230
|
|
},
|
|
{
|
|
"epoch": 1.47,
|
|
"learning_rate": 5.646845211075806e-05,
|
|
"loss": 0.1193,
|
|
"step": 3240
|
|
},
|
|
{
|
|
"epoch": 1.48,
|
|
"learning_rate": 5.6395823876532004e-05,
|
|
"loss": 0.1108,
|
|
"step": 3250
|
|
},
|
|
{
|
|
"epoch": 1.48,
|
|
"learning_rate": 5.632319564230595e-05,
|
|
"loss": 0.1122,
|
|
"step": 3260
|
|
},
|
|
{
|
|
"epoch": 1.48,
|
|
"learning_rate": 5.62505674080799e-05,
|
|
"loss": 0.1085,
|
|
"step": 3270
|
|
},
|
|
{
|
|
"epoch": 1.49,
|
|
"learning_rate": 5.617793917385384e-05,
|
|
"loss": 0.1168,
|
|
"step": 3280
|
|
},
|
|
{
|
|
"epoch": 1.49,
|
|
"learning_rate": 5.6105310939627784e-05,
|
|
"loss": 0.1183,
|
|
"step": 3290
|
|
},
|
|
{
|
|
"epoch": 1.5,
|
|
"learning_rate": 5.603268270540173e-05,
|
|
"loss": 0.1472,
|
|
"step": 3300
|
|
},
|
|
{
|
|
"epoch": 1.5,
|
|
"eval_loss": 0.14049187302589417,
|
|
"eval_mean_accuracy": 0.7717139026549295,
|
|
"eval_mean_iou": 0.6523727128294649,
|
|
"eval_overall_accuracy": 0.9561611561293013,
|
|
"eval_per_category_accuracy": [
|
|
0.9934544592848197,
|
|
0.7343383150633915,
|
|
0.8856195664135579,
|
|
0.726465002712968,
|
|
0.8905686838615036,
|
|
0.690963129142948,
|
|
0.8245953780357522,
|
|
0.7321515987377007,
|
|
0.25716502845318157,
|
|
0.623912087912088,
|
|
0.7878081410218576,
|
|
0.9305982216634445,
|
|
0.8350988256029351,
|
|
0.9062571493937314,
|
|
0.7898344542162442,
|
|
0.8042844295991778,
|
|
0.8639288061640433,
|
|
0.6138069705093834
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9834976876071128,
|
|
0.6155098938682446,
|
|
0.7997279625606921,
|
|
0.5913758833922261,
|
|
0.7902125234696338,
|
|
0.5654944340031806,
|
|
0.7642206021553233,
|
|
0.5704052509808999,
|
|
0.22546262699564587,
|
|
0.5149594872415044,
|
|
0.5752302012555126,
|
|
0.8332914408306752,
|
|
0.7620157296824934,
|
|
0.7798415216346063,
|
|
0.685743159785904,
|
|
0.6639006569494006,
|
|
0.7615679852895975,
|
|
0.2602517832277132
|
|
],
|
|
"eval_runtime": 76.2338,
|
|
"eval_samples_per_second": 1.167,
|
|
"eval_steps_per_second": 0.157,
|
|
"step": 3300
|
|
},
|
|
{
|
|
"epoch": 1.5,
|
|
"learning_rate": 5.596005447117567e-05,
|
|
"loss": 0.1187,
|
|
"step": 3310
|
|
},
|
|
{
|
|
"epoch": 1.51,
|
|
"learning_rate": 5.588742623694962e-05,
|
|
"loss": 0.1279,
|
|
"step": 3320
|
|
},
|
|
{
|
|
"epoch": 1.51,
|
|
"learning_rate": 5.5814798002723564e-05,
|
|
"loss": 0.1204,
|
|
"step": 3330
|
|
},
|
|
{
|
|
"epoch": 1.52,
|
|
"learning_rate": 5.574216976849751e-05,
|
|
"loss": 0.1206,
|
|
"step": 3340
|
|
},
|
|
{
|
|
"epoch": 1.52,
|
|
"learning_rate": 5.566954153427145e-05,
|
|
"loss": 0.1258,
|
|
"step": 3350
|
|
},
|
|
{
|
|
"epoch": 1.53,
|
|
"learning_rate": 5.5596913300045394e-05,
|
|
"loss": 0.1214,
|
|
"step": 3360
|
|
},
|
|
{
|
|
"epoch": 1.53,
|
|
"learning_rate": 5.5524285065819344e-05,
|
|
"loss": 0.1229,
|
|
"step": 3370
|
|
},
|
|
{
|
|
"epoch": 1.53,
|
|
"learning_rate": 5.545165683159329e-05,
|
|
"loss": 0.1376,
|
|
"step": 3380
|
|
},
|
|
{
|
|
"epoch": 1.54,
|
|
"learning_rate": 5.537902859736723e-05,
|
|
"loss": 0.1528,
|
|
"step": 3390
|
|
},
|
|
{
|
|
"epoch": 1.54,
|
|
"learning_rate": 5.5306400363141174e-05,
|
|
"loss": 0.1192,
|
|
"step": 3400
|
|
},
|
|
{
|
|
"epoch": 1.54,
|
|
"eval_loss": 0.13206158578395844,
|
|
"eval_mean_accuracy": 0.7787210569092442,
|
|
"eval_mean_iou": 0.6670240956145178,
|
|
"eval_overall_accuracy": 0.9592076419444566,
|
|
"eval_per_category_accuracy": [
|
|
0.9919085729931352,
|
|
0.6951537427743325,
|
|
0.9162230384065148,
|
|
0.6920103092783505,
|
|
0.8718161065914304,
|
|
0.7556494243482792,
|
|
0.8699885512254365,
|
|
0.8346854096782294,
|
|
0.2686497672012416,
|
|
0.6940659340659341,
|
|
0.6739558126266502,
|
|
0.9222165553723713,
|
|
0.8902397352660828,
|
|
0.8791642468718652,
|
|
0.7976105794102432,
|
|
0.7731500513874615,
|
|
0.8725287203809018,
|
|
0.6179624664879356
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9836782455582729,
|
|
0.6310559736594544,
|
|
0.8034299637824291,
|
|
0.59817083895175,
|
|
0.7950283471434801,
|
|
0.6311009231836003,
|
|
0.8056107833249014,
|
|
0.622326255424257,
|
|
0.2306052666637062,
|
|
0.532893848715295,
|
|
0.5422434329763912,
|
|
0.8347761669416341,
|
|
0.7853621193671092,
|
|
0.7871295318827107,
|
|
0.697165385484965,
|
|
0.6722443143737992,
|
|
0.7898672581203839,
|
|
0.26374506550718
|
|
],
|
|
"eval_runtime": 83.3802,
|
|
"eval_samples_per_second": 1.067,
|
|
"eval_steps_per_second": 0.144,
|
|
"step": 3400
|
|
},
|
|
{
|
|
"epoch": 1.55,
|
|
"learning_rate": 5.523377212891512e-05,
|
|
"loss": 0.1606,
|
|
"step": 3410
|
|
},
|
|
{
|
|
"epoch": 1.55,
|
|
"learning_rate": 5.516114389468906e-05,
|
|
"loss": 0.1167,
|
|
"step": 3420
|
|
},
|
|
{
|
|
"epoch": 1.56,
|
|
"learning_rate": 5.508851566046301e-05,
|
|
"loss": 0.132,
|
|
"step": 3430
|
|
},
|
|
{
|
|
"epoch": 1.56,
|
|
"learning_rate": 5.5015887426236954e-05,
|
|
"loss": 0.1212,
|
|
"step": 3440
|
|
},
|
|
{
|
|
"epoch": 1.57,
|
|
"learning_rate": 5.49432591920109e-05,
|
|
"loss": 0.124,
|
|
"step": 3450
|
|
},
|
|
{
|
|
"epoch": 1.57,
|
|
"learning_rate": 5.487063095778484e-05,
|
|
"loss": 0.1371,
|
|
"step": 3460
|
|
},
|
|
{
|
|
"epoch": 1.58,
|
|
"learning_rate": 5.4798002723558784e-05,
|
|
"loss": 0.1138,
|
|
"step": 3470
|
|
},
|
|
{
|
|
"epoch": 1.58,
|
|
"learning_rate": 5.4725374489332735e-05,
|
|
"loss": 0.1013,
|
|
"step": 3480
|
|
},
|
|
{
|
|
"epoch": 1.58,
|
|
"learning_rate": 5.465274625510668e-05,
|
|
"loss": 0.117,
|
|
"step": 3490
|
|
},
|
|
{
|
|
"epoch": 1.59,
|
|
"learning_rate": 5.458011802088062e-05,
|
|
"loss": 0.1087,
|
|
"step": 3500
|
|
},
|
|
{
|
|
"epoch": 1.59,
|
|
"eval_loss": 0.13412128388881683,
|
|
"eval_mean_accuracy": 0.7736412261063906,
|
|
"eval_mean_iou": 0.6695343192185108,
|
|
"eval_overall_accuracy": 0.9579999687966336,
|
|
"eval_per_category_accuracy": [
|
|
0.9908575877561856,
|
|
0.6992017357794006,
|
|
0.9181712198745461,
|
|
0.6848209441128594,
|
|
0.8999960890677092,
|
|
0.7552696311147815,
|
|
0.8742896713958109,
|
|
0.6774670199948337,
|
|
0.21112260734609414,
|
|
0.7002295482295482,
|
|
0.7326526195675697,
|
|
0.9190694768970439,
|
|
0.903854108591263,
|
|
0.8790894532143674,
|
|
0.8390131919296431,
|
|
0.7836459403905447,
|
|
0.9263622702292384,
|
|
0.5304289544235925
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9836469704651415,
|
|
0.633966585430308,
|
|
0.8044954078680937,
|
|
0.6015848427073404,
|
|
0.7850568307979773,
|
|
0.6209403648906235,
|
|
0.7959676026910106,
|
|
0.5669896894747982,
|
|
0.18577866800200302,
|
|
0.5234388690928209,
|
|
0.5514831104595199,
|
|
0.8385657341774435,
|
|
0.7894441280999038,
|
|
0.79647146518701,
|
|
0.700614242321971,
|
|
0.6860835236027849,
|
|
0.7656166318978664,
|
|
0.42147307876657614
|
|
],
|
|
"eval_runtime": 81.1535,
|
|
"eval_samples_per_second": 1.097,
|
|
"eval_steps_per_second": 0.148,
|
|
"step": 3500
|
|
},
|
|
{
|
|
"epoch": 1.59,
|
|
"learning_rate": 5.4507489786654565e-05,
|
|
"loss": 0.1186,
|
|
"step": 3510
|
|
},
|
|
{
|
|
"epoch": 1.6,
|
|
"learning_rate": 5.443486155242851e-05,
|
|
"loss": 0.1306,
|
|
"step": 3520
|
|
},
|
|
{
|
|
"epoch": 1.6,
|
|
"learning_rate": 5.436223331820246e-05,
|
|
"loss": 0.1179,
|
|
"step": 3530
|
|
},
|
|
{
|
|
"epoch": 1.61,
|
|
"learning_rate": 5.42896050839764e-05,
|
|
"loss": 0.1186,
|
|
"step": 3540
|
|
},
|
|
{
|
|
"epoch": 1.61,
|
|
"learning_rate": 5.4216976849750345e-05,
|
|
"loss": 0.113,
|
|
"step": 3550
|
|
},
|
|
{
|
|
"epoch": 1.62,
|
|
"learning_rate": 5.414434861552429e-05,
|
|
"loss": 0.1037,
|
|
"step": 3560
|
|
},
|
|
{
|
|
"epoch": 1.62,
|
|
"learning_rate": 5.407172038129823e-05,
|
|
"loss": 0.1094,
|
|
"step": 3570
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"learning_rate": 5.399909214707218e-05,
|
|
"loss": 0.1148,
|
|
"step": 3580
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"learning_rate": 5.3926463912846125e-05,
|
|
"loss": 0.1478,
|
|
"step": 3590
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"learning_rate": 5.385383567862007e-05,
|
|
"loss": 0.1159,
|
|
"step": 3600
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"eval_loss": 0.13787628710269928,
|
|
"eval_mean_accuracy": 0.7727645737217357,
|
|
"eval_mean_iou": 0.6571073050454355,
|
|
"eval_overall_accuracy": 0.9573489842789896,
|
|
"eval_per_category_accuracy": [
|
|
0.9921308957265332,
|
|
0.7184054145954436,
|
|
0.8962184989545504,
|
|
0.7112723819858926,
|
|
0.8583624995111334,
|
|
0.7488882215519411,
|
|
0.9106491002794336,
|
|
0.8166342421237263,
|
|
0.25333678220382827,
|
|
0.6308473748473749,
|
|
0.7146314112022665,
|
|
0.920907908418435,
|
|
0.834096181860691,
|
|
0.8645090896291995,
|
|
0.7966459734436971,
|
|
0.7785906988694759,
|
|
0.8413166169350733,
|
|
0.6223190348525469
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9837198229522633,
|
|
0.6457848160223568,
|
|
0.8032140149296972,
|
|
0.6036030850696443,
|
|
0.7737964315713036,
|
|
0.6443207897016582,
|
|
0.7965703918921593,
|
|
0.6002291604707481,
|
|
0.22283400072806697,
|
|
0.5215962137976707,
|
|
0.5572404905999033,
|
|
0.8371056660461305,
|
|
0.7682493011698933,
|
|
0.7694771756284191,
|
|
0.692395601101598,
|
|
0.6730952909817859,
|
|
0.7677478761119262,
|
|
0.1669513620426144
|
|
],
|
|
"eval_runtime": 76.762,
|
|
"eval_samples_per_second": 1.159,
|
|
"eval_steps_per_second": 0.156,
|
|
"step": 3600
|
|
},
|
|
{
|
|
"epoch": 1.64,
|
|
"learning_rate": 5.378120744439401e-05,
|
|
"loss": 0.1232,
|
|
"step": 3610
|
|
},
|
|
{
|
|
"epoch": 1.64,
|
|
"learning_rate": 5.3708579210167955e-05,
|
|
"loss": 0.1095,
|
|
"step": 3620
|
|
},
|
|
{
|
|
"epoch": 1.65,
|
|
"learning_rate": 5.3635950975941905e-05,
|
|
"loss": 0.1243,
|
|
"step": 3630
|
|
},
|
|
{
|
|
"epoch": 1.65,
|
|
"learning_rate": 5.356332274171585e-05,
|
|
"loss": 0.1032,
|
|
"step": 3640
|
|
},
|
|
{
|
|
"epoch": 1.66,
|
|
"learning_rate": 5.349069450748979e-05,
|
|
"loss": 0.121,
|
|
"step": 3650
|
|
},
|
|
{
|
|
"epoch": 1.66,
|
|
"learning_rate": 5.3418066273263735e-05,
|
|
"loss": 0.1096,
|
|
"step": 3660
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 5.334543803903768e-05,
|
|
"loss": 0.1148,
|
|
"step": 3670
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 5.327280980481162e-05,
|
|
"loss": 0.1102,
|
|
"step": 3680
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 5.320018157058557e-05,
|
|
"loss": 0.1158,
|
|
"step": 3690
|
|
},
|
|
{
|
|
"epoch": 1.68,
|
|
"learning_rate": 5.3127553336359515e-05,
|
|
"loss": 0.1216,
|
|
"step": 3700
|
|
},
|
|
{
|
|
"epoch": 1.68,
|
|
"eval_loss": 0.13317522406578064,
|
|
"eval_mean_accuracy": 0.7891147473677718,
|
|
"eval_mean_iou": 0.6805865558672544,
|
|
"eval_overall_accuracy": 0.960010357117385,
|
|
"eval_per_category_accuracy": [
|
|
0.9914235455528075,
|
|
0.719619812496964,
|
|
0.9111866540112248,
|
|
0.7301953336950624,
|
|
0.8926250108922895,
|
|
0.6292202314972244,
|
|
0.9373727672279698,
|
|
0.8306318902153749,
|
|
0.35830315571650284,
|
|
0.7416849816849816,
|
|
0.7602549725550375,
|
|
0.916444275756821,
|
|
0.8636989011384277,
|
|
0.8621156925892685,
|
|
0.8228735557854803,
|
|
0.7882130010277493,
|
|
0.8991400085783141,
|
|
0.5490616621983915
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.983791077686731,
|
|
0.6187506091024266,
|
|
0.8036680300831244,
|
|
0.6082142251850178,
|
|
0.820750824561979,
|
|
0.5742093437981723,
|
|
0.754999908907833,
|
|
0.6557655722789334,
|
|
0.2825900689542617,
|
|
0.5596544633051528,
|
|
0.5896273917421954,
|
|
0.8410978587834871,
|
|
0.7865993481073821,
|
|
0.784696276184641,
|
|
0.6989280287079588,
|
|
0.6807787048954773,
|
|
0.7828143889530667,
|
|
0.4236218843727376
|
|
],
|
|
"eval_runtime": 75.9913,
|
|
"eval_samples_per_second": 1.171,
|
|
"eval_steps_per_second": 0.158,
|
|
"step": 3700
|
|
},
|
|
{
|
|
"epoch": 1.68,
|
|
"learning_rate": 5.305492510213346e-05,
|
|
"loss": 0.1336,
|
|
"step": 3710
|
|
},
|
|
{
|
|
"epoch": 1.69,
|
|
"learning_rate": 5.29822968679074e-05,
|
|
"loss": 0.1175,
|
|
"step": 3720
|
|
},
|
|
{
|
|
"epoch": 1.69,
|
|
"learning_rate": 5.2909668633681345e-05,
|
|
"loss": 0.1318,
|
|
"step": 3730
|
|
},
|
|
{
|
|
"epoch": 1.7,
|
|
"learning_rate": 5.2837040399455295e-05,
|
|
"loss": 0.1106,
|
|
"step": 3740
|
|
},
|
|
{
|
|
"epoch": 1.7,
|
|
"learning_rate": 5.276441216522924e-05,
|
|
"loss": 0.099,
|
|
"step": 3750
|
|
},
|
|
{
|
|
"epoch": 1.71,
|
|
"learning_rate": 5.269178393100318e-05,
|
|
"loss": 0.1277,
|
|
"step": 3760
|
|
},
|
|
{
|
|
"epoch": 1.71,
|
|
"learning_rate": 5.2619155696777125e-05,
|
|
"loss": 0.1208,
|
|
"step": 3770
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"learning_rate": 5.254652746255107e-05,
|
|
"loss": 0.1033,
|
|
"step": 3780
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"learning_rate": 5.247389922832502e-05,
|
|
"loss": 0.1213,
|
|
"step": 3790
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"learning_rate": 5.240127099409896e-05,
|
|
"loss": 0.1368,
|
|
"step": 3800
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"eval_loss": 0.13429704308509827,
|
|
"eval_mean_accuracy": 0.775787841116801,
|
|
"eval_mean_iou": 0.6742017874273715,
|
|
"eval_overall_accuracy": 0.9586353087693118,
|
|
"eval_per_category_accuracy": [
|
|
0.9925796459543752,
|
|
0.7299017147298369,
|
|
0.9100345961263343,
|
|
0.6766820401519262,
|
|
0.8686804994740825,
|
|
0.6663163147690989,
|
|
0.9150859354066403,
|
|
0.8352399687134773,
|
|
0.2617692705638903,
|
|
0.6804200244200245,
|
|
0.7714985539750929,
|
|
0.9155982018953636,
|
|
0.8548459615488373,
|
|
0.8289117082871372,
|
|
0.8381401965856182,
|
|
0.7949640287769785,
|
|
0.8650674385092247,
|
|
0.5584450402144772
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9843675219674684,
|
|
0.6595364897290338,
|
|
0.8064101782721317,
|
|
0.6016039556198746,
|
|
0.7936248541800263,
|
|
0.597209880403655,
|
|
0.7952325150637706,
|
|
0.5868266284910741,
|
|
0.22507895556247498,
|
|
0.5462558520691034,
|
|
0.5834058601682622,
|
|
0.8430500293048027,
|
|
0.7802171647119654,
|
|
0.7659144588941692,
|
|
0.7167505391804457,
|
|
0.6883241842279434,
|
|
0.7807226164351516,
|
|
0.38110048941133423
|
|
],
|
|
"eval_runtime": 87.7077,
|
|
"eval_samples_per_second": 1.015,
|
|
"eval_steps_per_second": 0.137,
|
|
"step": 3800
|
|
},
|
|
{
|
|
"epoch": 1.73,
|
|
"learning_rate": 5.2328642759872905e-05,
|
|
"loss": 0.1086,
|
|
"step": 3810
|
|
},
|
|
{
|
|
"epoch": 1.73,
|
|
"learning_rate": 5.225601452564685e-05,
|
|
"loss": 0.124,
|
|
"step": 3820
|
|
},
|
|
{
|
|
"epoch": 1.74,
|
|
"learning_rate": 5.218338629142079e-05,
|
|
"loss": 0.0965,
|
|
"step": 3830
|
|
},
|
|
{
|
|
"epoch": 1.74,
|
|
"learning_rate": 5.211075805719474e-05,
|
|
"loss": 0.109,
|
|
"step": 3840
|
|
},
|
|
{
|
|
"epoch": 1.75,
|
|
"learning_rate": 5.2038129822968685e-05,
|
|
"loss": 0.1345,
|
|
"step": 3850
|
|
},
|
|
{
|
|
"epoch": 1.75,
|
|
"learning_rate": 5.196550158874263e-05,
|
|
"loss": 0.1153,
|
|
"step": 3860
|
|
},
|
|
{
|
|
"epoch": 1.76,
|
|
"learning_rate": 5.189287335451657e-05,
|
|
"loss": 0.1179,
|
|
"step": 3870
|
|
},
|
|
{
|
|
"epoch": 1.76,
|
|
"learning_rate": 5.1820245120290515e-05,
|
|
"loss": 0.1069,
|
|
"step": 3880
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"learning_rate": 5.1747616886064465e-05,
|
|
"loss": 0.1188,
|
|
"step": 3890
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"learning_rate": 5.167498865183841e-05,
|
|
"loss": 0.1007,
|
|
"step": 3900
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"eval_loss": 0.14351992309093475,
|
|
"eval_mean_accuracy": 0.7664950013504883,
|
|
"eval_mean_iou": 0.6695093049208686,
|
|
"eval_overall_accuracy": 0.9566625102182452,
|
|
"eval_per_category_accuracy": [
|
|
0.9922738412564651,
|
|
0.7236516135300118,
|
|
0.884450313634863,
|
|
0.7062533912099838,
|
|
0.8947739652668045,
|
|
0.6356767164666864,
|
|
0.9453869094224459,
|
|
0.7349442641073858,
|
|
0.2869115364718055,
|
|
0.6710525030525031,
|
|
0.6838910857974778,
|
|
0.927684406542444,
|
|
0.8642294480513641,
|
|
0.8367650423244108,
|
|
0.812623943783411,
|
|
0.7936343782117163,
|
|
0.8520364238921477,
|
|
0.5506702412868633
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9842857601121769,
|
|
0.6512020982077809,
|
|
0.8017007587329098,
|
|
0.6066767653227686,
|
|
0.7885149683436959,
|
|
0.5786775909271819,
|
|
0.7782877114577514,
|
|
0.5204883271991803,
|
|
0.24535480445938773,
|
|
0.5381951788917875,
|
|
0.5561617535298589,
|
|
0.8366714566388658,
|
|
0.7948352148203284,
|
|
0.7821566780857128,
|
|
0.7020867476162098,
|
|
0.6825906323547286,
|
|
0.7816882668342591,
|
|
0.4215927750410509
|
|
],
|
|
"eval_runtime": 95.7785,
|
|
"eval_samples_per_second": 0.929,
|
|
"eval_steps_per_second": 0.125,
|
|
"step": 3900
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"learning_rate": 5.160236041761235e-05,
|
|
"loss": 0.104,
|
|
"step": 3910
|
|
},
|
|
{
|
|
"epoch": 1.78,
|
|
"learning_rate": 5.1529732183386295e-05,
|
|
"loss": 0.0936,
|
|
"step": 3920
|
|
},
|
|
{
|
|
"epoch": 1.78,
|
|
"learning_rate": 5.145710394916024e-05,
|
|
"loss": 0.0966,
|
|
"step": 3930
|
|
},
|
|
{
|
|
"epoch": 1.79,
|
|
"learning_rate": 5.138447571493418e-05,
|
|
"loss": 0.1005,
|
|
"step": 3940
|
|
},
|
|
{
|
|
"epoch": 1.79,
|
|
"learning_rate": 5.131184748070813e-05,
|
|
"loss": 0.1222,
|
|
"step": 3950
|
|
},
|
|
{
|
|
"epoch": 1.8,
|
|
"learning_rate": 5.1239219246482076e-05,
|
|
"loss": 0.1345,
|
|
"step": 3960
|
|
},
|
|
{
|
|
"epoch": 1.8,
|
|
"learning_rate": 5.116659101225602e-05,
|
|
"loss": 0.114,
|
|
"step": 3970
|
|
},
|
|
{
|
|
"epoch": 1.81,
|
|
"learning_rate": 5.109396277802996e-05,
|
|
"loss": 0.1373,
|
|
"step": 3980
|
|
},
|
|
{
|
|
"epoch": 1.81,
|
|
"learning_rate": 5.1021334543803906e-05,
|
|
"loss": 0.1069,
|
|
"step": 3990
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"learning_rate": 5.0948706309577856e-05,
|
|
"loss": 0.137,
|
|
"step": 4000
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"eval_loss": 0.13786786794662476,
|
|
"eval_mean_accuracy": 0.7728080994273032,
|
|
"eval_mean_iou": 0.6646466356753247,
|
|
"eval_overall_accuracy": 0.9552087676659059,
|
|
"eval_per_category_accuracy": [
|
|
0.9923036839786628,
|
|
0.7142764617302741,
|
|
0.9116130873775724,
|
|
0.6671188279978296,
|
|
0.8388544947933141,
|
|
0.676539818670812,
|
|
0.888999084794009,
|
|
0.7545236301759596,
|
|
0.28965338851526123,
|
|
0.662954822954823,
|
|
0.6863011273092132,
|
|
0.930776134391321,
|
|
0.8912154020466522,
|
|
0.8834054872146841,
|
|
0.8102690118986032,
|
|
0.7838514902363823,
|
|
0.9125413101369136,
|
|
0.6153485254691688
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9843104882002514,
|
|
0.6397175051118813,
|
|
0.8053449145384836,
|
|
0.5962296175062133,
|
|
0.7682931274146803,
|
|
0.5578943535024309,
|
|
0.8146034185608627,
|
|
0.4992255233095971,
|
|
0.25318802568508636,
|
|
0.5288708797631108,
|
|
0.545629443093214,
|
|
0.8386195890684222,
|
|
0.7933122548627232,
|
|
0.8009565637104141,
|
|
0.6963529003126085,
|
|
0.6757295531314026,
|
|
0.7936759275163174,
|
|
0.37168535686814297
|
|
],
|
|
"eval_runtime": 87.0618,
|
|
"eval_samples_per_second": 1.022,
|
|
"eval_steps_per_second": 0.138,
|
|
"step": 4000
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"learning_rate": 5.08760780753518e-05,
|
|
"loss": 0.1123,
|
|
"step": 4010
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"learning_rate": 5.080344984112574e-05,
|
|
"loss": 0.1155,
|
|
"step": 4020
|
|
},
|
|
{
|
|
"epoch": 1.83,
|
|
"learning_rate": 5.0730821606899686e-05,
|
|
"loss": 0.1171,
|
|
"step": 4030
|
|
},
|
|
{
|
|
"epoch": 1.83,
|
|
"learning_rate": 5.065819337267363e-05,
|
|
"loss": 0.1139,
|
|
"step": 4040
|
|
},
|
|
{
|
|
"epoch": 1.84,
|
|
"learning_rate": 5.058556513844758e-05,
|
|
"loss": 0.0967,
|
|
"step": 4050
|
|
},
|
|
{
|
|
"epoch": 1.84,
|
|
"learning_rate": 5.051293690422152e-05,
|
|
"loss": 0.1091,
|
|
"step": 4060
|
|
},
|
|
{
|
|
"epoch": 1.85,
|
|
"learning_rate": 5.0440308669995466e-05,
|
|
"loss": 0.1095,
|
|
"step": 4070
|
|
},
|
|
{
|
|
"epoch": 1.85,
|
|
"learning_rate": 5.036768043576941e-05,
|
|
"loss": 0.1095,
|
|
"step": 4080
|
|
},
|
|
{
|
|
"epoch": 1.86,
|
|
"learning_rate": 5.029505220154335e-05,
|
|
"loss": 0.1035,
|
|
"step": 4090
|
|
},
|
|
{
|
|
"epoch": 1.86,
|
|
"learning_rate": 5.02224239673173e-05,
|
|
"loss": 0.1135,
|
|
"step": 4100
|
|
},
|
|
{
|
|
"epoch": 1.86,
|
|
"eval_loss": 0.13894569873809814,
|
|
"eval_mean_accuracy": 0.7828193365629676,
|
|
"eval_mean_iou": 0.6788571097023736,
|
|
"eval_overall_accuracy": 0.9589751597200886,
|
|
"eval_per_category_accuracy": [
|
|
0.9924508452091282,
|
|
0.7176120079664502,
|
|
0.8960138797182788,
|
|
0.7147314161692893,
|
|
0.8934531679580748,
|
|
0.6898369980436232,
|
|
0.9350029752894382,
|
|
0.7650439944146693,
|
|
0.3213140196585618,
|
|
0.7172063492063492,
|
|
0.7081488913809046,
|
|
0.9127634590978638,
|
|
0.8908422207434851,
|
|
0.8344860355841824,
|
|
0.8294641317468529,
|
|
0.8032695272353546,
|
|
0.838746208416005,
|
|
0.6303619302949062
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9845350675889538,
|
|
0.6610336341263331,
|
|
0.8014127569522836,
|
|
0.610014471780029,
|
|
0.788742171126752,
|
|
0.6212098013755968,
|
|
0.782461315794455,
|
|
0.5891649614037918,
|
|
0.2686186316062624,
|
|
0.541830552501273,
|
|
0.55696192738273,
|
|
0.839407061595349,
|
|
0.7836525150889516,
|
|
0.7900103295458332,
|
|
0.7043775912722746,
|
|
0.69891685855447,
|
|
0.7882008380333764,
|
|
0.4088774889140075
|
|
],
|
|
"eval_runtime": 81.5041,
|
|
"eval_samples_per_second": 1.092,
|
|
"eval_steps_per_second": 0.147,
|
|
"step": 4100
|
|
},
|
|
{
|
|
"epoch": 1.87,
|
|
"learning_rate": 5.0149795733091246e-05,
|
|
"loss": 0.1147,
|
|
"step": 4110
|
|
},
|
|
{
|
|
"epoch": 1.87,
|
|
"learning_rate": 5.007716749886519e-05,
|
|
"loss": 0.1361,
|
|
"step": 4120
|
|
},
|
|
{
|
|
"epoch": 1.87,
|
|
"learning_rate": 5.000453926463913e-05,
|
|
"loss": 0.1051,
|
|
"step": 4130
|
|
},
|
|
{
|
|
"epoch": 1.88,
|
|
"learning_rate": 4.9931911030413076e-05,
|
|
"loss": 0.127,
|
|
"step": 4140
|
|
},
|
|
{
|
|
"epoch": 1.88,
|
|
"learning_rate": 4.9859282796187026e-05,
|
|
"loss": 0.1275,
|
|
"step": 4150
|
|
},
|
|
{
|
|
"epoch": 1.89,
|
|
"learning_rate": 4.978665456196097e-05,
|
|
"loss": 0.1328,
|
|
"step": 4160
|
|
},
|
|
{
|
|
"epoch": 1.89,
|
|
"learning_rate": 4.971402632773491e-05,
|
|
"loss": 0.1083,
|
|
"step": 4170
|
|
},
|
|
{
|
|
"epoch": 1.9,
|
|
"learning_rate": 4.9641398093508856e-05,
|
|
"loss": 0.152,
|
|
"step": 4180
|
|
},
|
|
{
|
|
"epoch": 1.9,
|
|
"learning_rate": 4.95687698592828e-05,
|
|
"loss": 0.099,
|
|
"step": 4190
|
|
},
|
|
{
|
|
"epoch": 1.91,
|
|
"learning_rate": 4.949614162505674e-05,
|
|
"loss": 0.1281,
|
|
"step": 4200
|
|
},
|
|
{
|
|
"epoch": 1.91,
|
|
"eval_loss": 0.1334676295518875,
|
|
"eval_mean_accuracy": 0.7880120308161696,
|
|
"eval_mean_iou": 0.6755332328718038,
|
|
"eval_overall_accuracy": 0.9584213428283006,
|
|
"eval_per_category_accuracy": [
|
|
0.9914689198032866,
|
|
0.7314561440437831,
|
|
0.9120068504456916,
|
|
0.7297205642973413,
|
|
0.8634865069405325,
|
|
0.680596098728576,
|
|
0.900973319831435,
|
|
0.8149055352939795,
|
|
0.3026901189860321,
|
|
0.7354725274725274,
|
|
0.7353676051073206,
|
|
0.9191090130587942,
|
|
0.8866922648058558,
|
|
0.8776507752142618,
|
|
0.7936497672012416,
|
|
0.7969874100719424,
|
|
0.9242485489380479,
|
|
0.5877345844504022
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9841648128511729,
|
|
0.6392155198030309,
|
|
0.8042780067509167,
|
|
0.6069957686882934,
|
|
0.7780489101673204,
|
|
0.6120132319886901,
|
|
0.7829185799149376,
|
|
0.6028501099793939,
|
|
0.2566340628974955,
|
|
0.5437410633043026,
|
|
0.5742686823992134,
|
|
0.8401808496028104,
|
|
0.798286128325838,
|
|
0.7950475674253807,
|
|
0.691800383300139,
|
|
0.6887966113926932,
|
|
0.7933927204435226,
|
|
0.366965182457315
|
|
],
|
|
"eval_runtime": 80.5596,
|
|
"eval_samples_per_second": 1.105,
|
|
"eval_steps_per_second": 0.149,
|
|
"step": 4200
|
|
},
|
|
{
|
|
"epoch": 1.91,
|
|
"learning_rate": 4.942351339083069e-05,
|
|
"loss": 0.1076,
|
|
"step": 4210
|
|
},
|
|
{
|
|
"epoch": 1.92,
|
|
"learning_rate": 4.9350885156604636e-05,
|
|
"loss": 0.1024,
|
|
"step": 4220
|
|
},
|
|
{
|
|
"epoch": 1.92,
|
|
"learning_rate": 4.927825692237858e-05,
|
|
"loss": 0.1067,
|
|
"step": 4230
|
|
},
|
|
{
|
|
"epoch": 1.92,
|
|
"learning_rate": 4.920562868815252e-05,
|
|
"loss": 0.1173,
|
|
"step": 4240
|
|
},
|
|
{
|
|
"epoch": 1.93,
|
|
"learning_rate": 4.9133000453926466e-05,
|
|
"loss": 0.1102,
|
|
"step": 4250
|
|
},
|
|
{
|
|
"epoch": 1.93,
|
|
"learning_rate": 4.9060372219700416e-05,
|
|
"loss": 0.0959,
|
|
"step": 4260
|
|
},
|
|
{
|
|
"epoch": 1.94,
|
|
"learning_rate": 4.898774398547436e-05,
|
|
"loss": 0.1285,
|
|
"step": 4270
|
|
},
|
|
{
|
|
"epoch": 1.94,
|
|
"learning_rate": 4.89151157512483e-05,
|
|
"loss": 0.1278,
|
|
"step": 4280
|
|
},
|
|
{
|
|
"epoch": 1.95,
|
|
"learning_rate": 4.8842487517022246e-05,
|
|
"loss": 0.1186,
|
|
"step": 4290
|
|
},
|
|
{
|
|
"epoch": 1.95,
|
|
"learning_rate": 4.876985928279619e-05,
|
|
"loss": 0.1245,
|
|
"step": 4300
|
|
},
|
|
{
|
|
"epoch": 1.95,
|
|
"eval_loss": 0.13443762063980103,
|
|
"eval_mean_accuracy": 0.7825141928725435,
|
|
"eval_mean_iou": 0.6684080027410144,
|
|
"eval_overall_accuracy": 0.9573094228680215,
|
|
"eval_per_category_accuracy": [
|
|
0.9919866189227488,
|
|
0.7269709677941677,
|
|
0.9073831435017057,
|
|
0.7095767769940314,
|
|
0.8775246954790309,
|
|
0.6467724199453274,
|
|
0.8928182428740948,
|
|
0.7752861759451438,
|
|
0.3290222452146922,
|
|
0.645997557997558,
|
|
0.7973597749316336,
|
|
0.9167328897375985,
|
|
0.8575661385177059,
|
|
0.886784400682822,
|
|
0.8199204604242111,
|
|
0.7835881294964029,
|
|
0.9045492836490544,
|
|
0.6154155495978553
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9841008078941216,
|
|
0.6601237998617911,
|
|
0.8036749084318817,
|
|
0.6052296656253615,
|
|
0.7720933405454484,
|
|
0.5803545423384261,
|
|
0.7963377956006369,
|
|
0.5713431842385517,
|
|
0.2788862091646569,
|
|
0.5162121235774388,
|
|
0.5750017734269702,
|
|
0.8372946231899758,
|
|
0.7948864346738904,
|
|
0.8019631403880126,
|
|
0.7034095374586811,
|
|
0.6923875903874315,
|
|
0.7961282648075675,
|
|
0.26191630772741537
|
|
],
|
|
"eval_runtime": 80.2554,
|
|
"eval_samples_per_second": 1.109,
|
|
"eval_steps_per_second": 0.15,
|
|
"step": 4300
|
|
},
|
|
{
|
|
"epoch": 1.96,
|
|
"learning_rate": 4.869723104857014e-05,
|
|
"loss": 0.1116,
|
|
"step": 4310
|
|
},
|
|
{
|
|
"epoch": 1.96,
|
|
"learning_rate": 4.862460281434408e-05,
|
|
"loss": 0.1118,
|
|
"step": 4320
|
|
},
|
|
{
|
|
"epoch": 1.97,
|
|
"learning_rate": 4.8551974580118026e-05,
|
|
"loss": 0.1187,
|
|
"step": 4330
|
|
},
|
|
{
|
|
"epoch": 1.97,
|
|
"learning_rate": 4.847934634589197e-05,
|
|
"loss": 0.1381,
|
|
"step": 4340
|
|
},
|
|
{
|
|
"epoch": 1.97,
|
|
"learning_rate": 4.840671811166591e-05,
|
|
"loss": 0.1106,
|
|
"step": 4350
|
|
},
|
|
{
|
|
"epoch": 1.98,
|
|
"learning_rate": 4.833408987743986e-05,
|
|
"loss": 0.0977,
|
|
"step": 4360
|
|
},
|
|
{
|
|
"epoch": 1.98,
|
|
"learning_rate": 4.8261461643213807e-05,
|
|
"loss": 0.1118,
|
|
"step": 4370
|
|
},
|
|
{
|
|
"epoch": 1.99,
|
|
"learning_rate": 4.818883340898775e-05,
|
|
"loss": 0.143,
|
|
"step": 4380
|
|
},
|
|
{
|
|
"epoch": 1.99,
|
|
"learning_rate": 4.811620517476169e-05,
|
|
"loss": 0.1297,
|
|
"step": 4390
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"learning_rate": 4.8043576940535637e-05,
|
|
"loss": 0.1279,
|
|
"step": 4400
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"eval_loss": 0.12572704255580902,
|
|
"eval_mean_accuracy": 0.7862403936198279,
|
|
"eval_mean_iou": 0.6747473672019839,
|
|
"eval_overall_accuracy": 0.9576788912998156,
|
|
"eval_per_category_accuracy": [
|
|
0.9915765311063784,
|
|
0.7196683884130248,
|
|
0.9192923269505887,
|
|
0.7331795984807379,
|
|
0.8581648945111781,
|
|
0.7115117095553328,
|
|
0.9112928763567146,
|
|
0.7519079901660606,
|
|
0.29948266942576307,
|
|
0.7491965811965812,
|
|
0.683487772728167,
|
|
0.9157444856938398,
|
|
0.8900823696563135,
|
|
0.8693926755011175,
|
|
0.8411364028280738,
|
|
0.8102389516957862,
|
|
0.9084990110252937,
|
|
0.5884718498659517
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9844717060862208,
|
|
0.6535599800017645,
|
|
0.8063888964304353,
|
|
0.6055343939054447,
|
|
0.7820480995831319,
|
|
0.5855220360295534,
|
|
0.8090355387973314,
|
|
0.5450858581824229,
|
|
0.24937537692771602,
|
|
0.5302421724311955,
|
|
0.5408211714341311,
|
|
0.8375502175037155,
|
|
0.8053053786011244,
|
|
0.800377494521874,
|
|
0.7112820466082772,
|
|
0.7012536483669215,
|
|
0.8107974135557,
|
|
0.3868011806687519
|
|
],
|
|
"eval_runtime": 80.6668,
|
|
"eval_samples_per_second": 1.103,
|
|
"eval_steps_per_second": 0.149,
|
|
"step": 4400
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"learning_rate": 4.797094870630959e-05,
|
|
"loss": 0.1082,
|
|
"step": 4410
|
|
},
|
|
{
|
|
"epoch": 2.01,
|
|
"learning_rate": 4.789832047208353e-05,
|
|
"loss": 0.0905,
|
|
"step": 4420
|
|
},
|
|
{
|
|
"epoch": 2.01,
|
|
"learning_rate": 4.782569223785747e-05,
|
|
"loss": 0.0983,
|
|
"step": 4430
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"learning_rate": 4.775306400363142e-05,
|
|
"loss": 0.1054,
|
|
"step": 4440
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"learning_rate": 4.768043576940536e-05,
|
|
"loss": 0.0985,
|
|
"step": 4450
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"learning_rate": 4.760780753517931e-05,
|
|
"loss": 0.1075,
|
|
"step": 4460
|
|
},
|
|
{
|
|
"epoch": 2.03,
|
|
"learning_rate": 4.7535179300953253e-05,
|
|
"loss": 0.0946,
|
|
"step": 4470
|
|
},
|
|
{
|
|
"epoch": 2.03,
|
|
"learning_rate": 4.74625510667272e-05,
|
|
"loss": 0.0959,
|
|
"step": 4480
|
|
},
|
|
{
|
|
"epoch": 2.04,
|
|
"learning_rate": 4.738992283250114e-05,
|
|
"loss": 0.1003,
|
|
"step": 4490
|
|
},
|
|
{
|
|
"epoch": 2.04,
|
|
"learning_rate": 4.7317294598275083e-05,
|
|
"loss": 0.1002,
|
|
"step": 4500
|
|
},
|
|
{
|
|
"epoch": 2.04,
|
|
"eval_loss": 0.13972726464271545,
|
|
"eval_mean_accuracy": 0.7788409927090669,
|
|
"eval_mean_iou": 0.6773702141156384,
|
|
"eval_overall_accuracy": 0.9584378017468399,
|
|
"eval_per_category_accuracy": [
|
|
0.9922478814163378,
|
|
0.673666995903431,
|
|
0.9096373940794542,
|
|
0.7090341833966359,
|
|
0.8820256982557928,
|
|
0.7314044842098755,
|
|
0.8473119738870504,
|
|
0.7690487415845215,
|
|
0.28660113812726334,
|
|
0.7365860805860805,
|
|
0.7050010820594542,
|
|
0.9094859112887602,
|
|
0.8745706166933439,
|
|
0.9047216796015698,
|
|
0.8286342472840145,
|
|
0.810129753340185,
|
|
0.8998342965936688,
|
|
0.5491957104557641
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850577942021761,
|
|
0.6230905169831666,
|
|
0.8079639313816529,
|
|
0.6100606909430439,
|
|
0.7888574153189829,
|
|
0.6199027954268561,
|
|
0.7932197150475545,
|
|
0.5528114514838938,
|
|
0.25061069392924995,
|
|
0.5259310922025388,
|
|
0.5354586614467373,
|
|
0.841570176882694,
|
|
0.7832103915347324,
|
|
0.7804022770398482,
|
|
0.710008680715882,
|
|
0.6975333222719983,
|
|
0.8216968638169574,
|
|
0.4652773834535234
|
|
],
|
|
"eval_runtime": 80.0929,
|
|
"eval_samples_per_second": 1.111,
|
|
"eval_steps_per_second": 0.15,
|
|
"step": 4500
|
|
},
|
|
{
|
|
"epoch": 2.05,
|
|
"learning_rate": 4.724466636404903e-05,
|
|
"loss": 0.1034,
|
|
"step": 4510
|
|
},
|
|
{
|
|
"epoch": 2.05,
|
|
"learning_rate": 4.717203812982298e-05,
|
|
"loss": 0.1153,
|
|
"step": 4520
|
|
},
|
|
{
|
|
"epoch": 2.06,
|
|
"learning_rate": 4.709940989559692e-05,
|
|
"loss": 0.0998,
|
|
"step": 4530
|
|
},
|
|
{
|
|
"epoch": 2.06,
|
|
"learning_rate": 4.7026781661370864e-05,
|
|
"loss": 0.1161,
|
|
"step": 4540
|
|
},
|
|
{
|
|
"epoch": 2.07,
|
|
"learning_rate": 4.695415342714481e-05,
|
|
"loss": 0.1006,
|
|
"step": 4550
|
|
},
|
|
{
|
|
"epoch": 2.07,
|
|
"learning_rate": 4.688152519291875e-05,
|
|
"loss": 0.1049,
|
|
"step": 4560
|
|
},
|
|
{
|
|
"epoch": 2.07,
|
|
"learning_rate": 4.68088969586927e-05,
|
|
"loss": 0.1089,
|
|
"step": 4570
|
|
},
|
|
{
|
|
"epoch": 2.08,
|
|
"learning_rate": 4.6736268724466644e-05,
|
|
"loss": 0.1044,
|
|
"step": 4580
|
|
},
|
|
{
|
|
"epoch": 2.08,
|
|
"learning_rate": 4.666364049024059e-05,
|
|
"loss": 0.098,
|
|
"step": 4590
|
|
},
|
|
{
|
|
"epoch": 2.09,
|
|
"learning_rate": 4.659101225601453e-05,
|
|
"loss": 0.0907,
|
|
"step": 4600
|
|
},
|
|
{
|
|
"epoch": 2.09,
|
|
"eval_loss": 0.13046495616436005,
|
|
"eval_mean_accuracy": 0.7811944164041577,
|
|
"eval_mean_iou": 0.6782751969723191,
|
|
"eval_overall_accuracy": 0.9586188927125395,
|
|
"eval_per_category_accuracy": [
|
|
0.9928277621186691,
|
|
0.7083987758869152,
|
|
0.9055965940354352,
|
|
0.6987927292457949,
|
|
0.8713186634491815,
|
|
0.7051015946899607,
|
|
0.878982277018586,
|
|
0.8087493880883609,
|
|
0.31200206932229696,
|
|
0.7381587301587301,
|
|
0.6864093332546382,
|
|
0.9234421763866321,
|
|
0.8785721993417621,
|
|
0.8796526054590571,
|
|
0.841713010863942,
|
|
0.7948805241521069,
|
|
0.8484292119368168,
|
|
0.5884718498659517
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9844003862902916,
|
|
0.6621263715474839,
|
|
0.8059651700180577,
|
|
0.6015999065747986,
|
|
0.785216312190489,
|
|
0.6208224621761497,
|
|
0.8085146904898324,
|
|
0.5663717485850763,
|
|
0.2662693156732892,
|
|
0.5292836331036028,
|
|
0.5412749387197865,
|
|
0.8419965464907948,
|
|
0.807288546628603,
|
|
0.8064357394082153,
|
|
0.7138417523959252,
|
|
0.7065443292376559,
|
|
0.7819161642589012,
|
|
0.3790855317127931
|
|
],
|
|
"eval_runtime": 80.5469,
|
|
"eval_samples_per_second": 1.105,
|
|
"eval_steps_per_second": 0.149,
|
|
"step": 4600
|
|
},
|
|
{
|
|
"epoch": 2.09,
|
|
"learning_rate": 4.6518384021788474e-05,
|
|
"loss": 0.1112,
|
|
"step": 4610
|
|
},
|
|
{
|
|
"epoch": 2.1,
|
|
"learning_rate": 4.6445755787562424e-05,
|
|
"loss": 0.0881,
|
|
"step": 4620
|
|
},
|
|
{
|
|
"epoch": 2.1,
|
|
"learning_rate": 4.637312755333637e-05,
|
|
"loss": 0.1018,
|
|
"step": 4630
|
|
},
|
|
{
|
|
"epoch": 2.11,
|
|
"learning_rate": 4.630049931911031e-05,
|
|
"loss": 0.0866,
|
|
"step": 4640
|
|
},
|
|
{
|
|
"epoch": 2.11,
|
|
"learning_rate": 4.6227871084884254e-05,
|
|
"loss": 0.1176,
|
|
"step": 4650
|
|
},
|
|
{
|
|
"epoch": 2.12,
|
|
"learning_rate": 4.61552428506582e-05,
|
|
"loss": 0.088,
|
|
"step": 4660
|
|
},
|
|
{
|
|
"epoch": 2.12,
|
|
"learning_rate": 4.608261461643215e-05,
|
|
"loss": 0.096,
|
|
"step": 4670
|
|
},
|
|
{
|
|
"epoch": 2.12,
|
|
"learning_rate": 4.600998638220609e-05,
|
|
"loss": 0.1117,
|
|
"step": 4680
|
|
},
|
|
{
|
|
"epoch": 2.13,
|
|
"learning_rate": 4.5937358147980034e-05,
|
|
"loss": 0.0836,
|
|
"step": 4690
|
|
},
|
|
{
|
|
"epoch": 2.13,
|
|
"learning_rate": 4.586472991375398e-05,
|
|
"loss": 0.1096,
|
|
"step": 4700
|
|
},
|
|
{
|
|
"epoch": 2.13,
|
|
"eval_loss": 0.13663238286972046,
|
|
"eval_mean_accuracy": 0.7862924352091032,
|
|
"eval_mean_iou": 0.6753792995078041,
|
|
"eval_overall_accuracy": 0.9587173890531733,
|
|
"eval_per_category_accuracy": [
|
|
0.9912568589725028,
|
|
0.7287844686604381,
|
|
0.9090837184989545,
|
|
0.7161557243624526,
|
|
0.8716486912442459,
|
|
0.733279161275222,
|
|
0.9053092387086896,
|
|
0.7746160085116683,
|
|
0.2463528194516296,
|
|
0.724014652014652,
|
|
0.7265930866237778,
|
|
0.9221967872914961,
|
|
0.8754833372300056,
|
|
0.8802377558383049,
|
|
0.8189827987584066,
|
|
0.7959789311408016,
|
|
0.9121771946444165,
|
|
0.621112600536193
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9840687506780187,
|
|
0.6647319450598139,
|
|
0.8056522779326183,
|
|
0.6075723574428908,
|
|
0.779166766332831,
|
|
0.6165828987116383,
|
|
0.818051751624903,
|
|
0.5792731359655048,
|
|
0.21812019054598755,
|
|
0.5545737501309351,
|
|
0.5726912550299665,
|
|
0.8376480896923469,
|
|
0.798322325129454,
|
|
0.790813220866898,
|
|
0.7083919865012259,
|
|
0.6933484031243705,
|
|
0.8084097453722061,
|
|
0.31940854099886257
|
|
],
|
|
"eval_runtime": 81.8556,
|
|
"eval_samples_per_second": 1.087,
|
|
"eval_steps_per_second": 0.147,
|
|
"step": 4700
|
|
},
|
|
{
|
|
"epoch": 2.14,
|
|
"learning_rate": 4.579210167952792e-05,
|
|
"loss": 0.1073,
|
|
"step": 4710
|
|
},
|
|
{
|
|
"epoch": 2.14,
|
|
"learning_rate": 4.571947344530187e-05,
|
|
"loss": 0.1135,
|
|
"step": 4720
|
|
},
|
|
{
|
|
"epoch": 2.15,
|
|
"learning_rate": 4.5646845211075814e-05,
|
|
"loss": 0.0965,
|
|
"step": 4730
|
|
},
|
|
{
|
|
"epoch": 2.15,
|
|
"learning_rate": 4.557421697684976e-05,
|
|
"loss": 0.0962,
|
|
"step": 4740
|
|
},
|
|
{
|
|
"epoch": 2.16,
|
|
"learning_rate": 4.55015887426237e-05,
|
|
"loss": 0.1166,
|
|
"step": 4750
|
|
},
|
|
{
|
|
"epoch": 2.16,
|
|
"learning_rate": 4.5428960508397644e-05,
|
|
"loss": 0.0883,
|
|
"step": 4760
|
|
},
|
|
{
|
|
"epoch": 2.17,
|
|
"learning_rate": 4.535633227417159e-05,
|
|
"loss": 0.1268,
|
|
"step": 4770
|
|
},
|
|
{
|
|
"epoch": 2.17,
|
|
"learning_rate": 4.528370403994554e-05,
|
|
"loss": 0.0955,
|
|
"step": 4780
|
|
},
|
|
{
|
|
"epoch": 2.17,
|
|
"learning_rate": 4.521107580571948e-05,
|
|
"loss": 0.0928,
|
|
"step": 4790
|
|
},
|
|
{
|
|
"epoch": 2.18,
|
|
"learning_rate": 4.5138447571493424e-05,
|
|
"loss": 0.0943,
|
|
"step": 4800
|
|
},
|
|
{
|
|
"epoch": 2.18,
|
|
"eval_loss": 0.1498226672410965,
|
|
"eval_mean_accuracy": 0.7786997550272359,
|
|
"eval_mean_iou": 0.664842613989284,
|
|
"eval_overall_accuracy": 0.9562002031990652,
|
|
"eval_per_category_accuracy": [
|
|
0.9923740750836234,
|
|
0.7165109538690717,
|
|
0.8956235556289205,
|
|
0.6829896907216495,
|
|
0.8612250274965986,
|
|
0.6496959446031824,
|
|
0.9242658342815981,
|
|
0.7623831948742127,
|
|
0.30144852560786345,
|
|
0.7222466422466423,
|
|
0.7190875287729446,
|
|
0.9306417114413699,
|
|
0.8555203855907055,
|
|
0.8759173222110765,
|
|
0.8311023452319365,
|
|
0.8024922918807811,
|
|
0.846287719125012,
|
|
0.6467828418230563
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9839855105743271,
|
|
0.6651285134525777,
|
|
0.8047618311882947,
|
|
0.5993334126889656,
|
|
0.7710710817091653,
|
|
0.5758863388149268,
|
|
0.7987690817885456,
|
|
0.5345619201418574,
|
|
0.2649118021458447,
|
|
0.5450390682588825,
|
|
0.5745082166911609,
|
|
0.8368648587152832,
|
|
0.7960356772315004,
|
|
0.7922615920920681,
|
|
0.7041768637907733,
|
|
0.6976128653991122,
|
|
0.7758146139724083,
|
|
0.2464438031514161
|
|
],
|
|
"eval_runtime": 79.8882,
|
|
"eval_samples_per_second": 1.114,
|
|
"eval_steps_per_second": 0.15,
|
|
"step": 4800
|
|
},
|
|
{
|
|
"epoch": 2.18,
|
|
"learning_rate": 4.506581933726737e-05,
|
|
"loss": 0.1111,
|
|
"step": 4810
|
|
},
|
|
{
|
|
"epoch": 2.19,
|
|
"learning_rate": 4.499319110304131e-05,
|
|
"loss": 0.1019,
|
|
"step": 4820
|
|
},
|
|
{
|
|
"epoch": 2.19,
|
|
"learning_rate": 4.492056286881526e-05,
|
|
"loss": 0.1044,
|
|
"step": 4830
|
|
},
|
|
{
|
|
"epoch": 2.2,
|
|
"learning_rate": 4.4847934634589204e-05,
|
|
"loss": 0.1078,
|
|
"step": 4840
|
|
},
|
|
{
|
|
"epoch": 2.2,
|
|
"learning_rate": 4.477530640036315e-05,
|
|
"loss": 0.1105,
|
|
"step": 4850
|
|
},
|
|
{
|
|
"epoch": 2.21,
|
|
"learning_rate": 4.470267816613709e-05,
|
|
"loss": 0.1199,
|
|
"step": 4860
|
|
},
|
|
{
|
|
"epoch": 2.21,
|
|
"learning_rate": 4.4630049931911034e-05,
|
|
"loss": 0.0992,
|
|
"step": 4870
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 4.4557421697684984e-05,
|
|
"loss": 0.1098,
|
|
"step": 4880
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 4.448479346345893e-05,
|
|
"loss": 0.0953,
|
|
"step": 4890
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 4.441216522923287e-05,
|
|
"loss": 0.0964,
|
|
"step": 4900
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"eval_loss": 0.1363184005022049,
|
|
"eval_mean_accuracy": 0.779909349604894,
|
|
"eval_mean_iou": 0.6639330164541964,
|
|
"eval_overall_accuracy": 0.955620283491156,
|
|
"eval_per_category_accuracy": [
|
|
0.9925093657889024,
|
|
0.7195226606648424,
|
|
0.9051151370089139,
|
|
0.6953336950623983,
|
|
0.8525475607103625,
|
|
0.6841136023388197,
|
|
0.9184005122369653,
|
|
0.6965911779953865,
|
|
0.2818416968442835,
|
|
0.7013137973137973,
|
|
0.7799976391430089,
|
|
0.926304594497357,
|
|
0.8919932377749402,
|
|
0.8673468489872059,
|
|
0.845878599758579,
|
|
0.8095002569373073,
|
|
0.8552455773853422,
|
|
0.6148123324396783
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9848753150184794,
|
|
0.6500724139444387,
|
|
0.8076650433149517,
|
|
0.6053020015351007,
|
|
0.7691890553423301,
|
|
0.562433739489086,
|
|
0.8069825680415114,
|
|
0.486605441610325,
|
|
0.23783123062819225,
|
|
0.5461301487087818,
|
|
0.5789838775629418,
|
|
0.8411738729261734,
|
|
0.8066273632852206,
|
|
0.7983356280877946,
|
|
0.7129400009083889,
|
|
0.6961135231277411,
|
|
0.7901193038470858,
|
|
0.26941376879699247
|
|
],
|
|
"eval_runtime": 78.9031,
|
|
"eval_samples_per_second": 1.128,
|
|
"eval_steps_per_second": 0.152,
|
|
"step": 4900
|
|
},
|
|
{
|
|
"epoch": 2.23,
|
|
"learning_rate": 4.4339536995006814e-05,
|
|
"loss": 0.0879,
|
|
"step": 4910
|
|
},
|
|
{
|
|
"epoch": 2.23,
|
|
"learning_rate": 4.426690876078076e-05,
|
|
"loss": 0.0961,
|
|
"step": 4920
|
|
},
|
|
{
|
|
"epoch": 2.24,
|
|
"learning_rate": 4.419428052655471e-05,
|
|
"loss": 0.1006,
|
|
"step": 4930
|
|
},
|
|
{
|
|
"epoch": 2.24,
|
|
"learning_rate": 4.412165229232865e-05,
|
|
"loss": 0.1023,
|
|
"step": 4940
|
|
},
|
|
{
|
|
"epoch": 2.25,
|
|
"learning_rate": 4.4049024058102595e-05,
|
|
"loss": 0.1079,
|
|
"step": 4950
|
|
},
|
|
{
|
|
"epoch": 2.25,
|
|
"learning_rate": 4.397639582387654e-05,
|
|
"loss": 0.0982,
|
|
"step": 4960
|
|
},
|
|
{
|
|
"epoch": 2.26,
|
|
"learning_rate": 4.390376758965048e-05,
|
|
"loss": 0.1068,
|
|
"step": 4970
|
|
},
|
|
{
|
|
"epoch": 2.26,
|
|
"learning_rate": 4.383113935542443e-05,
|
|
"loss": 0.0893,
|
|
"step": 4980
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"learning_rate": 4.3758511121198375e-05,
|
|
"loss": 0.1011,
|
|
"step": 4990
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"learning_rate": 4.368588288697232e-05,
|
|
"loss": 0.1051,
|
|
"step": 5000
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"eval_loss": 0.13948704302310944,
|
|
"eval_mean_accuracy": 0.7862352771041999,
|
|
"eval_mean_iou": 0.6728738609815013,
|
|
"eval_overall_accuracy": 0.958048488316911,
|
|
"eval_per_category_accuracy": [
|
|
0.9922981370042766,
|
|
0.7038812156932592,
|
|
0.9110095466050402,
|
|
0.7093733043950081,
|
|
0.8633389893190381,
|
|
0.6356612597653231,
|
|
0.922238809605835,
|
|
0.8056749272479963,
|
|
0.34345576823590274,
|
|
0.6931965811965812,
|
|
0.7369119990556572,
|
|
0.9255296857270503,
|
|
0.8810495836555582,
|
|
0.8887510339123242,
|
|
0.8609135195723401,
|
|
0.7904098150051387,
|
|
0.8652834392251129,
|
|
0.6232573726541555
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9845730934839846,
|
|
0.6543289782647962,
|
|
0.8066816691078623,
|
|
0.6087892898719441,
|
|
0.776111375309405,
|
|
0.5984657762070579,
|
|
0.8045457788220866,
|
|
0.5666638292141569,
|
|
0.2786217894913547,
|
|
0.5356773526370218,
|
|
0.5755764369626516,
|
|
0.8442773419890001,
|
|
0.8032563783039286,
|
|
0.796446848608625,
|
|
0.7188146788330364,
|
|
0.704896715282472,
|
|
0.7796387273996102,
|
|
0.2743634378780279
|
|
],
|
|
"eval_runtime": 75.4373,
|
|
"eval_samples_per_second": 1.18,
|
|
"eval_steps_per_second": 0.159,
|
|
"step": 5000
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"learning_rate": 4.361325465274626e-05,
|
|
"loss": 0.0912,
|
|
"step": 5010
|
|
},
|
|
{
|
|
"epoch": 2.28,
|
|
"learning_rate": 4.3540626418520205e-05,
|
|
"loss": 0.1056,
|
|
"step": 5020
|
|
},
|
|
{
|
|
"epoch": 2.28,
|
|
"learning_rate": 4.346799818429415e-05,
|
|
"loss": 0.1234,
|
|
"step": 5030
|
|
},
|
|
{
|
|
"epoch": 2.29,
|
|
"learning_rate": 4.33953699500681e-05,
|
|
"loss": 0.1045,
|
|
"step": 5040
|
|
},
|
|
{
|
|
"epoch": 2.29,
|
|
"learning_rate": 4.332274171584204e-05,
|
|
"loss": 0.1124,
|
|
"step": 5050
|
|
},
|
|
{
|
|
"epoch": 2.3,
|
|
"learning_rate": 4.3250113481615985e-05,
|
|
"loss": 0.0884,
|
|
"step": 5060
|
|
},
|
|
{
|
|
"epoch": 2.3,
|
|
"learning_rate": 4.317748524738993e-05,
|
|
"loss": 0.0937,
|
|
"step": 5070
|
|
},
|
|
{
|
|
"epoch": 2.31,
|
|
"learning_rate": 4.310485701316387e-05,
|
|
"loss": 0.0971,
|
|
"step": 5080
|
|
},
|
|
{
|
|
"epoch": 2.31,
|
|
"learning_rate": 4.303222877893782e-05,
|
|
"loss": 0.0995,
|
|
"step": 5090
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"learning_rate": 4.2959600544711765e-05,
|
|
"loss": 0.0963,
|
|
"step": 5100
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"eval_loss": 0.13856534659862518,
|
|
"eval_mean_accuracy": 0.7810347220217441,
|
|
"eval_mean_iou": 0.6805732655966783,
|
|
"eval_overall_accuracy": 0.9593023235878247,
|
|
"eval_per_category_accuracy": [
|
|
0.9915753662417573,
|
|
0.7019867549668874,
|
|
0.9124057719819523,
|
|
0.6568095496473142,
|
|
0.8796743359820179,
|
|
0.7558415290652228,
|
|
0.8941475534769128,
|
|
0.7791572509045454,
|
|
0.29218830832902226,
|
|
0.7648644688644689,
|
|
0.6649747191563871,
|
|
0.9269015905397873,
|
|
0.8923484344369909,
|
|
0.8519305562888267,
|
|
0.7913433350577685,
|
|
0.8191932168550874,
|
|
0.8974243457492602,
|
|
0.585857908847185
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9842123272869723,
|
|
0.6475578790141897,
|
|
0.8065977864490307,
|
|
0.5950961715725435,
|
|
0.7896473057668808,
|
|
0.636168832737687,
|
|
0.8243746170060541,
|
|
0.5740744437745472,
|
|
0.25528837461580184,
|
|
0.5439937196490229,
|
|
0.5443798418399395,
|
|
0.8415165992455051,
|
|
0.8064600018691664,
|
|
0.7964307307583155,
|
|
0.6916026939198417,
|
|
0.693300860564613,
|
|
0.7965899286487079,
|
|
0.4230266660213909
|
|
],
|
|
"eval_runtime": 75.6227,
|
|
"eval_samples_per_second": 1.177,
|
|
"eval_steps_per_second": 0.159,
|
|
"step": 5100
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"learning_rate": 4.288697231048571e-05,
|
|
"loss": 0.1163,
|
|
"step": 5110
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"learning_rate": 4.281434407625965e-05,
|
|
"loss": 0.101,
|
|
"step": 5120
|
|
},
|
|
{
|
|
"epoch": 2.33,
|
|
"learning_rate": 4.2741715842033595e-05,
|
|
"loss": 0.1173,
|
|
"step": 5130
|
|
},
|
|
{
|
|
"epoch": 2.33,
|
|
"learning_rate": 4.2669087607807545e-05,
|
|
"loss": 0.1094,
|
|
"step": 5140
|
|
},
|
|
{
|
|
"epoch": 2.34,
|
|
"learning_rate": 4.259645937358149e-05,
|
|
"loss": 0.0956,
|
|
"step": 5150
|
|
},
|
|
{
|
|
"epoch": 2.34,
|
|
"learning_rate": 4.252383113935543e-05,
|
|
"loss": 0.1393,
|
|
"step": 5160
|
|
},
|
|
{
|
|
"epoch": 2.35,
|
|
"learning_rate": 4.2451202905129375e-05,
|
|
"loss": 0.1037,
|
|
"step": 5170
|
|
},
|
|
{
|
|
"epoch": 2.35,
|
|
"learning_rate": 4.237857467090332e-05,
|
|
"loss": 0.1083,
|
|
"step": 5180
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"learning_rate": 4.230594643667727e-05,
|
|
"loss": 0.0906,
|
|
"step": 5190
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"learning_rate": 4.223331820245121e-05,
|
|
"loss": 0.1062,
|
|
"step": 5200
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"eval_loss": 0.14069417119026184,
|
|
"eval_mean_accuracy": 0.793414086796973,
|
|
"eval_mean_iou": 0.6864809607686351,
|
|
"eval_overall_accuracy": 0.9608978957272647,
|
|
"eval_per_category_accuracy": [
|
|
0.9927017903303589,
|
|
0.7209151702585858,
|
|
0.9042227220204688,
|
|
0.7253798155181769,
|
|
0.8823262225265583,
|
|
0.7159411585460101,
|
|
0.9293168665852377,
|
|
0.7939551258776755,
|
|
0.3227625452664252,
|
|
0.7244737484737485,
|
|
0.7663440162112181,
|
|
0.9198048495056003,
|
|
0.872911533550348,
|
|
0.8701362124491843,
|
|
0.8376929211933092,
|
|
0.7936729188078109,
|
|
0.8828905832945046,
|
|
0.6260053619302949
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9851730866254734,
|
|
0.6629984811031361,
|
|
0.8096963332727704,
|
|
0.6109333942648235,
|
|
0.7896189845195383,
|
|
0.6560161214939089,
|
|
0.8236946973622523,
|
|
0.5911623165092575,
|
|
0.2780055253542465,
|
|
0.5391958008912928,
|
|
0.5810987207697759,
|
|
0.8444977639679405,
|
|
0.8053277583510663,
|
|
0.8013833511621122,
|
|
0.7164935148738466,
|
|
0.7034907223421033,
|
|
0.8141598208467665,
|
|
0.3437109001251196
|
|
],
|
|
"eval_runtime": 73.8903,
|
|
"eval_samples_per_second": 1.204,
|
|
"eval_steps_per_second": 0.162,
|
|
"step": 5200
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"learning_rate": 4.2160689968225155e-05,
|
|
"loss": 0.0955,
|
|
"step": 5210
|
|
},
|
|
{
|
|
"epoch": 2.37,
|
|
"learning_rate": 4.20880617339991e-05,
|
|
"loss": 0.0926,
|
|
"step": 5220
|
|
},
|
|
{
|
|
"epoch": 2.37,
|
|
"learning_rate": 4.201543349977304e-05,
|
|
"loss": 0.0952,
|
|
"step": 5230
|
|
},
|
|
{
|
|
"epoch": 2.38,
|
|
"learning_rate": 4.194280526554699e-05,
|
|
"loss": 0.0929,
|
|
"step": 5240
|
|
},
|
|
{
|
|
"epoch": 2.38,
|
|
"learning_rate": 4.1870177031320935e-05,
|
|
"loss": 0.0853,
|
|
"step": 5250
|
|
},
|
|
{
|
|
"epoch": 2.39,
|
|
"learning_rate": 4.179754879709488e-05,
|
|
"loss": 0.1079,
|
|
"step": 5260
|
|
},
|
|
{
|
|
"epoch": 2.39,
|
|
"learning_rate": 4.172492056286882e-05,
|
|
"loss": 0.1007,
|
|
"step": 5270
|
|
},
|
|
{
|
|
"epoch": 2.4,
|
|
"learning_rate": 4.1652292328642765e-05,
|
|
"loss": 0.093,
|
|
"step": 5280
|
|
},
|
|
{
|
|
"epoch": 2.4,
|
|
"learning_rate": 4.157966409441671e-05,
|
|
"loss": 0.0977,
|
|
"step": 5290
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"learning_rate": 4.150703586019066e-05,
|
|
"loss": 0.0995,
|
|
"step": 5300
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"eval_loss": 0.14190153777599335,
|
|
"eval_mean_accuracy": 0.7876905051576213,
|
|
"eval_mean_iou": 0.6820334465496983,
|
|
"eval_overall_accuracy": 0.9599077889088834,
|
|
"eval_per_category_accuracy": [
|
|
0.9924348144531521,
|
|
0.7192312051684775,
|
|
0.9121719214262133,
|
|
0.7101871947911015,
|
|
0.8761002927710191,
|
|
0.7157976320333511,
|
|
0.9125560694164605,
|
|
0.7983265684356557,
|
|
0.30067252974650804,
|
|
0.6966837606837607,
|
|
0.7734167502803517,
|
|
0.9045399374537921,
|
|
0.8795074006798195,
|
|
0.8910300406525526,
|
|
0.8141328246249353,
|
|
0.7826952723535457,
|
|
0.8835293282686308,
|
|
0.6154155495978553
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850695155037767,
|
|
0.6698890028352537,
|
|
0.8080701914728328,
|
|
0.6105183371231998,
|
|
0.7831849613766758,
|
|
0.6462454572095555,
|
|
0.8141802421069952,
|
|
0.5798586920155084,
|
|
0.26502507979936163,
|
|
0.5442675742498703,
|
|
0.5795879252515572,
|
|
0.8418875756193056,
|
|
0.8090034946959201,
|
|
0.8123608620834888,
|
|
0.7139204975049146,
|
|
0.6994793370876172,
|
|
0.7976399076243352,
|
|
0.3164133843344016
|
|
],
|
|
"eval_runtime": 75.1424,
|
|
"eval_samples_per_second": 1.184,
|
|
"eval_steps_per_second": 0.16,
|
|
"step": 5300
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"learning_rate": 4.14344076259646e-05,
|
|
"loss": 0.0978,
|
|
"step": 5310
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"learning_rate": 4.1361779391738545e-05,
|
|
"loss": 0.0935,
|
|
"step": 5320
|
|
},
|
|
{
|
|
"epoch": 2.42,
|
|
"learning_rate": 4.128915115751249e-05,
|
|
"loss": 0.1003,
|
|
"step": 5330
|
|
},
|
|
{
|
|
"epoch": 2.42,
|
|
"learning_rate": 4.121652292328643e-05,
|
|
"loss": 0.0929,
|
|
"step": 5340
|
|
},
|
|
{
|
|
"epoch": 2.43,
|
|
"learning_rate": 4.114389468906038e-05,
|
|
"loss": 0.0869,
|
|
"step": 5350
|
|
},
|
|
{
|
|
"epoch": 2.43,
|
|
"learning_rate": 4.1071266454834326e-05,
|
|
"loss": 0.1052,
|
|
"step": 5360
|
|
},
|
|
{
|
|
"epoch": 2.44,
|
|
"learning_rate": 4.099863822060827e-05,
|
|
"loss": 0.0841,
|
|
"step": 5370
|
|
},
|
|
{
|
|
"epoch": 2.44,
|
|
"learning_rate": 4.092600998638221e-05,
|
|
"loss": 0.1085,
|
|
"step": 5380
|
|
},
|
|
{
|
|
"epoch": 2.45,
|
|
"learning_rate": 4.0853381752156156e-05,
|
|
"loss": 0.1065,
|
|
"step": 5390
|
|
},
|
|
{
|
|
"epoch": 2.45,
|
|
"learning_rate": 4.0780753517930106e-05,
|
|
"loss": 0.1133,
|
|
"step": 5400
|
|
},
|
|
{
|
|
"epoch": 2.45,
|
|
"eval_loss": 0.1448107659816742,
|
|
"eval_mean_accuracy": 0.7860606643190547,
|
|
"eval_mean_iou": 0.6723266330890474,
|
|
"eval_overall_accuracy": 0.9580608753675826,
|
|
"eval_per_category_accuracy": [
|
|
0.9918118337598402,
|
|
0.7352774494405674,
|
|
0.9131021651810278,
|
|
0.7190043407487792,
|
|
0.8650838140235054,
|
|
0.7536422612712474,
|
|
0.8617081989233281,
|
|
0.7797081971773487,
|
|
0.29979306777030523,
|
|
0.7275604395604396,
|
|
0.7188120954573177,
|
|
0.9322508332246089,
|
|
0.8405931334640218,
|
|
0.8651514351582986,
|
|
0.8220221158820487,
|
|
0.8033915724563206,
|
|
0.9154202911072505,
|
|
0.6047587131367292
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9847888440066707,
|
|
0.668482261151185,
|
|
0.8071594358421708,
|
|
0.607855504587156,
|
|
0.7777053350472859,
|
|
0.6382738398013226,
|
|
0.7800267121108553,
|
|
0.5656370150411607,
|
|
0.2649142857142857,
|
|
0.5581039869922598,
|
|
0.5782921810699588,
|
|
0.8393263946236865,
|
|
0.7634740564689355,
|
|
0.7677998992624292,
|
|
0.7079388505235019,
|
|
0.6958418177165048,
|
|
0.8128287887422735,
|
|
0.28343018690120936
|
|
],
|
|
"eval_runtime": 76.8852,
|
|
"eval_samples_per_second": 1.158,
|
|
"eval_steps_per_second": 0.156,
|
|
"step": 5400
|
|
},
|
|
{
|
|
"epoch": 2.46,
|
|
"learning_rate": 4.070812528370405e-05,
|
|
"loss": 0.0992,
|
|
"step": 5410
|
|
},
|
|
{
|
|
"epoch": 2.46,
|
|
"learning_rate": 4.063549704947799e-05,
|
|
"loss": 0.1028,
|
|
"step": 5420
|
|
},
|
|
{
|
|
"epoch": 2.46,
|
|
"learning_rate": 4.0562868815251936e-05,
|
|
"loss": 0.1002,
|
|
"step": 5430
|
|
},
|
|
{
|
|
"epoch": 2.47,
|
|
"learning_rate": 4.049024058102588e-05,
|
|
"loss": 0.0933,
|
|
"step": 5440
|
|
},
|
|
{
|
|
"epoch": 2.47,
|
|
"learning_rate": 4.041761234679983e-05,
|
|
"loss": 0.0934,
|
|
"step": 5450
|
|
},
|
|
{
|
|
"epoch": 2.48,
|
|
"learning_rate": 4.034498411257377e-05,
|
|
"loss": 0.0845,
|
|
"step": 5460
|
|
},
|
|
{
|
|
"epoch": 2.48,
|
|
"learning_rate": 4.0272355878347716e-05,
|
|
"loss": 0.0949,
|
|
"step": 5470
|
|
},
|
|
{
|
|
"epoch": 2.49,
|
|
"learning_rate": 4.019972764412166e-05,
|
|
"loss": 0.1047,
|
|
"step": 5480
|
|
},
|
|
{
|
|
"epoch": 2.49,
|
|
"learning_rate": 4.01270994098956e-05,
|
|
"loss": 0.083,
|
|
"step": 5490
|
|
},
|
|
{
|
|
"epoch": 2.5,
|
|
"learning_rate": 4.005447117566955e-05,
|
|
"loss": 0.0874,
|
|
"step": 5500
|
|
},
|
|
{
|
|
"epoch": 2.5,
|
|
"eval_loss": 0.1436825543642044,
|
|
"eval_mean_accuracy": 0.7867991040853224,
|
|
"eval_mean_iou": 0.6832179193851488,
|
|
"eval_overall_accuracy": 0.9600632056761323,
|
|
"eval_per_category_accuracy": [
|
|
0.9926575809445011,
|
|
0.7220648002720251,
|
|
0.9126533784527346,
|
|
0.7089663591969615,
|
|
0.8688733015747334,
|
|
0.7494203736988769,
|
|
0.8806473951427964,
|
|
0.7973583481004999,
|
|
0.25659596482152097,
|
|
0.6673699633699633,
|
|
0.7635503354384309,
|
|
0.9160489141393175,
|
|
0.884889304533928,
|
|
0.8988437780476216,
|
|
0.8554222710812209,
|
|
0.8011626413155191,
|
|
0.8899476352550197,
|
|
0.595911528150134
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.984880249300407,
|
|
0.67195057635802,
|
|
0.8092375386688906,
|
|
0.613007271874267,
|
|
0.7855266414530675,
|
|
0.6435047723058371,
|
|
0.8203687735427027,
|
|
0.5757417658383376,
|
|
0.22948089201443508,
|
|
0.5395190902988906,
|
|
0.5825784129033227,
|
|
0.846772602018814,
|
|
0.8051629267494426,
|
|
0.8230038229595105,
|
|
0.7209706868204239,
|
|
0.7078282484336693,
|
|
0.8028281928515756,
|
|
0.3355600845410628
|
|
],
|
|
"eval_runtime": 75.9568,
|
|
"eval_samples_per_second": 1.172,
|
|
"eval_steps_per_second": 0.158,
|
|
"step": 5500
|
|
},
|
|
{
|
|
"epoch": 2.5,
|
|
"learning_rate": 3.998184294144349e-05,
|
|
"loss": 0.1144,
|
|
"step": 5510
|
|
},
|
|
{
|
|
"epoch": 2.51,
|
|
"learning_rate": 3.990921470721743e-05,
|
|
"loss": 0.0819,
|
|
"step": 5520
|
|
},
|
|
{
|
|
"epoch": 2.51,
|
|
"learning_rate": 3.9836586472991376e-05,
|
|
"loss": 0.0906,
|
|
"step": 5530
|
|
},
|
|
{
|
|
"epoch": 2.51,
|
|
"learning_rate": 3.9763958238765326e-05,
|
|
"loss": 0.1255,
|
|
"step": 5540
|
|
},
|
|
{
|
|
"epoch": 2.52,
|
|
"learning_rate": 3.969133000453927e-05,
|
|
"loss": 0.1035,
|
|
"step": 5550
|
|
},
|
|
{
|
|
"epoch": 2.52,
|
|
"learning_rate": 3.961870177031321e-05,
|
|
"loss": 0.0909,
|
|
"step": 5560
|
|
},
|
|
{
|
|
"epoch": 2.53,
|
|
"learning_rate": 3.9546073536087156e-05,
|
|
"loss": 0.0893,
|
|
"step": 5570
|
|
},
|
|
{
|
|
"epoch": 2.53,
|
|
"learning_rate": 3.94734453018611e-05,
|
|
"loss": 0.098,
|
|
"step": 5580
|
|
},
|
|
{
|
|
"epoch": 2.54,
|
|
"learning_rate": 3.940081706763505e-05,
|
|
"loss": 0.0941,
|
|
"step": 5590
|
|
},
|
|
{
|
|
"epoch": 2.54,
|
|
"learning_rate": 3.932818883340899e-05,
|
|
"loss": 0.0825,
|
|
"step": 5600
|
|
},
|
|
{
|
|
"epoch": 2.54,
|
|
"eval_loss": 0.1394958645105362,
|
|
"eval_mean_accuracy": 0.7899059096006436,
|
|
"eval_mean_iou": 0.6796236093353095,
|
|
"eval_overall_accuracy": 0.958950857098183,
|
|
"eval_per_category_accuracy": [
|
|
0.9926132606191554,
|
|
0.7325571981411616,
|
|
0.904631960493012,
|
|
0.7543407487791645,
|
|
0.8680581809499174,
|
|
0.7480116057746237,
|
|
0.8881900148590478,
|
|
0.7584741859091426,
|
|
0.27501293326435594,
|
|
0.7362246642246643,
|
|
0.7294654626296012,
|
|
0.9225763344442994,
|
|
0.8980900311134291,
|
|
0.8857988842546152,
|
|
0.8347667701327816,
|
|
0.8093653648509763,
|
|
0.871482659771101,
|
|
0.6086461126005361
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9847737868822146,
|
|
0.6712761695624434,
|
|
0.8086321934304831,
|
|
0.6129174473713215,
|
|
0.7762552482886116,
|
|
0.6416916391685009,
|
|
0.8251962470579106,
|
|
0.5465201941975035,
|
|
0.2496360648039446,
|
|
0.5549248280838156,
|
|
0.5758973642110495,
|
|
0.8395847950060266,
|
|
0.8004792953236646,
|
|
0.8124702388158479,
|
|
0.715316130700604,
|
|
0.7050640702814616,
|
|
0.8061817237235351,
|
|
0.30640753112663227
|
|
],
|
|
"eval_runtime": 75.446,
|
|
"eval_samples_per_second": 1.18,
|
|
"eval_steps_per_second": 0.159,
|
|
"step": 5600
|
|
},
|
|
{
|
|
"epoch": 2.55,
|
|
"learning_rate": 3.9255560599182936e-05,
|
|
"loss": 0.104,
|
|
"step": 5610
|
|
},
|
|
{
|
|
"epoch": 2.55,
|
|
"learning_rate": 3.918293236495688e-05,
|
|
"loss": 0.1045,
|
|
"step": 5620
|
|
},
|
|
{
|
|
"epoch": 2.56,
|
|
"learning_rate": 3.911030413073082e-05,
|
|
"loss": 0.0895,
|
|
"step": 5630
|
|
},
|
|
{
|
|
"epoch": 2.56,
|
|
"learning_rate": 3.903767589650477e-05,
|
|
"loss": 0.1053,
|
|
"step": 5640
|
|
},
|
|
{
|
|
"epoch": 2.56,
|
|
"learning_rate": 3.8965047662278716e-05,
|
|
"loss": 0.0967,
|
|
"step": 5650
|
|
},
|
|
{
|
|
"epoch": 2.57,
|
|
"learning_rate": 3.889241942805266e-05,
|
|
"loss": 0.0967,
|
|
"step": 5660
|
|
},
|
|
{
|
|
"epoch": 2.57,
|
|
"learning_rate": 3.88197911938266e-05,
|
|
"loss": 0.1163,
|
|
"step": 5670
|
|
},
|
|
{
|
|
"epoch": 2.58,
|
|
"learning_rate": 3.8747162959600546e-05,
|
|
"loss": 0.1213,
|
|
"step": 5680
|
|
},
|
|
{
|
|
"epoch": 2.58,
|
|
"learning_rate": 3.8674534725374496e-05,
|
|
"loss": 0.1102,
|
|
"step": 5690
|
|
},
|
|
{
|
|
"epoch": 2.59,
|
|
"learning_rate": 3.860190649114844e-05,
|
|
"loss": 0.1005,
|
|
"step": 5700
|
|
},
|
|
{
|
|
"epoch": 2.59,
|
|
"eval_loss": 0.1268976330757141,
|
|
"eval_mean_accuracy": 0.7871117980869821,
|
|
"eval_mean_iou": 0.686933111067596,
|
|
"eval_overall_accuracy": 0.9599076603235823,
|
|
"eval_per_category_accuracy": [
|
|
0.9926646256019716,
|
|
0.7225667514046535,
|
|
0.8977488445031363,
|
|
0.7453879544221378,
|
|
0.8739060539173475,
|
|
0.7545255013491492,
|
|
0.9012464896804435,
|
|
0.8032218615481048,
|
|
0.2687015002586653,
|
|
0.7288986568986568,
|
|
0.7437191367132936,
|
|
0.9206390625185326,
|
|
0.8672643562397713,
|
|
0.8803741442725657,
|
|
0.8389539144680117,
|
|
0.7926258992805756,
|
|
0.8485032693251212,
|
|
0.5870643431635388
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.985185230857788,
|
|
0.6716788584846022,
|
|
0.8071765933598732,
|
|
0.6127683300808475,
|
|
0.7851744404675747,
|
|
0.6319301249960702,
|
|
0.8206798319274474,
|
|
0.5758901107725064,
|
|
0.23264355460001793,
|
|
0.556702799889586,
|
|
0.5797618226015475,
|
|
0.8450397552629001,
|
|
0.7983956754429897,
|
|
0.8062322215686116,
|
|
0.7237232351218423,
|
|
0.7104749512036435,
|
|
0.803467177811932,
|
|
0.4178712847669481
|
|
],
|
|
"eval_runtime": 75.3313,
|
|
"eval_samples_per_second": 1.181,
|
|
"eval_steps_per_second": 0.159,
|
|
"step": 5700
|
|
},
|
|
{
|
|
"epoch": 2.59,
|
|
"learning_rate": 3.852927825692238e-05,
|
|
"loss": 0.0995,
|
|
"step": 5710
|
|
},
|
|
{
|
|
"epoch": 2.6,
|
|
"learning_rate": 3.8456650022696326e-05,
|
|
"loss": 0.0926,
|
|
"step": 5720
|
|
},
|
|
{
|
|
"epoch": 2.6,
|
|
"learning_rate": 3.838402178847027e-05,
|
|
"loss": 0.0974,
|
|
"step": 5730
|
|
},
|
|
{
|
|
"epoch": 2.61,
|
|
"learning_rate": 3.831139355424421e-05,
|
|
"loss": 0.1003,
|
|
"step": 5740
|
|
},
|
|
{
|
|
"epoch": 2.61,
|
|
"learning_rate": 3.823876532001816e-05,
|
|
"loss": 0.099,
|
|
"step": 5750
|
|
},
|
|
{
|
|
"epoch": 2.61,
|
|
"learning_rate": 3.8166137085792106e-05,
|
|
"loss": 0.103,
|
|
"step": 5760
|
|
},
|
|
{
|
|
"epoch": 2.62,
|
|
"learning_rate": 3.809350885156605e-05,
|
|
"loss": 0.095,
|
|
"step": 5770
|
|
},
|
|
{
|
|
"epoch": 2.62,
|
|
"learning_rate": 3.802088061733999e-05,
|
|
"loss": 0.1116,
|
|
"step": 5780
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"learning_rate": 3.7948252383113936e-05,
|
|
"loss": 0.1032,
|
|
"step": 5790
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"learning_rate": 3.7875624148887886e-05,
|
|
"loss": 0.0992,
|
|
"step": 5800
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"eval_loss": 0.13441899418830872,
|
|
"eval_mean_accuracy": 0.7896584094262842,
|
|
"eval_mean_iou": 0.6706728560954918,
|
|
"eval_overall_accuracy": 0.9575793234150062,
|
|
"eval_per_category_accuracy": [
|
|
0.9917902005597341,
|
|
0.734305931119351,
|
|
0.9237372069990095,
|
|
0.7105941399891481,
|
|
0.8412854479698487,
|
|
0.730797256656318,
|
|
0.9053161984500656,
|
|
0.766902760692422,
|
|
0.29694774961200204,
|
|
0.7168644688644689,
|
|
0.7139428279131992,
|
|
0.9216156057137661,
|
|
0.8879961512868011,
|
|
0.899992080671559,
|
|
0.8425967839282635,
|
|
0.8107464028776978,
|
|
0.8867847676295155,
|
|
0.6316353887399464
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9846100446977286,
|
|
0.6570368867897192,
|
|
0.8084709338373836,
|
|
0.6118670793669334,
|
|
0.7667789610342505,
|
|
0.6227294108792594,
|
|
0.8130077235702489,
|
|
0.5434300379522429,
|
|
0.2513024823781796,
|
|
0.5553503999273547,
|
|
0.5764596554490362,
|
|
0.8449028264068605,
|
|
0.7983644727587295,
|
|
0.803406685335229,
|
|
0.7228602072989191,
|
|
0.7092795207669527,
|
|
0.8002823710255025,
|
|
0.20197171024432062
|
|
],
|
|
"eval_runtime": 76.0516,
|
|
"eval_samples_per_second": 1.17,
|
|
"eval_steps_per_second": 0.158,
|
|
"step": 5800
|
|
},
|
|
{
|
|
"epoch": 2.64,
|
|
"learning_rate": 3.780299591466183e-05,
|
|
"loss": 0.1177,
|
|
"step": 5810
|
|
},
|
|
{
|
|
"epoch": 2.64,
|
|
"learning_rate": 3.773036768043577e-05,
|
|
"loss": 0.0972,
|
|
"step": 5820
|
|
},
|
|
{
|
|
"epoch": 2.65,
|
|
"learning_rate": 3.7657739446209716e-05,
|
|
"loss": 0.0787,
|
|
"step": 5830
|
|
},
|
|
{
|
|
"epoch": 2.65,
|
|
"learning_rate": 3.758511121198366e-05,
|
|
"loss": 0.0953,
|
|
"step": 5840
|
|
},
|
|
{
|
|
"epoch": 2.66,
|
|
"learning_rate": 3.751248297775761e-05,
|
|
"loss": 0.082,
|
|
"step": 5850
|
|
},
|
|
{
|
|
"epoch": 2.66,
|
|
"learning_rate": 3.743985474353155e-05,
|
|
"loss": 0.0967,
|
|
"step": 5860
|
|
},
|
|
{
|
|
"epoch": 2.66,
|
|
"learning_rate": 3.73672265093055e-05,
|
|
"loss": 0.0956,
|
|
"step": 5870
|
|
},
|
|
{
|
|
"epoch": 2.67,
|
|
"learning_rate": 3.729459827507944e-05,
|
|
"loss": 0.1307,
|
|
"step": 5880
|
|
},
|
|
{
|
|
"epoch": 2.67,
|
|
"learning_rate": 3.722197004085338e-05,
|
|
"loss": 0.0974,
|
|
"step": 5890
|
|
},
|
|
{
|
|
"epoch": 2.68,
|
|
"learning_rate": 3.7149341806627333e-05,
|
|
"loss": 0.122,
|
|
"step": 5900
|
|
},
|
|
{
|
|
"epoch": 2.68,
|
|
"eval_loss": 0.13111147284507751,
|
|
"eval_mean_accuracy": 0.7883410579171871,
|
|
"eval_mean_iou": 0.6767765658762446,
|
|
"eval_overall_accuracy": 0.9591094884979591,
|
|
"eval_per_category_accuracy": [
|
|
0.9920656078380079,
|
|
0.7211904337829305,
|
|
0.9036587295036865,
|
|
0.6854991861096039,
|
|
0.8738655723374956,
|
|
0.738629388047112,
|
|
0.9014152634088117,
|
|
0.7453923730972032,
|
|
0.2596999482669426,
|
|
0.6926788766788767,
|
|
0.766471895964902,
|
|
0.9355521027307627,
|
|
0.8793590273906084,
|
|
0.9128037942382486,
|
|
0.8493975254354199,
|
|
0.8116007194244604,
|
|
0.8823197242596575,
|
|
0.638538873994638
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9849246790162564,
|
|
0.6685579622041098,
|
|
0.8104779074411926,
|
|
0.6064078718425632,
|
|
0.7795262192778223,
|
|
0.6397543562370068,
|
|
0.8151970040281974,
|
|
0.5529108469380098,
|
|
0.22835827685029342,
|
|
0.5441911149651981,
|
|
0.5801441462906156,
|
|
0.837493098517774,
|
|
0.8077946108476929,
|
|
0.825658025644495,
|
|
0.7171273231875156,
|
|
0.7089273791288637,
|
|
0.7899439180042545,
|
|
0.2845834453505392
|
|
],
|
|
"eval_runtime": 76.2941,
|
|
"eval_samples_per_second": 1.167,
|
|
"eval_steps_per_second": 0.157,
|
|
"step": 5900
|
|
},
|
|
{
|
|
"epoch": 2.68,
|
|
"learning_rate": 3.707671357240128e-05,
|
|
"loss": 0.1152,
|
|
"step": 5910
|
|
},
|
|
{
|
|
"epoch": 2.69,
|
|
"learning_rate": 3.700408533817522e-05,
|
|
"loss": 0.0882,
|
|
"step": 5920
|
|
},
|
|
{
|
|
"epoch": 2.69,
|
|
"learning_rate": 3.6931457103949163e-05,
|
|
"loss": 0.0882,
|
|
"step": 5930
|
|
},
|
|
{
|
|
"epoch": 2.7,
|
|
"learning_rate": 3.685882886972311e-05,
|
|
"loss": 0.1144,
|
|
"step": 5940
|
|
},
|
|
{
|
|
"epoch": 2.7,
|
|
"learning_rate": 3.678620063549706e-05,
|
|
"loss": 0.0885,
|
|
"step": 5950
|
|
},
|
|
{
|
|
"epoch": 2.71,
|
|
"learning_rate": 3.6713572401271e-05,
|
|
"loss": 0.098,
|
|
"step": 5960
|
|
},
|
|
{
|
|
"epoch": 2.71,
|
|
"learning_rate": 3.6640944167044944e-05,
|
|
"loss": 0.0956,
|
|
"step": 5970
|
|
},
|
|
{
|
|
"epoch": 2.71,
|
|
"learning_rate": 3.656831593281889e-05,
|
|
"loss": 0.1106,
|
|
"step": 5980
|
|
},
|
|
{
|
|
"epoch": 2.72,
|
|
"learning_rate": 3.649568769859283e-05,
|
|
"loss": 0.1136,
|
|
"step": 5990
|
|
},
|
|
{
|
|
"epoch": 2.72,
|
|
"learning_rate": 3.6423059464366774e-05,
|
|
"loss": 0.0938,
|
|
"step": 6000
|
|
},
|
|
{
|
|
"epoch": 2.72,
|
|
"eval_loss": 0.1398986577987671,
|
|
"eval_mean_accuracy": 0.7859709892856883,
|
|
"eval_mean_iou": 0.687898138786663,
|
|
"eval_overall_accuracy": 0.9594852576095066,
|
|
"eval_per_category_accuracy": [
|
|
0.9923034066299434,
|
|
0.7256108421444648,
|
|
0.9027697534940025,
|
|
0.7238876831253391,
|
|
0.8872773255809965,
|
|
0.7376556158612253,
|
|
0.8999154391422814,
|
|
0.7498649730036326,
|
|
0.2488877392653906,
|
|
0.7262026862026862,
|
|
0.7584744929075922,
|
|
0.9152463300557855,
|
|
0.8725293599266226,
|
|
0.8950073033806734,
|
|
0.827357087428867,
|
|
0.8159750770811922,
|
|
0.86911590906987,
|
|
0.5993967828418231
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.984698799272596,
|
|
0.6722924824099494,
|
|
0.8096993288259389,
|
|
0.6134613173928037,
|
|
0.793041382524653,
|
|
0.6348061385039867,
|
|
0.8116350488899908,
|
|
0.5510817993792497,
|
|
0.22583673660986717,
|
|
0.5578189948828763,
|
|
0.5852149823536109,
|
|
0.8380704136120916,
|
|
0.7979941279514445,
|
|
0.8050082507924322,
|
|
0.710620591339122,
|
|
0.7023525834185719,
|
|
0.7917829784243446,
|
|
0.4967505415764039
|
|
],
|
|
"eval_runtime": 78.0349,
|
|
"eval_samples_per_second": 1.141,
|
|
"eval_steps_per_second": 0.154,
|
|
"step": 6000
|
|
},
|
|
{
|
|
"epoch": 2.73,
|
|
"learning_rate": 3.6350431230140724e-05,
|
|
"loss": 0.0994,
|
|
"step": 6010
|
|
},
|
|
{
|
|
"epoch": 2.73,
|
|
"learning_rate": 3.627780299591467e-05,
|
|
"loss": 0.0855,
|
|
"step": 6020
|
|
},
|
|
{
|
|
"epoch": 2.74,
|
|
"learning_rate": 3.620517476168861e-05,
|
|
"loss": 0.0998,
|
|
"step": 6030
|
|
},
|
|
{
|
|
"epoch": 2.74,
|
|
"learning_rate": 3.6132546527462554e-05,
|
|
"loss": 0.0939,
|
|
"step": 6040
|
|
},
|
|
{
|
|
"epoch": 2.75,
|
|
"learning_rate": 3.60599182932365e-05,
|
|
"loss": 0.1038,
|
|
"step": 6050
|
|
},
|
|
{
|
|
"epoch": 2.75,
|
|
"learning_rate": 3.598729005901045e-05,
|
|
"loss": 0.0786,
|
|
"step": 6060
|
|
},
|
|
{
|
|
"epoch": 2.76,
|
|
"learning_rate": 3.591466182478439e-05,
|
|
"loss": 0.1073,
|
|
"step": 6070
|
|
},
|
|
{
|
|
"epoch": 2.76,
|
|
"learning_rate": 3.5842033590558334e-05,
|
|
"loss": 0.1006,
|
|
"step": 6080
|
|
},
|
|
{
|
|
"epoch": 2.76,
|
|
"learning_rate": 3.576940535633228e-05,
|
|
"loss": 0.0964,
|
|
"step": 6090
|
|
},
|
|
{
|
|
"epoch": 2.77,
|
|
"learning_rate": 3.569677712210622e-05,
|
|
"loss": 0.0984,
|
|
"step": 6100
|
|
},
|
|
{
|
|
"epoch": 2.77,
|
|
"eval_loss": 0.137928768992424,
|
|
"eval_mean_accuracy": 0.7898073705363208,
|
|
"eval_mean_iou": 0.6917297696873022,
|
|
"eval_overall_accuracy": 0.9597389135467872,
|
|
"eval_per_category_accuracy": [
|
|
0.992448127191679,
|
|
0.7315209119318642,
|
|
0.9064425828106085,
|
|
0.7121540965816603,
|
|
0.8710037304805026,
|
|
0.7046511422502307,
|
|
0.8946434350499536,
|
|
0.8119159743710632,
|
|
0.32762545266425247,
|
|
0.7186910866910867,
|
|
0.7254224950323634,
|
|
0.9250908343316214,
|
|
0.8800739168749887,
|
|
0.8981046407264663,
|
|
0.8404304621486464,
|
|
0.8129946043165468,
|
|
0.8826221252619009,
|
|
0.5806970509383378
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9841650313607442,
|
|
0.6620651250036637,
|
|
0.8106955128303709,
|
|
0.6111402130260171,
|
|
0.7907538668824434,
|
|
0.6277769581083341,
|
|
0.816592235908447,
|
|
0.5801459816329033,
|
|
0.27775097583439323,
|
|
0.5573939393939394,
|
|
0.585910188775186,
|
|
0.8438983222009031,
|
|
0.7917123384634053,
|
|
0.8002634457290038,
|
|
0.7176573989710742,
|
|
0.7108468921825769,
|
|
0.7864883071889135,
|
|
0.4958791208791209
|
|
],
|
|
"eval_runtime": 76.5926,
|
|
"eval_samples_per_second": 1.162,
|
|
"eval_steps_per_second": 0.157,
|
|
"step": 6100
|
|
},
|
|
{
|
|
"epoch": 2.77,
|
|
"learning_rate": 3.562414888788017e-05,
|
|
"loss": 0.1072,
|
|
"step": 6110
|
|
},
|
|
{
|
|
"epoch": 2.78,
|
|
"learning_rate": 3.5551520653654114e-05,
|
|
"loss": 0.0932,
|
|
"step": 6120
|
|
},
|
|
{
|
|
"epoch": 2.78,
|
|
"learning_rate": 3.547889241942806e-05,
|
|
"loss": 0.099,
|
|
"step": 6130
|
|
},
|
|
{
|
|
"epoch": 2.79,
|
|
"learning_rate": 3.5406264185202e-05,
|
|
"loss": 0.0938,
|
|
"step": 6140
|
|
},
|
|
{
|
|
"epoch": 2.79,
|
|
"learning_rate": 3.5333635950975944e-05,
|
|
"loss": 0.1093,
|
|
"step": 6150
|
|
},
|
|
{
|
|
"epoch": 2.8,
|
|
"learning_rate": 3.5261007716749894e-05,
|
|
"loss": 0.0923,
|
|
"step": 6160
|
|
},
|
|
{
|
|
"epoch": 2.8,
|
|
"learning_rate": 3.518837948252384e-05,
|
|
"loss": 0.0922,
|
|
"step": 6170
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"learning_rate": 3.511575124829778e-05,
|
|
"loss": 0.1006,
|
|
"step": 6180
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"learning_rate": 3.5043123014071724e-05,
|
|
"loss": 0.0874,
|
|
"step": 6190
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"learning_rate": 3.497049477984567e-05,
|
|
"loss": 0.0839,
|
|
"step": 6200
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"eval_loss": 0.13149453699588776,
|
|
"eval_mean_accuracy": 0.7891440579467708,
|
|
"eval_mean_iou": 0.6860310300219666,
|
|
"eval_overall_accuracy": 0.9601006668605162,
|
|
"eval_per_category_accuracy": [
|
|
0.9920891270094053,
|
|
0.7334477566022766,
|
|
0.913305064927919,
|
|
0.7107976125881714,
|
|
0.8711361532756117,
|
|
0.7494358304002402,
|
|
0.8823664512626711,
|
|
0.8053714552026489,
|
|
0.24852560786342473,
|
|
0.7472625152625153,
|
|
0.6937476637352692,
|
|
0.9306179897443196,
|
|
0.8881715015376869,
|
|
0.9095656688312831,
|
|
0.8478509225728574,
|
|
0.8069373072970195,
|
|
0.8711494015237308,
|
|
0.6028150134048257
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850031413953964,
|
|
0.664549162289839,
|
|
0.810160003660713,
|
|
0.6136909293201382,
|
|
0.789657160520625,
|
|
0.6421957071280444,
|
|
0.8123257017182747,
|
|
0.5809162583014003,
|
|
0.2266144629463654,
|
|
0.5653423786931524,
|
|
0.5706552522130338,
|
|
0.843401208933351,
|
|
0.7949519704780417,
|
|
0.8085170454323247,
|
|
0.7231519485951454,
|
|
0.7109813176675666,
|
|
0.7935351491131911,
|
|
0.4129097419887981
|
|
],
|
|
"eval_runtime": 76.4248,
|
|
"eval_samples_per_second": 1.165,
|
|
"eval_steps_per_second": 0.157,
|
|
"step": 6200
|
|
},
|
|
{
|
|
"epoch": 2.82,
|
|
"learning_rate": 3.489786654561962e-05,
|
|
"loss": 0.09,
|
|
"step": 6210
|
|
},
|
|
{
|
|
"epoch": 2.82,
|
|
"learning_rate": 3.482523831139356e-05,
|
|
"loss": 0.0945,
|
|
"step": 6220
|
|
},
|
|
{
|
|
"epoch": 2.83,
|
|
"learning_rate": 3.4752610077167504e-05,
|
|
"loss": 0.0961,
|
|
"step": 6230
|
|
},
|
|
{
|
|
"epoch": 2.83,
|
|
"learning_rate": 3.467998184294145e-05,
|
|
"loss": 0.0924,
|
|
"step": 6240
|
|
},
|
|
{
|
|
"epoch": 2.84,
|
|
"learning_rate": 3.460735360871539e-05,
|
|
"loss": 0.0886,
|
|
"step": 6250
|
|
},
|
|
{
|
|
"epoch": 2.84,
|
|
"learning_rate": 3.4534725374489334e-05,
|
|
"loss": 0.0921,
|
|
"step": 6260
|
|
},
|
|
{
|
|
"epoch": 2.85,
|
|
"learning_rate": 3.4462097140263284e-05,
|
|
"loss": 0.0909,
|
|
"step": 6270
|
|
},
|
|
{
|
|
"epoch": 2.85,
|
|
"learning_rate": 3.438946890603723e-05,
|
|
"loss": 0.0939,
|
|
"step": 6280
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"learning_rate": 3.431684067181117e-05,
|
|
"loss": 0.1058,
|
|
"step": 6290
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"learning_rate": 3.4244212437585114e-05,
|
|
"loss": 0.0846,
|
|
"step": 6300
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"eval_loss": 0.13326804339885712,
|
|
"eval_mean_accuracy": 0.7897628723404772,
|
|
"eval_mean_iou": 0.6789238793132619,
|
|
"eval_overall_accuracy": 0.9593457425578257,
|
|
"eval_per_category_accuracy": [
|
|
0.9920089177597812,
|
|
0.7340792435110672,
|
|
0.9173836937383075,
|
|
0.7316196418882257,
|
|
0.8678104885715011,
|
|
0.7462804552219362,
|
|
0.9081000950004697,
|
|
0.7618846336568562,
|
|
0.25835488877392654,
|
|
0.693997557997558,
|
|
0.7528281099372406,
|
|
0.9127595054816888,
|
|
0.8971728144164883,
|
|
0.8777299684986714,
|
|
0.8506261855492326,
|
|
0.7959982014388489,
|
|
0.8912930111425512,
|
|
0.625804289544236
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9847307623055664,
|
|
0.6671081093011964,
|
|
0.8108197239522068,
|
|
0.6178121420389462,
|
|
0.7841557259556485,
|
|
0.6326530230224796,
|
|
0.8226430321008514,
|
|
0.5525983418407454,
|
|
0.23203085071783675,
|
|
0.542537512886106,
|
|
0.5779894115958885,
|
|
0.8460912619153201,
|
|
0.808235379891934,
|
|
0.812661156620813,
|
|
0.7267047861075815,
|
|
0.7150828640014772,
|
|
0.8132121208708603,
|
|
0.2735636225132577
|
|
],
|
|
"eval_runtime": 77.206,
|
|
"eval_samples_per_second": 1.153,
|
|
"eval_steps_per_second": 0.155,
|
|
"step": 6300
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"learning_rate": 3.417158420335906e-05,
|
|
"loss": 0.0909,
|
|
"step": 6310
|
|
},
|
|
{
|
|
"epoch": 2.87,
|
|
"learning_rate": 3.409895596913301e-05,
|
|
"loss": 0.0992,
|
|
"step": 6320
|
|
},
|
|
{
|
|
"epoch": 2.87,
|
|
"learning_rate": 3.402632773490695e-05,
|
|
"loss": 0.0878,
|
|
"step": 6330
|
|
},
|
|
{
|
|
"epoch": 2.88,
|
|
"learning_rate": 3.3953699500680894e-05,
|
|
"loss": 0.1106,
|
|
"step": 6340
|
|
},
|
|
{
|
|
"epoch": 2.88,
|
|
"learning_rate": 3.388107126645484e-05,
|
|
"loss": 0.0934,
|
|
"step": 6350
|
|
},
|
|
{
|
|
"epoch": 2.89,
|
|
"learning_rate": 3.380844303222878e-05,
|
|
"loss": 0.0987,
|
|
"step": 6360
|
|
},
|
|
{
|
|
"epoch": 2.89,
|
|
"learning_rate": 3.373581479800273e-05,
|
|
"loss": 0.0919,
|
|
"step": 6370
|
|
},
|
|
{
|
|
"epoch": 2.9,
|
|
"learning_rate": 3.3663186563776674e-05,
|
|
"loss": 0.0949,
|
|
"step": 6380
|
|
},
|
|
{
|
|
"epoch": 2.9,
|
|
"learning_rate": 3.359055832955062e-05,
|
|
"loss": 0.1013,
|
|
"step": 6390
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"learning_rate": 3.351793009532456e-05,
|
|
"loss": 0.0927,
|
|
"step": 6400
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"eval_loss": 0.1403910517692566,
|
|
"eval_mean_accuracy": 0.788558184698478,
|
|
"eval_mean_iou": 0.6774119670611869,
|
|
"eval_overall_accuracy": 0.9594932298981742,
|
|
"eval_per_category_accuracy": [
|
|
0.9924294893577413,
|
|
0.7339497077349051,
|
|
0.9050033701991856,
|
|
0.7270075963103635,
|
|
0.8760371689515888,
|
|
0.7769487588268805,
|
|
0.866345126615095,
|
|
0.7706600336348184,
|
|
0.25675116399379205,
|
|
0.7258119658119658,
|
|
0.7562120049577997,
|
|
0.9191722709175948,
|
|
0.8761802420732694,
|
|
0.8969739366101754,
|
|
0.7995182359027418,
|
|
0.8099113566289825,
|
|
0.8880437432306918,
|
|
0.6170911528150134
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850749687225412,
|
|
0.6666666666666666,
|
|
0.8110797424620212,
|
|
0.6137066300240467,
|
|
0.7827799603700107,
|
|
0.6567757301541971,
|
|
0.8086142972910207,
|
|
0.5685894548294493,
|
|
0.23417004812682835,
|
|
0.5639591365858101,
|
|
0.5987802408362283,
|
|
0.8434607835667925,
|
|
0.7802691480714792,
|
|
0.7886176031440264,
|
|
0.6986682614878928,
|
|
0.7000555216256732,
|
|
0.8084970698790307,
|
|
0.2836501432576481
|
|
],
|
|
"eval_runtime": 76.4223,
|
|
"eval_samples_per_second": 1.165,
|
|
"eval_steps_per_second": 0.157,
|
|
"step": 6400
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"learning_rate": 3.3445301861098504e-05,
|
|
"loss": 0.0984,
|
|
"step": 6410
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"learning_rate": 3.3372673626872455e-05,
|
|
"loss": 0.0938,
|
|
"step": 6420
|
|
},
|
|
{
|
|
"epoch": 2.92,
|
|
"learning_rate": 3.330004539264639e-05,
|
|
"loss": 0.0861,
|
|
"step": 6430
|
|
},
|
|
{
|
|
"epoch": 2.92,
|
|
"learning_rate": 3.3227417158420334e-05,
|
|
"loss": 0.0917,
|
|
"step": 6440
|
|
},
|
|
{
|
|
"epoch": 2.93,
|
|
"learning_rate": 3.315478892419428e-05,
|
|
"loss": 0.1024,
|
|
"step": 6450
|
|
},
|
|
{
|
|
"epoch": 2.93,
|
|
"learning_rate": 3.308216068996823e-05,
|
|
"loss": 0.0953,
|
|
"step": 6460
|
|
},
|
|
{
|
|
"epoch": 2.94,
|
|
"learning_rate": 3.300953245574217e-05,
|
|
"loss": 0.1043,
|
|
"step": 6470
|
|
},
|
|
{
|
|
"epoch": 2.94,
|
|
"learning_rate": 3.2936904221516115e-05,
|
|
"loss": 0.1355,
|
|
"step": 6480
|
|
},
|
|
{
|
|
"epoch": 2.95,
|
|
"learning_rate": 3.286427598729006e-05,
|
|
"loss": 0.098,
|
|
"step": 6490
|
|
},
|
|
{
|
|
"epoch": 2.95,
|
|
"learning_rate": 3.2791647753064e-05,
|
|
"loss": 0.0909,
|
|
"step": 6500
|
|
},
|
|
{
|
|
"epoch": 2.95,
|
|
"eval_loss": 0.13613344728946686,
|
|
"eval_mean_accuracy": 0.7934710007087689,
|
|
"eval_mean_iou": 0.6817604654493301,
|
|
"eval_overall_accuracy": 0.9605776754657874,
|
|
"eval_per_category_accuracy": [
|
|
0.9919079628259527,
|
|
0.721287585615052,
|
|
0.9140788351491141,
|
|
0.7199538795442214,
|
|
0.8629835747705071,
|
|
0.7538122849862435,
|
|
0.9116878416798032,
|
|
0.8138650597099313,
|
|
0.27568546301086394,
|
|
0.7141978021978022,
|
|
0.7089653544236558,
|
|
0.9209948879742856,
|
|
0.8852759743179325,
|
|
0.8958828291360893,
|
|
0.8405220727711674,
|
|
0.811157502569373,
|
|
0.9038920243278521,
|
|
0.6363270777479892
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9848577133289431,
|
|
0.668126527979842,
|
|
0.8098557003832959,
|
|
0.6173665232057695,
|
|
0.7912109618271054,
|
|
0.6499847682875747,
|
|
0.8221945356288797,
|
|
0.5992894452964657,
|
|
0.2481143495669988,
|
|
0.553552636559791,
|
|
0.5704064834747372,
|
|
0.8454001088731627,
|
|
0.8014164478905916,
|
|
0.799446429272506,
|
|
0.7140162786226344,
|
|
0.7018340465736675,
|
|
0.8094589366640875,
|
|
0.28515648465188925
|
|
],
|
|
"eval_runtime": 82.5196,
|
|
"eval_samples_per_second": 1.079,
|
|
"eval_steps_per_second": 0.145,
|
|
"step": 6500
|
|
},
|
|
{
|
|
"epoch": 2.96,
|
|
"learning_rate": 3.271901951883795e-05,
|
|
"loss": 0.0998,
|
|
"step": 6510
|
|
},
|
|
{
|
|
"epoch": 2.96,
|
|
"learning_rate": 3.2646391284611895e-05,
|
|
"loss": 0.0985,
|
|
"step": 6520
|
|
},
|
|
{
|
|
"epoch": 2.96,
|
|
"learning_rate": 3.257376305038584e-05,
|
|
"loss": 0.087,
|
|
"step": 6530
|
|
},
|
|
{
|
|
"epoch": 2.97,
|
|
"learning_rate": 3.250113481615978e-05,
|
|
"loss": 0.0805,
|
|
"step": 6540
|
|
},
|
|
{
|
|
"epoch": 2.97,
|
|
"learning_rate": 3.2428506581933725e-05,
|
|
"loss": 0.1091,
|
|
"step": 6550
|
|
},
|
|
{
|
|
"epoch": 2.98,
|
|
"learning_rate": 3.2355878347707675e-05,
|
|
"loss": 0.0962,
|
|
"step": 6560
|
|
},
|
|
{
|
|
"epoch": 2.98,
|
|
"learning_rate": 3.228325011348162e-05,
|
|
"loss": 0.1055,
|
|
"step": 6570
|
|
},
|
|
{
|
|
"epoch": 2.99,
|
|
"learning_rate": 3.221062187925556e-05,
|
|
"loss": 0.0825,
|
|
"step": 6580
|
|
},
|
|
{
|
|
"epoch": 2.99,
|
|
"learning_rate": 3.2137993645029505e-05,
|
|
"loss": 0.1204,
|
|
"step": 6590
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 3.206536541080345e-05,
|
|
"loss": 0.1103,
|
|
"step": 6600
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"eval_loss": 0.13448721170425415,
|
|
"eval_mean_accuracy": 0.7842717389115598,
|
|
"eval_mean_iou": 0.6855736569488313,
|
|
"eval_overall_accuracy": 0.9587874251804994,
|
|
"eval_per_category_accuracy": [
|
|
0.9925916274190494,
|
|
0.7192150131964572,
|
|
0.9178823456586332,
|
|
0.736096039066739,
|
|
0.871658983171327,
|
|
0.6566337954151008,
|
|
0.9092432325214795,
|
|
0.8086428115962448,
|
|
0.29896533885152615,
|
|
0.697982905982906,
|
|
0.7484310137913396,
|
|
0.9177805980239826,
|
|
0.8860223369242667,
|
|
0.8895209686218609,
|
|
0.829609630970857,
|
|
0.7929663412127441,
|
|
0.8548413474741802,
|
|
0.5888069705093834
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9849901246619346,
|
|
0.6681307441223808,
|
|
0.8113800624710255,
|
|
0.6178061137359823,
|
|
0.7902524897517402,
|
|
0.5755218161233199,
|
|
0.8346864812457373,
|
|
0.5466369043621395,
|
|
0.26839123165521084,
|
|
0.5516176595465458,
|
|
0.5830906471291499,
|
|
0.8445767965771167,
|
|
0.7993136988468356,
|
|
0.7957939400618746,
|
|
0.7185719019991318,
|
|
0.7044205673070888,
|
|
0.7959448704651303,
|
|
0.44919977501661806
|
|
],
|
|
"eval_runtime": 82.4042,
|
|
"eval_samples_per_second": 1.08,
|
|
"eval_steps_per_second": 0.146,
|
|
"step": 6600
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 3.19927371765774e-05,
|
|
"loss": 0.1185,
|
|
"step": 6610
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 3.192010894235134e-05,
|
|
"loss": 0.0955,
|
|
"step": 6620
|
|
},
|
|
{
|
|
"epoch": 3.01,
|
|
"learning_rate": 3.1847480708125285e-05,
|
|
"loss": 0.0764,
|
|
"step": 6630
|
|
},
|
|
{
|
|
"epoch": 3.01,
|
|
"learning_rate": 3.177485247389923e-05,
|
|
"loss": 0.0909,
|
|
"step": 6640
|
|
},
|
|
{
|
|
"epoch": 3.02,
|
|
"learning_rate": 3.170222423967317e-05,
|
|
"loss": 0.0899,
|
|
"step": 6650
|
|
},
|
|
{
|
|
"epoch": 3.02,
|
|
"learning_rate": 3.162959600544712e-05,
|
|
"loss": 0.0773,
|
|
"step": 6660
|
|
},
|
|
{
|
|
"epoch": 3.03,
|
|
"learning_rate": 3.1556967771221065e-05,
|
|
"loss": 0.076,
|
|
"step": 6670
|
|
},
|
|
{
|
|
"epoch": 3.03,
|
|
"learning_rate": 3.148433953699501e-05,
|
|
"loss": 0.1028,
|
|
"step": 6680
|
|
},
|
|
{
|
|
"epoch": 3.04,
|
|
"learning_rate": 3.141171130276895e-05,
|
|
"loss": 0.0983,
|
|
"step": 6690
|
|
},
|
|
{
|
|
"epoch": 3.04,
|
|
"learning_rate": 3.1339083068542895e-05,
|
|
"loss": 0.099,
|
|
"step": 6700
|
|
},
|
|
{
|
|
"epoch": 3.04,
|
|
"eval_loss": 0.13621346652507782,
|
|
"eval_mean_accuracy": 0.798219757662134,
|
|
"eval_mean_iou": 0.6859926694874328,
|
|
"eval_overall_accuracy": 0.9610216805104459,
|
|
"eval_per_category_accuracy": [
|
|
0.99322392702933,
|
|
0.7230848945093022,
|
|
0.9045305106195665,
|
|
0.7362995116657624,
|
|
0.8652773022526283,
|
|
0.7670388051528226,
|
|
0.9050082298941772,
|
|
0.8048457982669579,
|
|
0.3523538541127781,
|
|
0.7736556776556777,
|
|
0.7125656613350646,
|
|
0.9152147011263853,
|
|
0.8902442314263619,
|
|
0.8796218080706756,
|
|
0.8431195033626487,
|
|
0.7978096094552929,
|
|
0.8702144269963866,
|
|
0.6338471849865952
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.985045217091938,
|
|
0.6707773188133683,
|
|
0.8138994442450272,
|
|
0.6172741229317109,
|
|
0.7934656634943166,
|
|
0.662450250677468,
|
|
0.8354277692651164,
|
|
0.5877626455037622,
|
|
0.29528310066764935,
|
|
0.5570105420098036,
|
|
0.5721168275230227,
|
|
0.8490454620477177,
|
|
0.8029107394851664,
|
|
0.795529966019147,
|
|
0.7260577204193293,
|
|
0.7072379097695554,
|
|
0.8079467352721251,
|
|
0.26862661553756567
|
|
],
|
|
"eval_runtime": 83.2771,
|
|
"eval_samples_per_second": 1.069,
|
|
"eval_steps_per_second": 0.144,
|
|
"step": 6700
|
|
},
|
|
{
|
|
"epoch": 3.05,
|
|
"learning_rate": 3.126645483431684e-05,
|
|
"loss": 0.0841,
|
|
"step": 6710
|
|
},
|
|
{
|
|
"epoch": 3.05,
|
|
"learning_rate": 3.119382660009079e-05,
|
|
"loss": 0.1462,
|
|
"step": 6720
|
|
},
|
|
{
|
|
"epoch": 3.05,
|
|
"learning_rate": 3.112119836586473e-05,
|
|
"loss": 0.0919,
|
|
"step": 6730
|
|
},
|
|
{
|
|
"epoch": 3.06,
|
|
"learning_rate": 3.1048570131638675e-05,
|
|
"loss": 0.0887,
|
|
"step": 6740
|
|
},
|
|
{
|
|
"epoch": 3.06,
|
|
"learning_rate": 3.097594189741262e-05,
|
|
"loss": 0.0831,
|
|
"step": 6750
|
|
},
|
|
{
|
|
"epoch": 3.07,
|
|
"learning_rate": 3.090331366318656e-05,
|
|
"loss": 0.0868,
|
|
"step": 6760
|
|
},
|
|
{
|
|
"epoch": 3.07,
|
|
"learning_rate": 3.083068542896051e-05,
|
|
"loss": 0.0918,
|
|
"step": 6770
|
|
},
|
|
{
|
|
"epoch": 3.08,
|
|
"learning_rate": 3.0758057194734455e-05,
|
|
"loss": 0.087,
|
|
"step": 6780
|
|
},
|
|
{
|
|
"epoch": 3.08,
|
|
"learning_rate": 3.06854289605084e-05,
|
|
"loss": 0.1027,
|
|
"step": 6790
|
|
},
|
|
{
|
|
"epoch": 3.09,
|
|
"learning_rate": 3.061280072628234e-05,
|
|
"loss": 0.0885,
|
|
"step": 6800
|
|
},
|
|
{
|
|
"epoch": 3.09,
|
|
"eval_loss": 0.12955991923809052,
|
|
"eval_mean_accuracy": 0.8001941118044491,
|
|
"eval_mean_iou": 0.6919037217109923,
|
|
"eval_overall_accuracy": 0.9617451014143698,
|
|
"eval_per_category_accuracy": [
|
|
0.9927245883950861,
|
|
0.729610259233472,
|
|
0.9164878397711016,
|
|
0.7216494845360825,
|
|
0.8643489704299212,
|
|
0.7607081818944617,
|
|
0.9047802983641128,
|
|
0.8274327890706711,
|
|
0.32338334195550955,
|
|
0.7358534798534798,
|
|
0.7726494717582483,
|
|
0.9263401770429324,
|
|
0.886921568980091,
|
|
0.8920419548422294,
|
|
0.8435775564752543,
|
|
0.8017792908530319,
|
|
0.8809527483005373,
|
|
0.6222520107238606
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9853156922191287,
|
|
0.6713949399529159,
|
|
0.810264362049832,
|
|
0.6117755289788408,
|
|
0.7942374740717857,
|
|
0.6503930586076437,
|
|
0.8294382237534692,
|
|
0.6112378052279226,
|
|
0.2745761222876219,
|
|
0.5723913654633731,
|
|
0.5958625084396028,
|
|
0.8461885723985251,
|
|
0.8148460251564533,
|
|
0.8138547252817828,
|
|
0.7261893164474752,
|
|
0.7106314902532337,
|
|
0.8129164341279513,
|
|
0.3227533460803059
|
|
],
|
|
"eval_runtime": 80.8852,
|
|
"eval_samples_per_second": 1.1,
|
|
"eval_steps_per_second": 0.148,
|
|
"step": 6800
|
|
},
|
|
{
|
|
"epoch": 3.09,
|
|
"learning_rate": 3.0540172492056285e-05,
|
|
"loss": 0.0952,
|
|
"step": 6810
|
|
},
|
|
{
|
|
"epoch": 3.1,
|
|
"learning_rate": 3.0467544257830232e-05,
|
|
"loss": 0.0908,
|
|
"step": 6820
|
|
},
|
|
{
|
|
"epoch": 3.1,
|
|
"learning_rate": 3.039491602360418e-05,
|
|
"loss": 0.0908,
|
|
"step": 6830
|
|
},
|
|
{
|
|
"epoch": 3.1,
|
|
"learning_rate": 3.0322287789378122e-05,
|
|
"loss": 0.0889,
|
|
"step": 6840
|
|
},
|
|
{
|
|
"epoch": 3.11,
|
|
"learning_rate": 3.0249659555152065e-05,
|
|
"loss": 0.0812,
|
|
"step": 6850
|
|
},
|
|
{
|
|
"epoch": 3.11,
|
|
"learning_rate": 3.0177031320926012e-05,
|
|
"loss": 0.075,
|
|
"step": 6860
|
|
},
|
|
{
|
|
"epoch": 3.12,
|
|
"learning_rate": 3.0104403086699956e-05,
|
|
"loss": 0.0874,
|
|
"step": 6870
|
|
},
|
|
{
|
|
"epoch": 3.12,
|
|
"learning_rate": 3.0031774852473902e-05,
|
|
"loss": 0.0806,
|
|
"step": 6880
|
|
},
|
|
{
|
|
"epoch": 3.13,
|
|
"learning_rate": 2.9959146618247846e-05,
|
|
"loss": 0.0824,
|
|
"step": 6890
|
|
},
|
|
{
|
|
"epoch": 3.13,
|
|
"learning_rate": 2.988651838402179e-05,
|
|
"loss": 0.07,
|
|
"step": 6900
|
|
},
|
|
{
|
|
"epoch": 3.13,
|
|
"eval_loss": 0.14067216217517853,
|
|
"eval_mean_accuracy": 0.798203682941023,
|
|
"eval_mean_iou": 0.6923460941833122,
|
|
"eval_overall_accuracy": 0.9625431446546919,
|
|
"eval_per_category_accuracy": [
|
|
0.9924110734027792,
|
|
0.7320066710924723,
|
|
0.9121994332563003,
|
|
0.7206999457406402,
|
|
0.8750999174587448,
|
|
0.7592420033651447,
|
|
0.907545055625733,
|
|
0.8627854759724202,
|
|
0.2620279358510088,
|
|
0.7625592185592186,
|
|
0.7480473745302878,
|
|
0.9202120719716288,
|
|
0.8997895796989371,
|
|
0.8822791827253049,
|
|
0.8390886359717192,
|
|
0.7952530832476875,
|
|
0.8624384012244155,
|
|
0.6339812332439678
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9851986993485564,
|
|
0.6720680274131446,
|
|
0.8130522140671002,
|
|
0.6155004633920297,
|
|
0.8019457961284027,
|
|
0.6852198381430089,
|
|
0.8350000880469114,
|
|
0.626904146267834,
|
|
0.23910683094934618,
|
|
0.5700235115440221,
|
|
0.5920893837349632,
|
|
0.8446416971799552,
|
|
0.8095238095238095,
|
|
0.8059246220250296,
|
|
0.7220269505782411,
|
|
0.708903305600568,
|
|
0.7915114991518296,
|
|
0.3435888122048674
|
|
],
|
|
"eval_runtime": 81.7478,
|
|
"eval_samples_per_second": 1.089,
|
|
"eval_steps_per_second": 0.147,
|
|
"step": 6900
|
|
},
|
|
{
|
|
"epoch": 3.14,
|
|
"learning_rate": 2.9813890149795736e-05,
|
|
"loss": 0.0953,
|
|
"step": 6910
|
|
},
|
|
{
|
|
"epoch": 3.14,
|
|
"learning_rate": 2.974126191556968e-05,
|
|
"loss": 0.0813,
|
|
"step": 6920
|
|
},
|
|
{
|
|
"epoch": 3.15,
|
|
"learning_rate": 2.9668633681343622e-05,
|
|
"loss": 0.09,
|
|
"step": 6930
|
|
},
|
|
{
|
|
"epoch": 3.15,
|
|
"learning_rate": 2.959600544711757e-05,
|
|
"loss": 0.087,
|
|
"step": 6940
|
|
},
|
|
{
|
|
"epoch": 3.15,
|
|
"learning_rate": 2.9523377212891512e-05,
|
|
"loss": 0.089,
|
|
"step": 6950
|
|
},
|
|
{
|
|
"epoch": 3.16,
|
|
"learning_rate": 2.945074897866546e-05,
|
|
"loss": 0.0836,
|
|
"step": 6960
|
|
},
|
|
{
|
|
"epoch": 3.16,
|
|
"learning_rate": 2.9378120744439402e-05,
|
|
"loss": 0.0816,
|
|
"step": 6970
|
|
},
|
|
{
|
|
"epoch": 3.17,
|
|
"learning_rate": 2.9305492510213346e-05,
|
|
"loss": 0.0932,
|
|
"step": 6980
|
|
},
|
|
{
|
|
"epoch": 3.17,
|
|
"learning_rate": 2.9232864275987292e-05,
|
|
"loss": 0.0813,
|
|
"step": 6990
|
|
},
|
|
{
|
|
"epoch": 3.18,
|
|
"learning_rate": 2.9160236041761236e-05,
|
|
"loss": 0.089,
|
|
"step": 7000
|
|
},
|
|
{
|
|
"epoch": 3.18,
|
|
"eval_loss": 0.13973882794380188,
|
|
"eval_mean_accuracy": 0.7966314280330089,
|
|
"eval_mean_iou": 0.6919191740005027,
|
|
"eval_overall_accuracy": 0.962646570098534,
|
|
"eval_per_category_accuracy": [
|
|
0.9923436776639871,
|
|
0.7139040463738079,
|
|
0.9156005832507979,
|
|
0.7291779706999457,
|
|
0.8699217058800524,
|
|
0.7384792372338688,
|
|
0.9310446223818323,
|
|
0.8831686816849201,
|
|
0.26735644076564924,
|
|
0.7537680097680097,
|
|
0.7258749926223219,
|
|
0.9226593603839752,
|
|
0.8784373145333885,
|
|
0.8745534378684687,
|
|
0.8363079841351957,
|
|
0.7912769784172662,
|
|
0.8772869075794651,
|
|
0.6382037533512064
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9851632581820728,
|
|
0.6638860446907185,
|
|
0.8099934133767777,
|
|
0.6114776475941304,
|
|
0.8028141902459472,
|
|
0.6733761121726402,
|
|
0.8356453610308161,
|
|
0.6397615333884661,
|
|
0.24225378521539398,
|
|
0.56523904747255,
|
|
0.5834618212870935,
|
|
0.8470755203229015,
|
|
0.8098479164680475,
|
|
0.8077524797938958,
|
|
0.726401273140022,
|
|
0.7105381553902059,
|
|
0.790298100871731,
|
|
0.34955947136563875
|
|
],
|
|
"eval_runtime": 84.3262,
|
|
"eval_samples_per_second": 1.055,
|
|
"eval_steps_per_second": 0.142,
|
|
"step": 7000
|
|
},
|
|
{
|
|
"epoch": 3.18,
|
|
"learning_rate": 2.9087607807535183e-05,
|
|
"loss": 0.0774,
|
|
"step": 7010
|
|
},
|
|
{
|
|
"epoch": 3.19,
|
|
"learning_rate": 2.9014979573309126e-05,
|
|
"loss": 0.085,
|
|
"step": 7020
|
|
},
|
|
{
|
|
"epoch": 3.19,
|
|
"learning_rate": 2.894235133908307e-05,
|
|
"loss": 0.1044,
|
|
"step": 7030
|
|
},
|
|
{
|
|
"epoch": 3.2,
|
|
"learning_rate": 2.8869723104857016e-05,
|
|
"loss": 0.0782,
|
|
"step": 7040
|
|
},
|
|
{
|
|
"epoch": 3.2,
|
|
"learning_rate": 2.879709487063096e-05,
|
|
"loss": 0.0847,
|
|
"step": 7050
|
|
},
|
|
{
|
|
"epoch": 3.2,
|
|
"learning_rate": 2.8724466636404903e-05,
|
|
"loss": 0.0926,
|
|
"step": 7060
|
|
},
|
|
{
|
|
"epoch": 3.21,
|
|
"learning_rate": 2.865183840217885e-05,
|
|
"loss": 0.0853,
|
|
"step": 7070
|
|
},
|
|
{
|
|
"epoch": 3.21,
|
|
"learning_rate": 2.8579210167952793e-05,
|
|
"loss": 0.0882,
|
|
"step": 7080
|
|
},
|
|
{
|
|
"epoch": 3.22,
|
|
"learning_rate": 2.850658193372674e-05,
|
|
"loss": 0.076,
|
|
"step": 7090
|
|
},
|
|
{
|
|
"epoch": 3.22,
|
|
"learning_rate": 2.8433953699500683e-05,
|
|
"loss": 0.0812,
|
|
"step": 7100
|
|
},
|
|
{
|
|
"epoch": 3.22,
|
|
"eval_loss": 0.1491415947675705,
|
|
"eval_mean_accuracy": 0.7976677477854113,
|
|
"eval_mean_iou": 0.6869542568563782,
|
|
"eval_overall_accuracy": 0.9612365894103319,
|
|
"eval_per_category_accuracy": [
|
|
0.9923379642803694,
|
|
0.7278777182273029,
|
|
0.9126310250907891,
|
|
0.7221920781334781,
|
|
0.8433726507818777,
|
|
0.7604984123759599,
|
|
0.919574968594167,
|
|
0.8715464249006039,
|
|
0.2744438696326953,
|
|
0.7457094017094017,
|
|
0.6974266658797144,
|
|
0.9302344889753413,
|
|
0.8945470568134812,
|
|
0.8943737571053975,
|
|
0.8326812812553889,
|
|
0.8170092497430627,
|
|
0.8775275940914548,
|
|
0.6440348525469168
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9852574724071208,
|
|
0.6644151467675663,
|
|
0.8121287515033556,
|
|
0.6136114792831211,
|
|
0.7844025163654371,
|
|
0.6836819768065079,
|
|
0.8484344022153549,
|
|
0.6007086759264332,
|
|
0.24860583907399597,
|
|
0.5591632546931421,
|
|
0.5687937231243181,
|
|
0.8454285765616859,
|
|
0.8128664288836865,
|
|
0.8062410267393253,
|
|
0.7190475304103419,
|
|
0.7080619481835287,
|
|
0.7912495478701205,
|
|
0.3130783265997654
|
|
],
|
|
"eval_runtime": 82.047,
|
|
"eval_samples_per_second": 1.085,
|
|
"eval_steps_per_second": 0.146,
|
|
"step": 7100
|
|
},
|
|
{
|
|
"epoch": 3.23,
|
|
"learning_rate": 2.8361325465274626e-05,
|
|
"loss": 0.0896,
|
|
"step": 7110
|
|
},
|
|
{
|
|
"epoch": 3.23,
|
|
"learning_rate": 2.8288697231048573e-05,
|
|
"loss": 0.0718,
|
|
"step": 7120
|
|
},
|
|
{
|
|
"epoch": 3.24,
|
|
"learning_rate": 2.8216068996822516e-05,
|
|
"loss": 0.0919,
|
|
"step": 7130
|
|
},
|
|
{
|
|
"epoch": 3.24,
|
|
"learning_rate": 2.8143440762596463e-05,
|
|
"loss": 0.0891,
|
|
"step": 7140
|
|
},
|
|
{
|
|
"epoch": 3.25,
|
|
"learning_rate": 2.8070812528370406e-05,
|
|
"loss": 0.0797,
|
|
"step": 7150
|
|
},
|
|
{
|
|
"epoch": 3.25,
|
|
"learning_rate": 2.799818429414435e-05,
|
|
"loss": 0.0898,
|
|
"step": 7160
|
|
},
|
|
{
|
|
"epoch": 3.25,
|
|
"learning_rate": 2.7925556059918296e-05,
|
|
"loss": 0.0868,
|
|
"step": 7170
|
|
},
|
|
{
|
|
"epoch": 3.26,
|
|
"learning_rate": 2.785292782569224e-05,
|
|
"loss": 0.0898,
|
|
"step": 7180
|
|
},
|
|
{
|
|
"epoch": 3.26,
|
|
"learning_rate": 2.7780299591466183e-05,
|
|
"loss": 0.1017,
|
|
"step": 7190
|
|
},
|
|
{
|
|
"epoch": 3.27,
|
|
"learning_rate": 2.770767135724013e-05,
|
|
"loss": 0.0826,
|
|
"step": 7200
|
|
},
|
|
{
|
|
"epoch": 3.27,
|
|
"eval_loss": 0.1519315540790558,
|
|
"eval_mean_accuracy": 0.7974368292032994,
|
|
"eval_mean_iou": 0.6892930701953764,
|
|
"eval_overall_accuracy": 0.9600244157769707,
|
|
"eval_per_category_accuracy": [
|
|
0.9929770311994012,
|
|
0.7247850515714309,
|
|
0.9172547320347749,
|
|
0.7556972327726533,
|
|
0.8699738516439295,
|
|
0.7603394291619376,
|
|
0.8973420747685016,
|
|
0.7366657454122433,
|
|
0.34811174340403517,
|
|
0.7283223443223443,
|
|
0.7590647071553641,
|
|
0.9218172401386928,
|
|
0.8768636584356959,
|
|
0.8855921017897682,
|
|
0.8276642524573202,
|
|
0.8179342240493319,
|
|
0.9057095160658247,
|
|
0.6277479892761394
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9849399335448075,
|
|
0.6687882862692365,
|
|
0.8103227492089629,
|
|
0.6225276567214214,
|
|
0.7786314294329074,
|
|
0.6449094181523288,
|
|
0.8361890895412523,
|
|
0.5534463080553863,
|
|
0.2995059420483376,
|
|
0.5553801004066918,
|
|
0.5828524382137894,
|
|
0.8478996883443704,
|
|
0.8065883618015633,
|
|
0.8069498843423146,
|
|
0.7226446343207739,
|
|
0.7120386058501507,
|
|
0.8248469128239139,
|
|
0.3488138244385684
|
|
],
|
|
"eval_runtime": 83.9979,
|
|
"eval_samples_per_second": 1.06,
|
|
"eval_steps_per_second": 0.143,
|
|
"step": 7200
|
|
},
|
|
{
|
|
"epoch": 3.27,
|
|
"learning_rate": 2.7635043123014073e-05,
|
|
"loss": 0.096,
|
|
"step": 7210
|
|
},
|
|
{
|
|
"epoch": 3.28,
|
|
"learning_rate": 2.756241488878802e-05,
|
|
"loss": 0.0858,
|
|
"step": 7220
|
|
},
|
|
{
|
|
"epoch": 3.28,
|
|
"learning_rate": 2.7489786654561963e-05,
|
|
"loss": 0.0909,
|
|
"step": 7230
|
|
},
|
|
{
|
|
"epoch": 3.29,
|
|
"learning_rate": 2.7417158420335906e-05,
|
|
"loss": 0.0905,
|
|
"step": 7240
|
|
},
|
|
{
|
|
"epoch": 3.29,
|
|
"learning_rate": 2.7344530186109853e-05,
|
|
"loss": 0.0888,
|
|
"step": 7250
|
|
},
|
|
{
|
|
"epoch": 3.3,
|
|
"learning_rate": 2.7271901951883796e-05,
|
|
"loss": 0.0942,
|
|
"step": 7260
|
|
},
|
|
{
|
|
"epoch": 3.3,
|
|
"learning_rate": 2.7199273717657743e-05,
|
|
"loss": 0.0925,
|
|
"step": 7270
|
|
},
|
|
{
|
|
"epoch": 3.3,
|
|
"learning_rate": 2.7126645483431686e-05,
|
|
"loss": 0.0876,
|
|
"step": 7280
|
|
},
|
|
{
|
|
"epoch": 3.31,
|
|
"learning_rate": 2.705401724920563e-05,
|
|
"loss": 0.0889,
|
|
"step": 7290
|
|
},
|
|
{
|
|
"epoch": 3.31,
|
|
"learning_rate": 2.6981389014979577e-05,
|
|
"loss": 0.0791,
|
|
"step": 7300
|
|
},
|
|
{
|
|
"epoch": 3.31,
|
|
"eval_loss": 0.1503271609544754,
|
|
"eval_mean_accuracy": 0.7965757722351132,
|
|
"eval_mean_iou": 0.6894053238568253,
|
|
"eval_overall_accuracy": 0.9597292696492056,
|
|
"eval_per_category_accuracy": [
|
|
0.9920047575289915,
|
|
0.7369937984747162,
|
|
0.9152773192472763,
|
|
0.7202930005425936,
|
|
0.8684081064706718,
|
|
0.76895101992148,
|
|
0.9000302748749857,
|
|
0.7476033837133056,
|
|
0.3396275219865494,
|
|
0.7115115995115995,
|
|
0.7837159889039721,
|
|
0.9278227831085702,
|
|
0.8741299929859899,
|
|
0.8907572637840311,
|
|
0.8435613898948094,
|
|
0.7977068345323741,
|
|
0.9058267735973068,
|
|
0.614142091152815
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9851162074854626,
|
|
0.6692446810074841,
|
|
0.812920074221703,
|
|
0.6180527265320375,
|
|
0.779525915031291,
|
|
0.6397384766023269,
|
|
0.8374950821583133,
|
|
0.5406745694779389,
|
|
0.28814080056179775,
|
|
0.5568287797942116,
|
|
0.5911057032415067,
|
|
0.8447939465499367,
|
|
0.8142915780126991,
|
|
0.8119265319217196,
|
|
0.728554740041236,
|
|
0.7093116900177632,
|
|
0.8260610187806374,
|
|
0.35551330798479086
|
|
],
|
|
"eval_runtime": 83.7008,
|
|
"eval_samples_per_second": 1.063,
|
|
"eval_steps_per_second": 0.143,
|
|
"step": 7300
|
|
},
|
|
{
|
|
"epoch": 3.32,
|
|
"learning_rate": 2.690876078075352e-05,
|
|
"loss": 0.0941,
|
|
"step": 7310
|
|
},
|
|
{
|
|
"epoch": 3.32,
|
|
"learning_rate": 2.6836132546527463e-05,
|
|
"loss": 0.0751,
|
|
"step": 7320
|
|
},
|
|
{
|
|
"epoch": 3.33,
|
|
"learning_rate": 2.676350431230141e-05,
|
|
"loss": 0.0778,
|
|
"step": 7330
|
|
},
|
|
{
|
|
"epoch": 3.33,
|
|
"learning_rate": 2.6690876078075353e-05,
|
|
"loss": 0.0805,
|
|
"step": 7340
|
|
},
|
|
{
|
|
"epoch": 3.34,
|
|
"learning_rate": 2.66182478438493e-05,
|
|
"loss": 0.0864,
|
|
"step": 7350
|
|
},
|
|
{
|
|
"epoch": 3.34,
|
|
"learning_rate": 2.6545619609623243e-05,
|
|
"loss": 0.0833,
|
|
"step": 7360
|
|
},
|
|
{
|
|
"epoch": 3.35,
|
|
"learning_rate": 2.6472991375397187e-05,
|
|
"loss": 0.0852,
|
|
"step": 7370
|
|
},
|
|
{
|
|
"epoch": 3.35,
|
|
"learning_rate": 2.6400363141171133e-05,
|
|
"loss": 0.1008,
|
|
"step": 7380
|
|
},
|
|
{
|
|
"epoch": 3.35,
|
|
"learning_rate": 2.6327734906945077e-05,
|
|
"loss": 0.0867,
|
|
"step": 7390
|
|
},
|
|
{
|
|
"epoch": 3.36,
|
|
"learning_rate": 2.6255106672719023e-05,
|
|
"loss": 0.0871,
|
|
"step": 7400
|
|
},
|
|
{
|
|
"epoch": 3.36,
|
|
"eval_loss": 0.13909876346588135,
|
|
"eval_mean_accuracy": 0.7960482843801368,
|
|
"eval_mean_iou": 0.6890686903763231,
|
|
"eval_overall_accuracy": 0.9619064331054688,
|
|
"eval_per_category_accuracy": [
|
|
0.9923535512783945,
|
|
0.7248012435434512,
|
|
0.9101773137449103,
|
|
0.7408437330439501,
|
|
0.871394137581109,
|
|
0.7620529149130671,
|
|
0.9019198446585724,
|
|
0.8367790055148818,
|
|
0.2545783755819969,
|
|
0.7233601953601954,
|
|
0.7375907454405949,
|
|
0.9220307354121448,
|
|
0.8765983849792277,
|
|
0.89043169139257,
|
|
0.8468108725642353,
|
|
0.7999357656731757,
|
|
0.9095296430125311,
|
|
0.6276809651474531
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850871328712067,
|
|
0.6697238098087914,
|
|
0.8138968117903024,
|
|
0.6260675187711354,
|
|
0.7880510625834967,
|
|
0.685249495172118,
|
|
0.8415382517387154,
|
|
0.6059089975658216,
|
|
0.23402130492676432,
|
|
0.5607517681089186,
|
|
0.5906652487297649,
|
|
0.8481546675201117,
|
|
0.808253081224945,
|
|
0.8084169825564906,
|
|
0.7309461678365267,
|
|
0.7167424460431655,
|
|
0.8236273761921128,
|
|
0.2661343033334281
|
|
],
|
|
"eval_runtime": 84.5569,
|
|
"eval_samples_per_second": 1.053,
|
|
"eval_steps_per_second": 0.142,
|
|
"step": 7400
|
|
},
|
|
{
|
|
"epoch": 3.36,
|
|
"learning_rate": 2.6182478438492967e-05,
|
|
"loss": 0.0843,
|
|
"step": 7410
|
|
},
|
|
{
|
|
"epoch": 3.37,
|
|
"learning_rate": 2.610985020426691e-05,
|
|
"loss": 0.101,
|
|
"step": 7420
|
|
},
|
|
{
|
|
"epoch": 3.37,
|
|
"learning_rate": 2.6037221970040857e-05,
|
|
"loss": 0.0947,
|
|
"step": 7430
|
|
},
|
|
{
|
|
"epoch": 3.38,
|
|
"learning_rate": 2.59645937358148e-05,
|
|
"loss": 0.0794,
|
|
"step": 7440
|
|
},
|
|
{
|
|
"epoch": 3.38,
|
|
"learning_rate": 2.5891965501588744e-05,
|
|
"loss": 0.0911,
|
|
"step": 7450
|
|
},
|
|
{
|
|
"epoch": 3.39,
|
|
"learning_rate": 2.581933726736269e-05,
|
|
"loss": 0.0844,
|
|
"step": 7460
|
|
},
|
|
{
|
|
"epoch": 3.39,
|
|
"learning_rate": 2.5746709033136634e-05,
|
|
"loss": 0.0918,
|
|
"step": 7470
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"learning_rate": 2.567408079891058e-05,
|
|
"loss": 0.083,
|
|
"step": 7480
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"learning_rate": 2.5601452564684524e-05,
|
|
"loss": 0.0809,
|
|
"step": 7490
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"learning_rate": 2.5528824330458467e-05,
|
|
"loss": 0.0942,
|
|
"step": 7500
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"eval_loss": 0.15069787204265594,
|
|
"eval_mean_accuracy": 0.7945067044242303,
|
|
"eval_mean_iou": 0.6872757002457379,
|
|
"eval_overall_accuracy": 0.9596373311589274,
|
|
"eval_per_category_accuracy": [
|
|
0.9923054035407225,
|
|
0.7146326851147201,
|
|
0.917318353141851,
|
|
0.7225990233315247,
|
|
0.8705529440743544,
|
|
0.7270037405217299,
|
|
0.9133233809031657,
|
|
0.7338261141307784,
|
|
0.31764097258147955,
|
|
0.7381001221001221,
|
|
0.7534871825139192,
|
|
0.9268541471456868,
|
|
0.8925237846878765,
|
|
0.8954164686834556,
|
|
0.8394065787204691,
|
|
0.8224563206577595,
|
|
0.9096654148910893,
|
|
0.6140080428954423
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9849579847340046,
|
|
0.6687931870529761,
|
|
0.812854635323659,
|
|
0.6228224014965509,
|
|
0.7807169755593295,
|
|
0.6371301239838145,
|
|
0.8182895233225562,
|
|
0.5428083152727263,
|
|
0.27799157875673475,
|
|
0.5710971038152246,
|
|
0.5941007205404442,
|
|
0.8458027715742268,
|
|
0.8026979268179263,
|
|
0.8048221071904523,
|
|
0.7280090482933965,
|
|
0.7212545909285505,
|
|
0.8197395057059595,
|
|
0.3370741040547502
|
|
],
|
|
"eval_runtime": 81.7189,
|
|
"eval_samples_per_second": 1.089,
|
|
"eval_steps_per_second": 0.147,
|
|
"step": 7500
|
|
},
|
|
{
|
|
"epoch": 3.41,
|
|
"learning_rate": 2.5456196096232414e-05,
|
|
"loss": 0.0807,
|
|
"step": 7510
|
|
},
|
|
{
|
|
"epoch": 3.41,
|
|
"learning_rate": 2.5383567862006357e-05,
|
|
"loss": 0.0887,
|
|
"step": 7520
|
|
},
|
|
{
|
|
"epoch": 3.42,
|
|
"learning_rate": 2.5310939627780304e-05,
|
|
"loss": 0.0824,
|
|
"step": 7530
|
|
},
|
|
{
|
|
"epoch": 3.42,
|
|
"learning_rate": 2.5238311393554247e-05,
|
|
"loss": 0.0711,
|
|
"step": 7540
|
|
},
|
|
{
|
|
"epoch": 3.43,
|
|
"learning_rate": 2.516568315932819e-05,
|
|
"loss": 0.0764,
|
|
"step": 7550
|
|
},
|
|
{
|
|
"epoch": 3.43,
|
|
"learning_rate": 2.5093054925102137e-05,
|
|
"loss": 0.0866,
|
|
"step": 7560
|
|
},
|
|
{
|
|
"epoch": 3.44,
|
|
"learning_rate": 2.502042669087608e-05,
|
|
"loss": 0.0827,
|
|
"step": 7570
|
|
},
|
|
{
|
|
"epoch": 3.44,
|
|
"learning_rate": 2.4947798456650024e-05,
|
|
"loss": 0.0825,
|
|
"step": 7580
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"learning_rate": 2.487517022242397e-05,
|
|
"loss": 0.0796,
|
|
"step": 7590
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"learning_rate": 2.4802541988197914e-05,
|
|
"loss": 0.0941,
|
|
"step": 7600
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"eval_loss": 0.14226721227169037,
|
|
"eval_mean_accuracy": 0.8043973312375484,
|
|
"eval_mean_iou": 0.6950231914394323,
|
|
"eval_overall_accuracy": 0.9625836918863018,
|
|
"eval_per_category_accuracy": [
|
|
0.992655972321929,
|
|
0.7354879450768309,
|
|
0.9027353637063937,
|
|
0.7259224091155725,
|
|
0.8658872704642963,
|
|
0.7484466015129903,
|
|
0.9037432968990873,
|
|
0.868287713175564,
|
|
0.33507501293326436,
|
|
0.752986568986569,
|
|
0.7652029353321923,
|
|
0.9223391174737974,
|
|
0.8734151035016097,
|
|
0.8883418686095419,
|
|
0.8428770046559751,
|
|
0.8087808324768756,
|
|
0.9187497878564398,
|
|
0.6282171581769437
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850488257103458,
|
|
0.6723854636962475,
|
|
0.8136399633630944,
|
|
0.6180274858528698,
|
|
0.7959631356402571,
|
|
0.682059650916174,
|
|
0.8380410879116093,
|
|
0.6294282638927081,
|
|
0.3037992495309568,
|
|
0.5688573705649684,
|
|
0.5968892913047482,
|
|
0.8476645531675235,
|
|
0.8067695246796935,
|
|
0.8121415183132356,
|
|
0.730002193586327,
|
|
0.7154399940906069,
|
|
0.8157739717957921,
|
|
0.2784859018926226
|
|
],
|
|
"eval_runtime": 82.5733,
|
|
"eval_samples_per_second": 1.078,
|
|
"eval_steps_per_second": 0.145,
|
|
"step": 7600
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"learning_rate": 2.472991375397186e-05,
|
|
"loss": 0.0845,
|
|
"step": 7610
|
|
},
|
|
{
|
|
"epoch": 3.46,
|
|
"learning_rate": 2.4657285519745804e-05,
|
|
"loss": 0.0874,
|
|
"step": 7620
|
|
},
|
|
{
|
|
"epoch": 3.46,
|
|
"learning_rate": 2.4584657285519747e-05,
|
|
"loss": 0.0838,
|
|
"step": 7630
|
|
},
|
|
{
|
|
"epoch": 3.47,
|
|
"learning_rate": 2.4512029051293694e-05,
|
|
"loss": 0.0739,
|
|
"step": 7640
|
|
},
|
|
{
|
|
"epoch": 3.47,
|
|
"learning_rate": 2.4439400817067637e-05,
|
|
"loss": 0.0805,
|
|
"step": 7650
|
|
},
|
|
{
|
|
"epoch": 3.48,
|
|
"learning_rate": 2.4366772582841584e-05,
|
|
"loss": 0.0998,
|
|
"step": 7660
|
|
},
|
|
{
|
|
"epoch": 3.48,
|
|
"learning_rate": 2.4294144348615527e-05,
|
|
"loss": 0.0787,
|
|
"step": 7670
|
|
},
|
|
{
|
|
"epoch": 3.49,
|
|
"learning_rate": 2.422151611438947e-05,
|
|
"loss": 0.1046,
|
|
"step": 7680
|
|
},
|
|
{
|
|
"epoch": 3.49,
|
|
"learning_rate": 2.4148887880163417e-05,
|
|
"loss": 0.0791,
|
|
"step": 7690
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"learning_rate": 2.407625964593736e-05,
|
|
"loss": 0.1024,
|
|
"step": 7700
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"eval_loss": 0.1411096453666687,
|
|
"eval_mean_accuracy": 0.793477814813021,
|
|
"eval_mean_iou": 0.6876190052060385,
|
|
"eval_overall_accuracy": 0.9597613302509437,
|
|
"eval_per_category_accuracy": [
|
|
0.9928391888859046,
|
|
0.7323790864489386,
|
|
0.913808875316386,
|
|
0.7333830710797613,
|
|
0.8687285284671272,
|
|
0.7570427355711693,
|
|
0.9060295719411067,
|
|
0.7227078377074855,
|
|
0.29130884635281945,
|
|
0.7281269841269842,
|
|
0.7375809085364654,
|
|
0.9135146461711204,
|
|
0.8883333633077352,
|
|
0.9011139855340267,
|
|
0.8490149163648905,
|
|
0.8106821685508736,
|
|
0.9079898664807003,
|
|
0.6280160857908847
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9846414923408805,
|
|
0.6744553628677512,
|
|
0.8143157468925446,
|
|
0.6213296558064701,
|
|
0.7800849995194287,
|
|
0.6416433351798895,
|
|
0.8314123085029099,
|
|
0.5326802302281645,
|
|
0.2638212143928036,
|
|
0.5724928190713178,
|
|
0.5924401286315906,
|
|
0.8512147948939932,
|
|
0.8182251137827216,
|
|
0.8160195064423851,
|
|
0.7297326089272398,
|
|
0.711005323793696,
|
|
0.8306726852361555,
|
|
0.3109547671987522
|
|
],
|
|
"eval_runtime": 79.8015,
|
|
"eval_samples_per_second": 1.115,
|
|
"eval_steps_per_second": 0.15,
|
|
"step": 7700
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"learning_rate": 2.4003631411711304e-05,
|
|
"loss": 0.0825,
|
|
"step": 7710
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"learning_rate": 2.393100317748525e-05,
|
|
"loss": 0.0873,
|
|
"step": 7720
|
|
},
|
|
{
|
|
"epoch": 3.51,
|
|
"learning_rate": 2.3858374943259194e-05,
|
|
"loss": 0.0785,
|
|
"step": 7730
|
|
},
|
|
{
|
|
"epoch": 3.51,
|
|
"learning_rate": 2.378574670903314e-05,
|
|
"loss": 0.0828,
|
|
"step": 7740
|
|
},
|
|
{
|
|
"epoch": 3.52,
|
|
"learning_rate": 2.3713118474807084e-05,
|
|
"loss": 0.0728,
|
|
"step": 7750
|
|
},
|
|
{
|
|
"epoch": 3.52,
|
|
"learning_rate": 2.3640490240581028e-05,
|
|
"loss": 0.0829,
|
|
"step": 7760
|
|
},
|
|
{
|
|
"epoch": 3.53,
|
|
"learning_rate": 2.3567862006354974e-05,
|
|
"loss": 0.0822,
|
|
"step": 7770
|
|
},
|
|
{
|
|
"epoch": 3.53,
|
|
"learning_rate": 2.3495233772128918e-05,
|
|
"loss": 0.093,
|
|
"step": 7780
|
|
},
|
|
{
|
|
"epoch": 3.54,
|
|
"learning_rate": 2.3422605537902864e-05,
|
|
"loss": 0.087,
|
|
"step": 7790
|
|
},
|
|
{
|
|
"epoch": 3.54,
|
|
"learning_rate": 2.3349977303676808e-05,
|
|
"loss": 0.0898,
|
|
"step": 7800
|
|
},
|
|
{
|
|
"epoch": 3.54,
|
|
"eval_loss": 0.14239805936813354,
|
|
"eval_mean_accuracy": 0.7949254811371873,
|
|
"eval_mean_iou": 0.6902498843796094,
|
|
"eval_overall_accuracy": 0.9612459761373112,
|
|
"eval_per_category_accuracy": [
|
|
0.993142109157134,
|
|
0.731925711232371,
|
|
0.9112519946076812,
|
|
0.6812262615301139,
|
|
0.8621513009338895,
|
|
0.7472674760089913,
|
|
0.8922179651804139,
|
|
0.8149019225315349,
|
|
0.3208484221417486,
|
|
0.7321514041514041,
|
|
0.6952527100670877,
|
|
0.9328992262773146,
|
|
0.9065652932395734,
|
|
0.8931110641817573,
|
|
0.8570928177271944,
|
|
0.8205421377183967,
|
|
0.8944620502170807,
|
|
0.6216487935656837
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.985075545768495,
|
|
0.6730743459551214,
|
|
0.8130401231016586,
|
|
0.6100583090379009,
|
|
0.7897747164235203,
|
|
0.6521364622636056,
|
|
0.8290202684678771,
|
|
0.5880449815488054,
|
|
0.2901520467836257,
|
|
0.5639794735970866,
|
|
0.580207691991955,
|
|
0.8465515261972073,
|
|
0.8064788371851176,
|
|
0.8109337860780985,
|
|
0.735473101075587,
|
|
0.7208550355794571,
|
|
0.8313334212835765,
|
|
0.29830824649427506
|
|
],
|
|
"eval_runtime": 84.1376,
|
|
"eval_samples_per_second": 1.058,
|
|
"eval_steps_per_second": 0.143,
|
|
"step": 7800
|
|
},
|
|
{
|
|
"epoch": 3.55,
|
|
"learning_rate": 2.327734906945075e-05,
|
|
"loss": 0.0771,
|
|
"step": 7810
|
|
},
|
|
{
|
|
"epoch": 3.55,
|
|
"learning_rate": 2.3204720835224698e-05,
|
|
"loss": 0.08,
|
|
"step": 7820
|
|
},
|
|
{
|
|
"epoch": 3.55,
|
|
"learning_rate": 2.313209260099864e-05,
|
|
"loss": 0.0957,
|
|
"step": 7830
|
|
},
|
|
{
|
|
"epoch": 3.56,
|
|
"learning_rate": 2.3059464366772584e-05,
|
|
"loss": 0.0752,
|
|
"step": 7840
|
|
},
|
|
{
|
|
"epoch": 3.56,
|
|
"learning_rate": 2.298683613254653e-05,
|
|
"loss": 0.0889,
|
|
"step": 7850
|
|
},
|
|
{
|
|
"epoch": 3.57,
|
|
"learning_rate": 2.2914207898320474e-05,
|
|
"loss": 0.0747,
|
|
"step": 7860
|
|
},
|
|
{
|
|
"epoch": 3.57,
|
|
"learning_rate": 2.284157966409442e-05,
|
|
"loss": 0.0913,
|
|
"step": 7870
|
|
},
|
|
{
|
|
"epoch": 3.58,
|
|
"learning_rate": 2.2768951429868365e-05,
|
|
"loss": 0.0857,
|
|
"step": 7880
|
|
},
|
|
{
|
|
"epoch": 3.58,
|
|
"learning_rate": 2.2696323195642308e-05,
|
|
"loss": 0.0987,
|
|
"step": 7890
|
|
},
|
|
{
|
|
"epoch": 3.59,
|
|
"learning_rate": 2.2623694961416255e-05,
|
|
"loss": 0.0784,
|
|
"step": 7900
|
|
},
|
|
{
|
|
"epoch": 3.59,
|
|
"eval_loss": 0.1483561098575592,
|
|
"eval_mean_accuracy": 0.7962622662878549,
|
|
"eval_mean_iou": 0.6917747207606494,
|
|
"eval_overall_accuracy": 0.959908646144224,
|
|
"eval_per_category_accuracy": [
|
|
0.9926994606011167,
|
|
0.7252870027040593,
|
|
0.9169589798613403,
|
|
0.7194791101465002,
|
|
0.864814851662455,
|
|
0.7613750281532775,
|
|
0.9122063424123159,
|
|
0.7299171051657084,
|
|
0.3656492498706674,
|
|
0.7317802197802198,
|
|
0.735928308642704,
|
|
0.9242170851569388,
|
|
0.8911389673219071,
|
|
0.8768940393854601,
|
|
0.8574700379375755,
|
|
0.8056333504624872,
|
|
0.9110169622276463,
|
|
0.610254691689008
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850454586836164,
|
|
0.6737309167481387,
|
|
0.814432328415651,
|
|
0.6164574616457462,
|
|
0.7742703745339059,
|
|
0.6492239851933024,
|
|
0.8477682606776312,
|
|
0.5281189348145727,
|
|
0.32247467834656446,
|
|
0.5707624794295117,
|
|
0.595128431536326,
|
|
0.8513640982310973,
|
|
0.8199196627656362,
|
|
0.8131956457877729,
|
|
0.7408912914926409,
|
|
0.7208103448275862,
|
|
0.8240899689329169,
|
|
0.3042606516290727
|
|
],
|
|
"eval_runtime": 81.0689,
|
|
"eval_samples_per_second": 1.098,
|
|
"eval_steps_per_second": 0.148,
|
|
"step": 7900
|
|
},
|
|
{
|
|
"epoch": 3.59,
|
|
"learning_rate": 2.2551066727190198e-05,
|
|
"loss": 0.0839,
|
|
"step": 7910
|
|
},
|
|
{
|
|
"epoch": 3.6,
|
|
"learning_rate": 2.2478438492964145e-05,
|
|
"loss": 0.0772,
|
|
"step": 7920
|
|
},
|
|
{
|
|
"epoch": 3.6,
|
|
"learning_rate": 2.2405810258738088e-05,
|
|
"loss": 0.0883,
|
|
"step": 7930
|
|
},
|
|
{
|
|
"epoch": 3.6,
|
|
"learning_rate": 2.233318202451203e-05,
|
|
"loss": 0.0827,
|
|
"step": 7940
|
|
},
|
|
{
|
|
"epoch": 3.61,
|
|
"learning_rate": 2.2260553790285978e-05,
|
|
"loss": 0.0916,
|
|
"step": 7950
|
|
},
|
|
{
|
|
"epoch": 3.61,
|
|
"learning_rate": 2.218792555605992e-05,
|
|
"loss": 0.0903,
|
|
"step": 7960
|
|
},
|
|
{
|
|
"epoch": 3.62,
|
|
"learning_rate": 2.2115297321833865e-05,
|
|
"loss": 0.1006,
|
|
"step": 7970
|
|
},
|
|
{
|
|
"epoch": 3.62,
|
|
"learning_rate": 2.204266908760781e-05,
|
|
"loss": 0.0864,
|
|
"step": 7980
|
|
},
|
|
{
|
|
"epoch": 3.63,
|
|
"learning_rate": 2.1970040853381755e-05,
|
|
"loss": 0.1101,
|
|
"step": 7990
|
|
},
|
|
{
|
|
"epoch": 3.63,
|
|
"learning_rate": 2.18974126191557e-05,
|
|
"loss": 0.084,
|
|
"step": 8000
|
|
},
|
|
{
|
|
"epoch": 3.63,
|
|
"eval_loss": 0.14724896848201752,
|
|
"eval_mean_accuracy": 0.801441710329252,
|
|
"eval_mean_iou": 0.6890652338941418,
|
|
"eval_overall_accuracy": 0.9606205372328169,
|
|
"eval_per_category_accuracy": [
|
|
0.9925550728578445,
|
|
0.7332534529380333,
|
|
0.91065189281391,
|
|
0.7396907216494846,
|
|
0.8666385811412101,
|
|
0.7315413864219503,
|
|
0.9267191431166418,
|
|
0.7671664923508787,
|
|
0.3653388515261252,
|
|
0.7739291819291819,
|
|
0.7120836530327176,
|
|
0.9230744900823539,
|
|
0.9066102548423646,
|
|
0.863554370589374,
|
|
0.8586016985687187,
|
|
0.805511305241521,
|
|
0.9154511483523774,
|
|
0.6335790884718498
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9853020987467782,
|
|
0.6722732738528229,
|
|
0.8148160684339277,
|
|
0.6205052344105598,
|
|
0.7830637741753895,
|
|
0.6515298115611984,
|
|
0.8499542003249064,
|
|
0.5557586789447995,
|
|
0.31802215617400703,
|
|
0.5738465995509524,
|
|
0.5848103924641709,
|
|
0.8502836998244631,
|
|
0.8176778777138871,
|
|
0.811335105261635,
|
|
0.738910247789006,
|
|
0.7220123903180489,
|
|
0.8094385799294439,
|
|
0.2436340206185567
|
|
],
|
|
"eval_runtime": 81.6433,
|
|
"eval_samples_per_second": 1.09,
|
|
"eval_steps_per_second": 0.147,
|
|
"step": 8000
|
|
},
|
|
{
|
|
"epoch": 3.64,
|
|
"learning_rate": 2.1824784384929645e-05,
|
|
"loss": 0.0765,
|
|
"step": 8010
|
|
},
|
|
{
|
|
"epoch": 3.64,
|
|
"learning_rate": 2.1752156150703588e-05,
|
|
"loss": 0.0851,
|
|
"step": 8020
|
|
},
|
|
{
|
|
"epoch": 3.65,
|
|
"learning_rate": 2.1679527916477535e-05,
|
|
"loss": 0.0775,
|
|
"step": 8030
|
|
},
|
|
{
|
|
"epoch": 3.65,
|
|
"learning_rate": 2.1606899682251478e-05,
|
|
"loss": 0.0853,
|
|
"step": 8040
|
|
},
|
|
{
|
|
"epoch": 3.65,
|
|
"learning_rate": 2.1534271448025425e-05,
|
|
"loss": 0.0858,
|
|
"step": 8050
|
|
},
|
|
{
|
|
"epoch": 3.66,
|
|
"learning_rate": 2.1461643213799368e-05,
|
|
"loss": 0.102,
|
|
"step": 8060
|
|
},
|
|
{
|
|
"epoch": 3.66,
|
|
"learning_rate": 2.138901497957331e-05,
|
|
"loss": 0.0888,
|
|
"step": 8070
|
|
},
|
|
{
|
|
"epoch": 3.67,
|
|
"learning_rate": 2.131638674534726e-05,
|
|
"loss": 0.0821,
|
|
"step": 8080
|
|
},
|
|
{
|
|
"epoch": 3.67,
|
|
"learning_rate": 2.12437585111212e-05,
|
|
"loss": 0.0951,
|
|
"step": 8090
|
|
},
|
|
{
|
|
"epoch": 3.68,
|
|
"learning_rate": 2.1171130276895145e-05,
|
|
"loss": 0.076,
|
|
"step": 8100
|
|
},
|
|
{
|
|
"epoch": 3.68,
|
|
"eval_loss": 0.14653557538986206,
|
|
"eval_mean_accuracy": 0.7983003293456227,
|
|
"eval_mean_iou": 0.6930321233038834,
|
|
"eval_overall_accuracy": 0.961235389280855,
|
|
"eval_per_category_accuracy": [
|
|
0.9925582901029885,
|
|
0.724315484382843,
|
|
0.920943036755805,
|
|
0.7198182311448725,
|
|
0.8697680131023093,
|
|
0.7394044312154708,
|
|
0.9183552739180212,
|
|
0.7878025914345015,
|
|
0.3520434557682359,
|
|
0.718046398046398,
|
|
0.743561746247221,
|
|
0.9272811376925906,
|
|
0.8918088952034963,
|
|
0.8836518663217359,
|
|
0.8556971029487843,
|
|
0.8165274922918808,
|
|
0.8969645727968698,
|
|
0.6108579088471849
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850240593046129,
|
|
0.672747507256403,
|
|
0.812150004624905,
|
|
0.6193394024276377,
|
|
0.786230670566666,
|
|
0.6648691149374166,
|
|
0.8268529426508924,
|
|
0.577714664946371,
|
|
0.3062418433013816,
|
|
0.5654745878751049,
|
|
0.5948891896996789,
|
|
0.8507597494223438,
|
|
0.8241910753389651,
|
|
0.8191116675706869,
|
|
0.7335596978726353,
|
|
0.7211002825019004,
|
|
0.8257213791849651,
|
|
0.28860037998733373
|
|
],
|
|
"eval_runtime": 79.5832,
|
|
"eval_samples_per_second": 1.118,
|
|
"eval_steps_per_second": 0.151,
|
|
"step": 8100
|
|
},
|
|
{
|
|
"epoch": 3.68,
|
|
"learning_rate": 2.1098502042669092e-05,
|
|
"loss": 0.0831,
|
|
"step": 8110
|
|
},
|
|
{
|
|
"epoch": 3.69,
|
|
"learning_rate": 2.1025873808443035e-05,
|
|
"loss": 0.0771,
|
|
"step": 8120
|
|
},
|
|
{
|
|
"epoch": 3.69,
|
|
"learning_rate": 2.0953245574216982e-05,
|
|
"loss": 0.1087,
|
|
"step": 8130
|
|
},
|
|
{
|
|
"epoch": 3.69,
|
|
"learning_rate": 2.0880617339990925e-05,
|
|
"loss": 0.0749,
|
|
"step": 8140
|
|
},
|
|
{
|
|
"epoch": 3.7,
|
|
"learning_rate": 2.080798910576487e-05,
|
|
"loss": 0.0833,
|
|
"step": 8150
|
|
},
|
|
{
|
|
"epoch": 3.7,
|
|
"learning_rate": 2.0735360871538815e-05,
|
|
"loss": 0.075,
|
|
"step": 8160
|
|
},
|
|
{
|
|
"epoch": 3.71,
|
|
"learning_rate": 2.066273263731276e-05,
|
|
"loss": 0.0843,
|
|
"step": 8170
|
|
},
|
|
{
|
|
"epoch": 3.71,
|
|
"learning_rate": 2.0590104403086705e-05,
|
|
"loss": 0.0824,
|
|
"step": 8180
|
|
},
|
|
{
|
|
"epoch": 3.72,
|
|
"learning_rate": 2.051747616886065e-05,
|
|
"loss": 0.0899,
|
|
"step": 8190
|
|
},
|
|
{
|
|
"epoch": 3.72,
|
|
"learning_rate": 2.0444847934634592e-05,
|
|
"loss": 0.0821,
|
|
"step": 8200
|
|
},
|
|
{
|
|
"epoch": 3.72,
|
|
"eval_loss": 0.14239969849586487,
|
|
"eval_mean_accuracy": 0.7946176582333029,
|
|
"eval_mean_iou": 0.6883602339194399,
|
|
"eval_overall_accuracy": 0.960033716780416,
|
|
"eval_per_category_accuracy": [
|
|
0.9926663451640313,
|
|
0.7392120986414935,
|
|
0.9139017277429294,
|
|
0.6989962018448183,
|
|
0.8748522250803284,
|
|
0.7411885761728324,
|
|
0.92021526480076,
|
|
0.7268932229995683,
|
|
0.34604242110708744,
|
|
0.7233211233211233,
|
|
0.7549233705168309,
|
|
0.9279572060585214,
|
|
0.8988498822006007,
|
|
0.8787154849268782,
|
|
0.8400478530781169,
|
|
0.803365878725591,
|
|
0.898510520777726,
|
|
0.6234584450402145
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9851649722468053,
|
|
0.669850632391349,
|
|
0.8142412868632708,
|
|
0.6157614865268567,
|
|
0.7849918610693152,
|
|
0.6346062589447933,
|
|
0.826438002968982,
|
|
0.5387425863200032,
|
|
0.3039487435815877,
|
|
0.5563360705618957,
|
|
0.5873385171125941,
|
|
0.84827840442949,
|
|
0.8220391950459304,
|
|
0.8170482763134749,
|
|
0.7302443891676152,
|
|
0.719451443297783,
|
|
0.8246380669717703,
|
|
0.3113640167364017
|
|
],
|
|
"eval_runtime": 80.7089,
|
|
"eval_samples_per_second": 1.103,
|
|
"eval_steps_per_second": 0.149,
|
|
"step": 8200
|
|
},
|
|
{
|
|
"epoch": 3.73,
|
|
"learning_rate": 2.037221970040854e-05,
|
|
"loss": 0.0822,
|
|
"step": 8210
|
|
},
|
|
{
|
|
"epoch": 3.73,
|
|
"learning_rate": 2.0299591466182482e-05,
|
|
"loss": 0.0858,
|
|
"step": 8220
|
|
},
|
|
{
|
|
"epoch": 3.74,
|
|
"learning_rate": 2.0226963231956425e-05,
|
|
"loss": 0.0896,
|
|
"step": 8230
|
|
},
|
|
{
|
|
"epoch": 3.74,
|
|
"learning_rate": 2.0154334997730372e-05,
|
|
"loss": 0.0871,
|
|
"step": 8240
|
|
},
|
|
{
|
|
"epoch": 3.74,
|
|
"learning_rate": 2.0081706763504315e-05,
|
|
"loss": 0.0911,
|
|
"step": 8250
|
|
},
|
|
{
|
|
"epoch": 3.75,
|
|
"learning_rate": 2.0009078529278262e-05,
|
|
"loss": 0.082,
|
|
"step": 8260
|
|
},
|
|
{
|
|
"epoch": 3.75,
|
|
"learning_rate": 1.9936450295052202e-05,
|
|
"loss": 0.0837,
|
|
"step": 8270
|
|
},
|
|
{
|
|
"epoch": 3.76,
|
|
"learning_rate": 1.986382206082615e-05,
|
|
"loss": 0.0807,
|
|
"step": 8280
|
|
},
|
|
{
|
|
"epoch": 3.76,
|
|
"learning_rate": 1.9791193826600092e-05,
|
|
"loss": 0.0798,
|
|
"step": 8290
|
|
},
|
|
{
|
|
"epoch": 3.77,
|
|
"learning_rate": 1.9718565592374035e-05,
|
|
"loss": 0.0924,
|
|
"step": 8300
|
|
},
|
|
{
|
|
"epoch": 3.77,
|
|
"eval_loss": 0.14418603479862213,
|
|
"eval_mean_accuracy": 0.7959393875842294,
|
|
"eval_mean_iou": 0.687513928062221,
|
|
"eval_overall_accuracy": 0.960693445098534,
|
|
"eval_per_category_accuracy": [
|
|
0.9926985730852149,
|
|
0.7343545070354118,
|
|
0.9097921481236931,
|
|
0.7383342376559957,
|
|
0.8746937294032809,
|
|
0.7354629723678342,
|
|
0.9263032985694252,
|
|
0.7636097277241584,
|
|
0.3122090015519917,
|
|
0.7296019536019536,
|
|
0.727429223474788,
|
|
0.9245531425318168,
|
|
0.8809506681294175,
|
|
0.8907264663956497,
|
|
0.8506639075702708,
|
|
0.8141957862281604,
|
|
0.8894261478123756,
|
|
0.6319034852546916
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9851814454753052,
|
|
0.6659276117759342,
|
|
0.8151579764714583,
|
|
0.6203202461678728,
|
|
0.7890716829217115,
|
|
0.6438546028137649,
|
|
0.8253026803292666,
|
|
0.5633787612648031,
|
|
0.2740316941379467,
|
|
0.5648205562529303,
|
|
0.5850580714579575,
|
|
0.8508184374920412,
|
|
0.8190878307763053,
|
|
0.8183537933571281,
|
|
0.735997463621147,
|
|
0.7236056402352001,
|
|
0.813958545125946,
|
|
0.2813236654432608
|
|
],
|
|
"eval_runtime": 82.9377,
|
|
"eval_samples_per_second": 1.073,
|
|
"eval_steps_per_second": 0.145,
|
|
"step": 8300
|
|
},
|
|
{
|
|
"epoch": 3.77,
|
|
"learning_rate": 1.9645937358147982e-05,
|
|
"loss": 0.0788,
|
|
"step": 8310
|
|
},
|
|
{
|
|
"epoch": 3.78,
|
|
"learning_rate": 1.9573309123921926e-05,
|
|
"loss": 0.0843,
|
|
"step": 8320
|
|
},
|
|
{
|
|
"epoch": 3.78,
|
|
"learning_rate": 1.9500680889695872e-05,
|
|
"loss": 0.0696,
|
|
"step": 8330
|
|
},
|
|
{
|
|
"epoch": 3.79,
|
|
"learning_rate": 1.9428052655469816e-05,
|
|
"loss": 0.0742,
|
|
"step": 8340
|
|
},
|
|
{
|
|
"epoch": 3.79,
|
|
"learning_rate": 1.935542442124376e-05,
|
|
"loss": 0.0831,
|
|
"step": 8350
|
|
},
|
|
{
|
|
"epoch": 3.79,
|
|
"learning_rate": 1.9282796187017706e-05,
|
|
"loss": 0.0905,
|
|
"step": 8360
|
|
},
|
|
{
|
|
"epoch": 3.8,
|
|
"learning_rate": 1.921016795279165e-05,
|
|
"loss": 0.0997,
|
|
"step": 8370
|
|
},
|
|
{
|
|
"epoch": 3.8,
|
|
"learning_rate": 1.9137539718565596e-05,
|
|
"loss": 0.0804,
|
|
"step": 8380
|
|
},
|
|
{
|
|
"epoch": 3.81,
|
|
"learning_rate": 1.906491148433954e-05,
|
|
"loss": 0.0944,
|
|
"step": 8390
|
|
},
|
|
{
|
|
"epoch": 3.81,
|
|
"learning_rate": 1.8992283250113482e-05,
|
|
"loss": 0.0804,
|
|
"step": 8400
|
|
},
|
|
{
|
|
"epoch": 3.81,
|
|
"eval_loss": 0.14236502349376678,
|
|
"eval_mean_accuracy": 0.7955755743850114,
|
|
"eval_mean_iou": 0.6886094268621761,
|
|
"eval_overall_accuracy": 0.9601393281743767,
|
|
"eval_per_category_accuracy": [
|
|
0.9924674306625428,
|
|
0.7279586780874042,
|
|
0.9179425277869484,
|
|
0.7224633749321758,
|
|
0.8754855216600467,
|
|
0.7505155913954752,
|
|
0.9079417608841656,
|
|
0.7424678418982898,
|
|
0.32928091050181063,
|
|
0.7566495726495727,
|
|
0.7350823348875641,
|
|
0.9229637888294528,
|
|
0.8719223782889413,
|
|
0.8943077627017229,
|
|
0.8335111657182273,
|
|
0.8231564748201439,
|
|
0.8926507299281334,
|
|
0.6235924932975871
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9852852475973827,
|
|
0.673134797646319,
|
|
0.8135391236235183,
|
|
0.6244577324422559,
|
|
0.7840482873547397,
|
|
0.648832112886846,
|
|
0.8229458092964135,
|
|
0.5457687568382923,
|
|
0.28741081910954575,
|
|
0.5683365616012209,
|
|
0.5921737683353012,
|
|
0.8525880990902484,
|
|
0.8006655491606318,
|
|
0.7989317129538648,
|
|
0.7314250855921464,
|
|
0.7214093910615469,
|
|
0.8279734850653142,
|
|
0.31604334386358235
|
|
],
|
|
"eval_runtime": 86.7054,
|
|
"eval_samples_per_second": 1.026,
|
|
"eval_steps_per_second": 0.138,
|
|
"step": 8400
|
|
},
|
|
{
|
|
"epoch": 3.82,
|
|
"learning_rate": 1.891965501588743e-05,
|
|
"loss": 0.1009,
|
|
"step": 8410
|
|
},
|
|
{
|
|
"epoch": 3.82,
|
|
"learning_rate": 1.8847026781661372e-05,
|
|
"loss": 0.0694,
|
|
"step": 8420
|
|
},
|
|
{
|
|
"epoch": 3.83,
|
|
"learning_rate": 1.8774398547435316e-05,
|
|
"loss": 0.0786,
|
|
"step": 8430
|
|
},
|
|
{
|
|
"epoch": 3.83,
|
|
"learning_rate": 1.8701770313209262e-05,
|
|
"loss": 0.0801,
|
|
"step": 8440
|
|
},
|
|
{
|
|
"epoch": 3.84,
|
|
"learning_rate": 1.8629142078983206e-05,
|
|
"loss": 0.0862,
|
|
"step": 8450
|
|
},
|
|
{
|
|
"epoch": 3.84,
|
|
"learning_rate": 1.8556513844757153e-05,
|
|
"loss": 0.0799,
|
|
"step": 8460
|
|
},
|
|
{
|
|
"epoch": 3.84,
|
|
"learning_rate": 1.8483885610531096e-05,
|
|
"loss": 0.0841,
|
|
"step": 8470
|
|
},
|
|
{
|
|
"epoch": 3.85,
|
|
"learning_rate": 1.841125737630504e-05,
|
|
"loss": 0.0777,
|
|
"step": 8480
|
|
},
|
|
{
|
|
"epoch": 3.85,
|
|
"learning_rate": 1.8338629142078986e-05,
|
|
"loss": 0.0887,
|
|
"step": 8490
|
|
},
|
|
{
|
|
"epoch": 3.86,
|
|
"learning_rate": 1.826600090785293e-05,
|
|
"loss": 0.0896,
|
|
"step": 8500
|
|
},
|
|
{
|
|
"epoch": 3.86,
|
|
"eval_loss": 0.1432206928730011,
|
|
"eval_mean_accuracy": 0.8011006689138838,
|
|
"eval_mean_iou": 0.6992310087988485,
|
|
"eval_overall_accuracy": 0.9620774086941494,
|
|
"eval_per_category_accuracy": [
|
|
0.9925154674607272,
|
|
0.7424990689616089,
|
|
0.9183053400462199,
|
|
0.7421323928377646,
|
|
0.8784132318503581,
|
|
0.752926836808147,
|
|
0.8923345408484621,
|
|
0.8249977871830027,
|
|
0.3462493533367822,
|
|
0.7310476190476191,
|
|
0.7342461980365539,
|
|
0.9241775489951884,
|
|
0.9014351743610957,
|
|
0.8902073104200764,
|
|
0.8597387480600104,
|
|
0.8217754367934224,
|
|
0.8713006020248524,
|
|
0.5955093833780161
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9850162649833359,
|
|
0.6722866483895087,
|
|
0.8149722191278287,
|
|
0.6246503396700348,
|
|
0.7913935183725647,
|
|
0.6781338313177295,
|
|
0.8403050858565998,
|
|
0.5984890775903867,
|
|
0.29448257655755017,
|
|
0.5724327301096817,
|
|
0.5893751085703457,
|
|
0.8525572523260182,
|
|
0.8102996035194219,
|
|
0.809072915208829,
|
|
0.7390627605759075,
|
|
0.7261797995163872,
|
|
0.8113796888559393,
|
|
0.3760687378312029
|
|
],
|
|
"eval_runtime": 85.2104,
|
|
"eval_samples_per_second": 1.044,
|
|
"eval_steps_per_second": 0.141,
|
|
"step": 8500
|
|
},
|
|
{
|
|
"epoch": 3.86,
|
|
"learning_rate": 1.8193372673626876e-05,
|
|
"loss": 0.076,
|
|
"step": 8510
|
|
},
|
|
{
|
|
"epoch": 3.87,
|
|
"learning_rate": 1.812074443940082e-05,
|
|
"loss": 0.0866,
|
|
"step": 8520
|
|
},
|
|
{
|
|
"epoch": 3.87,
|
|
"learning_rate": 1.8048116205174763e-05,
|
|
"loss": 0.075,
|
|
"step": 8530
|
|
},
|
|
{
|
|
"epoch": 3.88,
|
|
"learning_rate": 1.797548797094871e-05,
|
|
"loss": 0.1002,
|
|
"step": 8540
|
|
},
|
|
{
|
|
"epoch": 3.88,
|
|
"learning_rate": 1.7902859736722653e-05,
|
|
"loss": 0.0928,
|
|
"step": 8550
|
|
},
|
|
{
|
|
"epoch": 3.89,
|
|
"learning_rate": 1.7830231502496596e-05,
|
|
"loss": 0.0759,
|
|
"step": 8560
|
|
},
|
|
{
|
|
"epoch": 3.89,
|
|
"learning_rate": 1.7757603268270543e-05,
|
|
"loss": 0.0882,
|
|
"step": 8570
|
|
},
|
|
{
|
|
"epoch": 3.89,
|
|
"learning_rate": 1.7684975034044486e-05,
|
|
"loss": 0.0862,
|
|
"step": 8580
|
|
},
|
|
{
|
|
"epoch": 3.9,
|
|
"learning_rate": 1.7612346799818433e-05,
|
|
"loss": 0.0934,
|
|
"step": 8590
|
|
},
|
|
{
|
|
"epoch": 3.9,
|
|
"learning_rate": 1.7539718565592376e-05,
|
|
"loss": 0.0839,
|
|
"step": 8600
|
|
},
|
|
{
|
|
"epoch": 3.9,
|
|
"eval_loss": 0.1454068124294281,
|
|
"eval_mean_accuracy": 0.7997635750166031,
|
|
"eval_mean_iou": 0.6893662286822321,
|
|
"eval_overall_accuracy": 0.9600287876772077,
|
|
"eval_per_category_accuracy": [
|
|
0.9921701128354435,
|
|
0.7270033517382082,
|
|
0.9204925305381314,
|
|
0.7166304937601736,
|
|
0.8673377460542467,
|
|
0.76710725625886,
|
|
0.9120828070028918,
|
|
0.7287068297467634,
|
|
0.3685980341438179,
|
|
0.7192869352869353,
|
|
0.7775187393023668,
|
|
0.92664460548841,
|
|
0.8900913619768718,
|
|
0.8818744170494343,
|
|
0.847468313502328,
|
|
0.8173047276464542,
|
|
0.911029305125697,
|
|
0.6243967828418231
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.98499100347911,
|
|
0.6740377109229568,
|
|
0.8138506749288889,
|
|
0.6264674493062967,
|
|
0.7778713799496764,
|
|
0.6535076993690768,
|
|
0.8535941965267092,
|
|
0.5348813774613862,
|
|
0.30828141225337485,
|
|
0.5664123194314109,
|
|
0.5975008693285759,
|
|
0.8513121333744983,
|
|
0.8093334150978108,
|
|
0.8059208806938117,
|
|
0.7369089401103046,
|
|
0.7218235350847833,
|
|
0.8204598065288858,
|
|
0.2714373124326214
|
|
],
|
|
"eval_runtime": 87.0327,
|
|
"eval_samples_per_second": 1.023,
|
|
"eval_steps_per_second": 0.138,
|
|
"step": 8600
|
|
},
|
|
{
|
|
"epoch": 3.91,
|
|
"learning_rate": 1.746709033136632e-05,
|
|
"loss": 0.0763,
|
|
"step": 8610
|
|
},
|
|
{
|
|
"epoch": 3.91,
|
|
"learning_rate": 1.7394462097140266e-05,
|
|
"loss": 0.085,
|
|
"step": 8620
|
|
},
|
|
{
|
|
"epoch": 3.92,
|
|
"learning_rate": 1.732183386291421e-05,
|
|
"loss": 0.0936,
|
|
"step": 8630
|
|
},
|
|
{
|
|
"epoch": 3.92,
|
|
"learning_rate": 1.7249205628688156e-05,
|
|
"loss": 0.0972,
|
|
"step": 8640
|
|
},
|
|
{
|
|
"epoch": 3.93,
|
|
"learning_rate": 1.71765773944621e-05,
|
|
"loss": 0.0845,
|
|
"step": 8650
|
|
},
|
|
{
|
|
"epoch": 3.93,
|
|
"learning_rate": 1.7103949160236043e-05,
|
|
"loss": 0.0784,
|
|
"step": 8660
|
|
},
|
|
{
|
|
"epoch": 3.94,
|
|
"learning_rate": 1.703132092600999e-05,
|
|
"loss": 0.0805,
|
|
"step": 8670
|
|
},
|
|
{
|
|
"epoch": 3.94,
|
|
"learning_rate": 1.6958692691783933e-05,
|
|
"loss": 0.0815,
|
|
"step": 8680
|
|
},
|
|
{
|
|
"epoch": 3.94,
|
|
"learning_rate": 1.6886064457557876e-05,
|
|
"loss": 0.0906,
|
|
"step": 8690
|
|
},
|
|
{
|
|
"epoch": 3.95,
|
|
"learning_rate": 1.6813436223331823e-05,
|
|
"loss": 0.0882,
|
|
"step": 8700
|
|
},
|
|
{
|
|
"epoch": 3.95,
|
|
"eval_loss": 0.14950454235076904,
|
|
"eval_mean_accuracy": 0.797425182067469,
|
|
"eval_mean_iou": 0.6949051175362407,
|
|
"eval_overall_accuracy": 0.9602914017237975,
|
|
"eval_per_category_accuracy": [
|
|
0.9922564792266364,
|
|
0.7303550899464045,
|
|
0.9147993011995158,
|
|
0.7272788931090614,
|
|
0.8826363525959328,
|
|
0.7583786361889957,
|
|
0.904900353902849,
|
|
0.7235459985946354,
|
|
0.3380237972064149,
|
|
0.7512185592185592,
|
|
0.7655373900725964,
|
|
0.9206469697508827,
|
|
0.8773897091883531,
|
|
0.8885442514474773,
|
|
0.8606656320055182,
|
|
0.8113116649537513,
|
|
0.8995288098669127,
|
|
0.6066353887399464
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9853183642068217,
|
|
0.6753305086014583,
|
|
0.8137482085143938,
|
|
0.6271860560332222,
|
|
0.7828199631106115,
|
|
0.6470856436604248,
|
|
0.8389340018002148,
|
|
0.5378408753694257,
|
|
0.289833215046132,
|
|
0.5719449071870538,
|
|
0.5950771536496964,
|
|
0.8532900450349763,
|
|
0.8113708842496539,
|
|
0.8133570677760639,
|
|
0.7382999574712006,
|
|
0.7222917989100288,
|
|
0.8034911027320235,
|
|
0.40107236229893206
|
|
],
|
|
"eval_runtime": 83.3489,
|
|
"eval_samples_per_second": 1.068,
|
|
"eval_steps_per_second": 0.144,
|
|
"step": 8700
|
|
},
|
|
{
|
|
"epoch": 3.95,
|
|
"learning_rate": 1.6740807989105766e-05,
|
|
"loss": 0.0761,
|
|
"step": 8710
|
|
},
|
|
{
|
|
"epoch": 3.96,
|
|
"learning_rate": 1.6668179754879713e-05,
|
|
"loss": 0.1032,
|
|
"step": 8720
|
|
},
|
|
{
|
|
"epoch": 3.96,
|
|
"learning_rate": 1.6595551520653653e-05,
|
|
"loss": 0.0881,
|
|
"step": 8730
|
|
},
|
|
{
|
|
"epoch": 3.97,
|
|
"learning_rate": 1.65229232864276e-05,
|
|
"loss": 0.0764,
|
|
"step": 8740
|
|
},
|
|
{
|
|
"epoch": 3.97,
|
|
"learning_rate": 1.6450295052201543e-05,
|
|
"loss": 0.0841,
|
|
"step": 8750
|
|
},
|
|
{
|
|
"epoch": 3.98,
|
|
"learning_rate": 1.637766681797549e-05,
|
|
"loss": 0.0933,
|
|
"step": 8760
|
|
},
|
|
{
|
|
"epoch": 3.98,
|
|
"learning_rate": 1.6305038583749433e-05,
|
|
"loss": 0.0821,
|
|
"step": 8770
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"learning_rate": 1.6232410349523377e-05,
|
|
"loss": 0.0974,
|
|
"step": 8780
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"learning_rate": 1.6159782115297323e-05,
|
|
"loss": 0.0939,
|
|
"step": 8790
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"learning_rate": 1.6087153881071267e-05,
|
|
"loss": 0.079,
|
|
"step": 8800
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"eval_loss": 0.1423356533050537,
|
|
"eval_mean_accuracy": 0.7975943044649596,
|
|
"eval_mean_iou": 0.6952737592404045,
|
|
"eval_overall_accuracy": 0.9609540875038404,
|
|
"eval_per_category_accuracy": [
|
|
0.9924551163794055,
|
|
0.720510370958079,
|
|
0.9179837955320788,
|
|
0.7331795984807379,
|
|
0.8734463478410625,
|
|
0.7626137723625347,
|
|
0.9092954305817996,
|
|
0.765502815245135,
|
|
0.3264355923435075,
|
|
0.740971916971917,
|
|
0.7363611324244034,
|
|
0.9256996912225768,
|
|
0.8719718360520116,
|
|
0.8972907097478133,
|
|
0.8565916537334023,
|
|
0.824441161356629,
|
|
0.8905524372595064,
|
|
0.6113941018766756
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9853967989607081,
|
|
0.67163751075423,
|
|
0.8135335458585021,
|
|
0.6271392933805187,
|
|
0.7852261105676173,
|
|
0.6634668766989079,
|
|
0.8395876308534634,
|
|
0.5557512491311817,
|
|
0.2873275351759938,
|
|
0.5724192574705704,
|
|
0.590634369575509,
|
|
0.8507994186046511,
|
|
0.8067195780401162,
|
|
0.8133511998053831,
|
|
0.7381697610268508,
|
|
0.7266587027045389,
|
|
0.814508549963452,
|
|
0.37260027775508536
|
|
],
|
|
"eval_runtime": 86.6765,
|
|
"eval_samples_per_second": 1.027,
|
|
"eval_steps_per_second": 0.138,
|
|
"step": 8800
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"learning_rate": 1.601452564684521e-05,
|
|
"loss": 0.0973,
|
|
"step": 8810
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"learning_rate": 1.5941897412619157e-05,
|
|
"loss": 0.0778,
|
|
"step": 8820
|
|
},
|
|
{
|
|
"epoch": 4.01,
|
|
"learning_rate": 1.58692691783931e-05,
|
|
"loss": 0.0686,
|
|
"step": 8830
|
|
},
|
|
{
|
|
"epoch": 4.01,
|
|
"learning_rate": 1.5796640944167047e-05,
|
|
"loss": 0.0774,
|
|
"step": 8840
|
|
},
|
|
{
|
|
"epoch": 4.02,
|
|
"learning_rate": 1.572401270994099e-05,
|
|
"loss": 0.0732,
|
|
"step": 8850
|
|
},
|
|
{
|
|
"epoch": 4.02,
|
|
"learning_rate": 1.5651384475714933e-05,
|
|
"loss": 0.0742,
|
|
"step": 8860
|
|
},
|
|
{
|
|
"epoch": 4.03,
|
|
"learning_rate": 1.557875624148888e-05,
|
|
"loss": 0.0801,
|
|
"step": 8870
|
|
},
|
|
{
|
|
"epoch": 4.03,
|
|
"learning_rate": 1.5506128007262823e-05,
|
|
"loss": 0.0733,
|
|
"step": 8880
|
|
},
|
|
{
|
|
"epoch": 4.04,
|
|
"learning_rate": 1.543349977303677e-05,
|
|
"loss": 0.0754,
|
|
"step": 8890
|
|
},
|
|
{
|
|
"epoch": 4.04,
|
|
"learning_rate": 1.5360871538810714e-05,
|
|
"loss": 0.0812,
|
|
"step": 8900
|
|
},
|
|
{
|
|
"epoch": 4.04,
|
|
"eval_loss": 0.14698636531829834,
|
|
"eval_mean_accuracy": 0.7996153806330561,
|
|
"eval_mean_iou": 0.6972965406743133,
|
|
"eval_overall_accuracy": 0.9617347288667486,
|
|
"eval_per_category_accuracy": [
|
|
0.9925776490435962,
|
|
0.7288816204925598,
|
|
0.9133944783757015,
|
|
0.7270754205100379,
|
|
0.8765661740035527,
|
|
0.7589814475421637,
|
|
0.90428789666176,
|
|
0.7953893925681864,
|
|
0.3345059493016037,
|
|
0.7661929181929182,
|
|
0.720907356036908,
|
|
0.9212360585609628,
|
|
0.8905679549664587,
|
|
0.8877083223342661,
|
|
0.8648366097603035,
|
|
0.8118191161356629,
|
|
0.8911109533963026,
|
|
0.6070375335120644
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9854201188515314,
|
|
0.674301207345936,
|
|
0.8140986541037422,
|
|
0.6266074351180734,
|
|
0.7923066093873357,
|
|
0.6654663236008743,
|
|
0.8454510997493212,
|
|
0.5763550473378125,
|
|
0.28996816000717524,
|
|
0.5647766137451848,
|
|
0.5786224103082365,
|
|
0.8535889339067617,
|
|
0.8102106180283142,
|
|
0.8070114390848732,
|
|
0.7428222301422363,
|
|
0.727096996893338,
|
|
0.8217051837266607,
|
|
0.3755286508002322
|
|
],
|
|
"eval_runtime": 84.7255,
|
|
"eval_samples_per_second": 1.05,
|
|
"eval_steps_per_second": 0.142,
|
|
"step": 8900
|
|
},
|
|
{
|
|
"epoch": 4.04,
|
|
"learning_rate": 1.5288243304584657e-05,
|
|
"loss": 0.0733,
|
|
"step": 8910
|
|
},
|
|
{
|
|
"epoch": 4.05,
|
|
"learning_rate": 1.5215615070358602e-05,
|
|
"loss": 0.0821,
|
|
"step": 8920
|
|
},
|
|
{
|
|
"epoch": 4.05,
|
|
"learning_rate": 1.5142986836132547e-05,
|
|
"loss": 0.0899,
|
|
"step": 8930
|
|
},
|
|
{
|
|
"epoch": 4.06,
|
|
"learning_rate": 1.5070358601906492e-05,
|
|
"loss": 0.0831,
|
|
"step": 8940
|
|
},
|
|
{
|
|
"epoch": 4.06,
|
|
"learning_rate": 1.4997730367680437e-05,
|
|
"loss": 0.0876,
|
|
"step": 8950
|
|
},
|
|
{
|
|
"epoch": 4.07,
|
|
"learning_rate": 1.492510213345438e-05,
|
|
"loss": 0.0681,
|
|
"step": 8960
|
|
},
|
|
{
|
|
"epoch": 4.07,
|
|
"learning_rate": 1.4852473899228325e-05,
|
|
"loss": 0.0744,
|
|
"step": 8970
|
|
},
|
|
{
|
|
"epoch": 4.08,
|
|
"learning_rate": 1.477984566500227e-05,
|
|
"loss": 0.0736,
|
|
"step": 8980
|
|
},
|
|
{
|
|
"epoch": 4.08,
|
|
"learning_rate": 1.4707217430776215e-05,
|
|
"loss": 0.0774,
|
|
"step": 8990
|
|
},
|
|
{
|
|
"epoch": 4.09,
|
|
"learning_rate": 1.463458919655016e-05,
|
|
"loss": 0.0764,
|
|
"step": 9000
|
|
},
|
|
{
|
|
"epoch": 4.09,
|
|
"eval_loss": 0.14982560276985168,
|
|
"eval_mean_accuracy": 0.8015064865594175,
|
|
"eval_mean_iou": 0.6920986484162183,
|
|
"eval_overall_accuracy": 0.9613027679786253,
|
|
"eval_per_category_accuracy": [
|
|
0.9924349253926398,
|
|
0.7299826745899383,
|
|
0.9157725321888412,
|
|
0.7145279435702658,
|
|
0.8650234347179635,
|
|
0.7611056399295174,
|
|
0.9126813447612286,
|
|
0.7848672219482544,
|
|
0.3223486808070357,
|
|
0.7378754578754578,
|
|
0.7460012984713451,
|
|
0.9237347439835846,
|
|
0.8957475316080068,
|
|
0.8944353518821604,
|
|
0.8760616054492154,
|
|
0.8233299075025694,
|
|
0.8943231926140098,
|
|
0.6368632707774798
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855178734297589,
|
|
0.6729709961039543,
|
|
0.8143660567657411,
|
|
0.627307371680362,
|
|
0.7865533688324895,
|
|
0.6713737833726459,
|
|
0.8465178204000613,
|
|
0.5682239641774492,
|
|
0.28379486245217705,
|
|
0.5729933097683451,
|
|
0.5916953397467406,
|
|
0.8528486326271372,
|
|
0.8104870041373586,
|
|
0.8055202253735424,
|
|
0.7439990480854161,
|
|
0.7292381391273675,
|
|
0.8059565356432753,
|
|
0.2884113397681054
|
|
],
|
|
"eval_runtime": 86.8211,
|
|
"eval_samples_per_second": 1.025,
|
|
"eval_steps_per_second": 0.138,
|
|
"step": 9000
|
|
},
|
|
{
|
|
"epoch": 4.09,
|
|
"learning_rate": 1.4561960962324104e-05,
|
|
"loss": 0.0809,
|
|
"step": 9010
|
|
},
|
|
{
|
|
"epoch": 4.09,
|
|
"learning_rate": 1.4489332728098049e-05,
|
|
"loss": 0.0832,
|
|
"step": 9020
|
|
},
|
|
{
|
|
"epoch": 4.1,
|
|
"learning_rate": 1.4416704493871994e-05,
|
|
"loss": 0.0721,
|
|
"step": 9030
|
|
},
|
|
{
|
|
"epoch": 4.1,
|
|
"learning_rate": 1.4344076259645939e-05,
|
|
"loss": 0.0727,
|
|
"step": 9040
|
|
},
|
|
{
|
|
"epoch": 4.11,
|
|
"learning_rate": 1.4271448025419882e-05,
|
|
"loss": 0.0814,
|
|
"step": 9050
|
|
},
|
|
{
|
|
"epoch": 4.11,
|
|
"learning_rate": 1.4198819791193827e-05,
|
|
"loss": 0.0735,
|
|
"step": 9060
|
|
},
|
|
{
|
|
"epoch": 4.12,
|
|
"learning_rate": 1.4126191556967772e-05,
|
|
"loss": 0.0731,
|
|
"step": 9070
|
|
},
|
|
{
|
|
"epoch": 4.12,
|
|
"learning_rate": 1.4053563322741717e-05,
|
|
"loss": 0.0808,
|
|
"step": 9080
|
|
},
|
|
{
|
|
"epoch": 4.13,
|
|
"learning_rate": 1.398093508851566e-05,
|
|
"loss": 0.0947,
|
|
"step": 9090
|
|
},
|
|
{
|
|
"epoch": 4.13,
|
|
"learning_rate": 1.3908306854289606e-05,
|
|
"loss": 0.0749,
|
|
"step": 9100
|
|
},
|
|
{
|
|
"epoch": 4.13,
|
|
"eval_loss": 0.15402460098266602,
|
|
"eval_mean_accuracy": 0.8009089659315938,
|
|
"eval_mean_iou": 0.694753290646456,
|
|
"eval_overall_accuracy": 0.96209283893028,
|
|
"eval_per_category_accuracy": [
|
|
0.9929531237397967,
|
|
0.7231496623973834,
|
|
0.9194608369098712,
|
|
0.7209712425393381,
|
|
0.8628079258816579,
|
|
0.7717906367719342,
|
|
0.8977840183458783,
|
|
0.8285148114228323,
|
|
0.3270563890325918,
|
|
0.7431013431013431,
|
|
0.7329772374038442,
|
|
0.9246519829361926,
|
|
0.8941109292664065,
|
|
0.8941405768790807,
|
|
0.8529110622521124,
|
|
0.8189298561151079,
|
|
0.8821623523095106,
|
|
0.628887399463807
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9853784134118165,
|
|
0.6752698902295201,
|
|
0.8145830953088226,
|
|
0.6272866753216099,
|
|
0.7905764692125179,
|
|
0.6927389913548095,
|
|
0.8495780816504622,
|
|
0.5937353964212252,
|
|
0.28859673148908976,
|
|
0.5682879275699014,
|
|
0.5839439511610228,
|
|
0.8511850169600094,
|
|
0.8083813348834752,
|
|
0.8079117157156998,
|
|
0.7390650609610886,
|
|
0.7236816711131293,
|
|
0.7995642579659965,
|
|
0.30579455090600965
|
|
],
|
|
"eval_runtime": 85.3031,
|
|
"eval_samples_per_second": 1.043,
|
|
"eval_steps_per_second": 0.141,
|
|
"step": 9100
|
|
},
|
|
{
|
|
"epoch": 4.14,
|
|
"learning_rate": 1.383567862006355e-05,
|
|
"loss": 0.0708,
|
|
"step": 9110
|
|
},
|
|
{
|
|
"epoch": 4.14,
|
|
"learning_rate": 1.3763050385837496e-05,
|
|
"loss": 0.0728,
|
|
"step": 9120
|
|
},
|
|
{
|
|
"epoch": 4.14,
|
|
"learning_rate": 1.369042215161144e-05,
|
|
"loss": 0.0869,
|
|
"step": 9130
|
|
},
|
|
{
|
|
"epoch": 4.15,
|
|
"learning_rate": 1.3617793917385384e-05,
|
|
"loss": 0.0803,
|
|
"step": 9140
|
|
},
|
|
{
|
|
"epoch": 4.15,
|
|
"learning_rate": 1.3545165683159329e-05,
|
|
"loss": 0.0928,
|
|
"step": 9150
|
|
},
|
|
{
|
|
"epoch": 4.16,
|
|
"learning_rate": 1.3472537448933274e-05,
|
|
"loss": 0.0717,
|
|
"step": 9160
|
|
},
|
|
{
|
|
"epoch": 4.16,
|
|
"learning_rate": 1.3399909214707219e-05,
|
|
"loss": 0.0688,
|
|
"step": 9170
|
|
},
|
|
{
|
|
"epoch": 4.17,
|
|
"learning_rate": 1.3327280980481162e-05,
|
|
"loss": 0.0926,
|
|
"step": 9180
|
|
},
|
|
{
|
|
"epoch": 4.17,
|
|
"learning_rate": 1.3254652746255108e-05,
|
|
"loss": 0.0712,
|
|
"step": 9190
|
|
},
|
|
{
|
|
"epoch": 4.18,
|
|
"learning_rate": 1.3182024512029053e-05,
|
|
"loss": 0.0697,
|
|
"step": 9200
|
|
},
|
|
{
|
|
"epoch": 4.18,
|
|
"eval_loss": 0.15140081942081451,
|
|
"eval_mean_accuracy": 0.8005548648580436,
|
|
"eval_mean_iou": 0.6961780056115205,
|
|
"eval_overall_accuracy": 0.9621129410990169,
|
|
"eval_per_category_accuracy": [
|
|
0.992782831626141,
|
|
0.7380624686280542,
|
|
0.9173097556949489,
|
|
0.7299240368963646,
|
|
0.8709996137096703,
|
|
0.7526817376865292,
|
|
0.9037885352180313,
|
|
0.8238380904382823,
|
|
0.31841696844283496,
|
|
0.7420952380952381,
|
|
0.7270554211178658,
|
|
0.9283248923627996,
|
|
0.8986790281099941,
|
|
0.8894769723527445,
|
|
0.8599543024659424,
|
|
0.8222764645426516,
|
|
0.8756885022818933,
|
|
0.618632707774799
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855298212243018,
|
|
0.6737816144624618,
|
|
0.8137645522661264,
|
|
0.6260616637579989,
|
|
0.7914383525346843,
|
|
0.6828467489593224,
|
|
0.8360892428585915,
|
|
0.5947329986307622,
|
|
0.2835229628264775,
|
|
0.5737156493305442,
|
|
0.586409076483656,
|
|
0.8514208841137287,
|
|
0.8150094803156027,
|
|
0.8100124203694058,
|
|
0.742073788863778,
|
|
0.7272541344498668,
|
|
0.8038699699172299,
|
|
0.33367073964283134
|
|
],
|
|
"eval_runtime": 81.7415,
|
|
"eval_samples_per_second": 1.089,
|
|
"eval_steps_per_second": 0.147,
|
|
"step": 9200
|
|
},
|
|
{
|
|
"epoch": 4.18,
|
|
"learning_rate": 1.3109396277802998e-05,
|
|
"loss": 0.0786,
|
|
"step": 9210
|
|
},
|
|
{
|
|
"epoch": 4.19,
|
|
"learning_rate": 1.3036768043576941e-05,
|
|
"loss": 0.0687,
|
|
"step": 9220
|
|
},
|
|
{
|
|
"epoch": 4.19,
|
|
"learning_rate": 1.2964139809350886e-05,
|
|
"loss": 0.077,
|
|
"step": 9230
|
|
},
|
|
{
|
|
"epoch": 4.19,
|
|
"learning_rate": 1.2891511575124831e-05,
|
|
"loss": 0.0727,
|
|
"step": 9240
|
|
},
|
|
{
|
|
"epoch": 4.2,
|
|
"learning_rate": 1.2818883340898776e-05,
|
|
"loss": 0.075,
|
|
"step": 9250
|
|
},
|
|
{
|
|
"epoch": 4.2,
|
|
"learning_rate": 1.2746255106672721e-05,
|
|
"loss": 0.0908,
|
|
"step": 9260
|
|
},
|
|
{
|
|
"epoch": 4.21,
|
|
"learning_rate": 1.2673626872446664e-05,
|
|
"loss": 0.0789,
|
|
"step": 9270
|
|
},
|
|
{
|
|
"epoch": 4.21,
|
|
"learning_rate": 1.260099863822061e-05,
|
|
"loss": 0.0802,
|
|
"step": 9280
|
|
},
|
|
{
|
|
"epoch": 4.22,
|
|
"learning_rate": 1.2528370403994554e-05,
|
|
"loss": 0.0796,
|
|
"step": 9290
|
|
},
|
|
{
|
|
"epoch": 4.22,
|
|
"learning_rate": 1.24557421697685e-05,
|
|
"loss": 0.0763,
|
|
"step": 9300
|
|
},
|
|
{
|
|
"epoch": 4.22,
|
|
"eval_loss": 0.1495458036661148,
|
|
"eval_mean_accuracy": 0.8018444069331001,
|
|
"eval_mean_iou": 0.6997454892585403,
|
|
"eval_overall_accuracy": 0.9618220382861877,
|
|
"eval_per_category_accuracy": [
|
|
0.9929321007068731,
|
|
0.7280720218915462,
|
|
0.9101068146803125,
|
|
0.7258545849158979,
|
|
0.8758930819724546,
|
|
0.7673435229796988,
|
|
0.8916803251591171,
|
|
0.7866609585020041,
|
|
0.3637351267459907,
|
|
0.7121465201465201,
|
|
0.7782466702079521,
|
|
0.9229044845868274,
|
|
0.8839855763178246,
|
|
0.8962172007813738,
|
|
0.8624331781341611,
|
|
0.813225847893114,
|
|
0.9073511215065742,
|
|
0.6144101876675603
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855452046413777,
|
|
0.6761450783435085,
|
|
0.8157453058607579,
|
|
0.628457337483117,
|
|
0.7905338290736328,
|
|
0.6714772913909204,
|
|
0.837535892947904,
|
|
0.5760236393361092,
|
|
0.3095856633349478,
|
|
0.565627569941192,
|
|
0.5921559821862954,
|
|
0.8530411331345378,
|
|
0.8028986335829855,
|
|
0.8034290176775446,
|
|
0.74795532083937,
|
|
0.7289440350069093,
|
|
0.8228870785840817,
|
|
0.38743079328853386
|
|
],
|
|
"eval_runtime": 85.8083,
|
|
"eval_samples_per_second": 1.037,
|
|
"eval_steps_per_second": 0.14,
|
|
"step": 9300
|
|
},
|
|
{
|
|
"epoch": 4.23,
|
|
"learning_rate": 1.2383113935542443e-05,
|
|
"loss": 0.0892,
|
|
"step": 9310
|
|
},
|
|
{
|
|
"epoch": 4.23,
|
|
"learning_rate": 1.2310485701316388e-05,
|
|
"loss": 0.0864,
|
|
"step": 9320
|
|
},
|
|
{
|
|
"epoch": 4.24,
|
|
"learning_rate": 1.2237857467090333e-05,
|
|
"loss": 0.0665,
|
|
"step": 9330
|
|
},
|
|
{
|
|
"epoch": 4.24,
|
|
"learning_rate": 1.2165229232864278e-05,
|
|
"loss": 0.0729,
|
|
"step": 9340
|
|
},
|
|
{
|
|
"epoch": 4.24,
|
|
"learning_rate": 1.2092600998638221e-05,
|
|
"loss": 0.0737,
|
|
"step": 9350
|
|
},
|
|
{
|
|
"epoch": 4.25,
|
|
"learning_rate": 1.2019972764412166e-05,
|
|
"loss": 0.0797,
|
|
"step": 9360
|
|
},
|
|
{
|
|
"epoch": 4.25,
|
|
"learning_rate": 1.1947344530186111e-05,
|
|
"loss": 0.0736,
|
|
"step": 9370
|
|
},
|
|
{
|
|
"epoch": 4.26,
|
|
"learning_rate": 1.1874716295960056e-05,
|
|
"loss": 0.0785,
|
|
"step": 9380
|
|
},
|
|
{
|
|
"epoch": 4.26,
|
|
"learning_rate": 1.1802088061734001e-05,
|
|
"loss": 0.091,
|
|
"step": 9390
|
|
},
|
|
{
|
|
"epoch": 4.27,
|
|
"learning_rate": 1.1729459827507945e-05,
|
|
"loss": 0.0847,
|
|
"step": 9400
|
|
},
|
|
{
|
|
"epoch": 4.27,
|
|
"eval_loss": 0.14821504056453705,
|
|
"eval_mean_accuracy": 0.8064361413894966,
|
|
"eval_mean_iou": 0.6972761209085664,
|
|
"eval_overall_accuracy": 0.9622693865486746,
|
|
"eval_per_category_accuracy": [
|
|
0.9929973885953984,
|
|
0.7311646885474182,
|
|
0.9175608011444921,
|
|
0.7520347259902334,
|
|
0.864897187079103,
|
|
0.7666568038191302,
|
|
0.8978501358889504,
|
|
0.8139553787710466,
|
|
0.37211588204862905,
|
|
0.7375921855921856,
|
|
0.7173562336461469,
|
|
0.922758200788351,
|
|
0.8940524791827779,
|
|
0.8880382943526389,
|
|
0.876913045352647,
|
|
0.8234712230215827,
|
|
0.9088353549971766,
|
|
0.6376005361930295
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855848828569402,
|
|
0.6755127380435921,
|
|
0.8165006502945452,
|
|
0.6316868911297214,
|
|
0.7884445422970183,
|
|
0.6859489927157852,
|
|
0.8430362210281603,
|
|
0.589946556550292,
|
|
0.3199875439298901,
|
|
0.562432033845283,
|
|
0.5734720518385706,
|
|
0.8528903392252232,
|
|
0.8178703486626441,
|
|
0.81427124893599,
|
|
0.7500702930181747,
|
|
0.7368294000666721,
|
|
0.829614753013219,
|
|
0.27687068890247096
|
|
],
|
|
"eval_runtime": 75.2666,
|
|
"eval_samples_per_second": 1.182,
|
|
"eval_steps_per_second": 0.159,
|
|
"step": 9400
|
|
},
|
|
{
|
|
"epoch": 4.27,
|
|
"learning_rate": 1.165683159328189e-05,
|
|
"loss": 0.0917,
|
|
"step": 9410
|
|
},
|
|
{
|
|
"epoch": 4.28,
|
|
"learning_rate": 1.1584203359055835e-05,
|
|
"loss": 0.0773,
|
|
"step": 9420
|
|
},
|
|
{
|
|
"epoch": 4.28,
|
|
"learning_rate": 1.151157512482978e-05,
|
|
"loss": 0.0734,
|
|
"step": 9430
|
|
},
|
|
{
|
|
"epoch": 4.29,
|
|
"learning_rate": 1.1438946890603723e-05,
|
|
"loss": 0.0798,
|
|
"step": 9440
|
|
},
|
|
{
|
|
"epoch": 4.29,
|
|
"learning_rate": 1.1366318656377668e-05,
|
|
"loss": 0.0875,
|
|
"step": 9450
|
|
},
|
|
{
|
|
"epoch": 4.29,
|
|
"learning_rate": 1.1293690422151613e-05,
|
|
"loss": 0.068,
|
|
"step": 9460
|
|
},
|
|
{
|
|
"epoch": 4.3,
|
|
"learning_rate": 1.1221062187925558e-05,
|
|
"loss": 0.0806,
|
|
"step": 9470
|
|
},
|
|
{
|
|
"epoch": 4.3,
|
|
"learning_rate": 1.1148433953699501e-05,
|
|
"loss": 0.0884,
|
|
"step": 9480
|
|
},
|
|
{
|
|
"epoch": 4.31,
|
|
"learning_rate": 1.1075805719473447e-05,
|
|
"loss": 0.0987,
|
|
"step": 9490
|
|
},
|
|
{
|
|
"epoch": 4.31,
|
|
"learning_rate": 1.1003177485247392e-05,
|
|
"loss": 0.0896,
|
|
"step": 9500
|
|
},
|
|
{
|
|
"epoch": 4.31,
|
|
"eval_loss": 0.14339512586593628,
|
|
"eval_mean_accuracy": 0.805191471580113,
|
|
"eval_mean_iou": 0.7010373604415414,
|
|
"eval_overall_accuracy": 0.9629785773459445,
|
|
"eval_per_category_accuracy": [
|
|
0.9929541221951863,
|
|
0.7355365209928917,
|
|
0.9082514856388247,
|
|
0.7241589799240369,
|
|
0.8703498500466225,
|
|
0.7673280662783355,
|
|
0.9190181892840862,
|
|
0.82143921617506,
|
|
0.34997413347128814,
|
|
0.7357362637362638,
|
|
0.748381829270692,
|
|
0.9288349088493791,
|
|
0.8834865025268421,
|
|
0.8941273779983457,
|
|
0.8731677875495775,
|
|
0.8239208633093525,
|
|
0.8968072008467228,
|
|
0.6199731903485255
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.985366983039312,
|
|
0.6760626265031552,
|
|
0.8168545355290974,
|
|
0.6312895405900787,
|
|
0.8007127860896879,
|
|
0.6777893720560556,
|
|
0.8562505572532078,
|
|
0.5992923036373221,
|
|
0.3036900700305261,
|
|
0.5772520347634157,
|
|
0.6002998382451572,
|
|
0.8507595647214325,
|
|
0.8116230561120175,
|
|
0.8071746029224274,
|
|
0.7468426777779826,
|
|
0.7331317622985957,
|
|
0.8194452276495201,
|
|
0.324834948728754
|
|
],
|
|
"eval_runtime": 73.2841,
|
|
"eval_samples_per_second": 1.214,
|
|
"eval_steps_per_second": 0.164,
|
|
"step": 9500
|
|
},
|
|
{
|
|
"epoch": 4.32,
|
|
"learning_rate": 1.0930549251021337e-05,
|
|
"loss": 0.0843,
|
|
"step": 9510
|
|
},
|
|
{
|
|
"epoch": 4.32,
|
|
"learning_rate": 1.0857921016795282e-05,
|
|
"loss": 0.0712,
|
|
"step": 9520
|
|
},
|
|
{
|
|
"epoch": 4.33,
|
|
"learning_rate": 1.0785292782569225e-05,
|
|
"loss": 0.0696,
|
|
"step": 9530
|
|
},
|
|
{
|
|
"epoch": 4.33,
|
|
"learning_rate": 1.071266454834317e-05,
|
|
"loss": 0.077,
|
|
"step": 9540
|
|
},
|
|
{
|
|
"epoch": 4.33,
|
|
"learning_rate": 1.0640036314117115e-05,
|
|
"loss": 0.078,
|
|
"step": 9550
|
|
},
|
|
{
|
|
"epoch": 4.34,
|
|
"learning_rate": 1.056740807989106e-05,
|
|
"loss": 0.0828,
|
|
"step": 9560
|
|
},
|
|
{
|
|
"epoch": 4.34,
|
|
"learning_rate": 1.0494779845665003e-05,
|
|
"loss": 0.0753,
|
|
"step": 9570
|
|
},
|
|
{
|
|
"epoch": 4.35,
|
|
"learning_rate": 1.0422151611438948e-05,
|
|
"loss": 0.0753,
|
|
"step": 9580
|
|
},
|
|
{
|
|
"epoch": 4.35,
|
|
"learning_rate": 1.0349523377212893e-05,
|
|
"loss": 0.0761,
|
|
"step": 9590
|
|
},
|
|
{
|
|
"epoch": 4.36,
|
|
"learning_rate": 1.0276895142986838e-05,
|
|
"loss": 0.0869,
|
|
"step": 9600
|
|
},
|
|
{
|
|
"epoch": 4.36,
|
|
"eval_loss": 0.14896970987319946,
|
|
"eval_mean_accuracy": 0.8013344143351102,
|
|
"eval_mean_iou": 0.6940233914355434,
|
|
"eval_overall_accuracy": 0.9610248094194391,
|
|
"eval_per_category_accuracy": [
|
|
0.9930181342796028,
|
|
0.7365728072021892,
|
|
0.9134374656102124,
|
|
0.7518312533912099,
|
|
0.8709138476506618,
|
|
0.733261496473664,
|
|
0.9116895816151471,
|
|
0.765255341017679,
|
|
0.3265390584583549,
|
|
0.7439609279609279,
|
|
0.7598910071022448,
|
|
0.9220228281797946,
|
|
0.8925462654892722,
|
|
0.8905504813191841,
|
|
0.8588441972753923,
|
|
0.8227582219938335,
|
|
0.9009667574898248,
|
|
0.6299597855227882
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855236516057857,
|
|
0.6757981370611917,
|
|
0.8169465043105967,
|
|
0.6337907375643225,
|
|
0.7813916266013309,
|
|
0.6447653275395748,
|
|
0.8497944355695392,
|
|
0.5533401688860444,
|
|
0.28893161219445207,
|
|
0.5823527162900944,
|
|
0.6036257081461223,
|
|
0.8525782806587823,
|
|
0.8160930408472012,
|
|
0.8133297438040439,
|
|
0.7446861201317665,
|
|
0.7286737474471074,
|
|
0.8354884954688658,
|
|
0.28531099171295876
|
|
],
|
|
"eval_runtime": 77.035,
|
|
"eval_samples_per_second": 1.155,
|
|
"eval_steps_per_second": 0.156,
|
|
"step": 9600
|
|
},
|
|
{
|
|
"epoch": 4.36,
|
|
"learning_rate": 1.0204266908760782e-05,
|
|
"loss": 0.0711,
|
|
"step": 9610
|
|
},
|
|
{
|
|
"epoch": 4.37,
|
|
"learning_rate": 1.0131638674534727e-05,
|
|
"loss": 0.0716,
|
|
"step": 9620
|
|
},
|
|
{
|
|
"epoch": 4.37,
|
|
"learning_rate": 1.0059010440308672e-05,
|
|
"loss": 0.0704,
|
|
"step": 9630
|
|
},
|
|
{
|
|
"epoch": 4.38,
|
|
"learning_rate": 9.986382206082615e-06,
|
|
"loss": 0.0795,
|
|
"step": 9640
|
|
},
|
|
{
|
|
"epoch": 4.38,
|
|
"learning_rate": 9.91375397185656e-06,
|
|
"loss": 0.0751,
|
|
"step": 9650
|
|
},
|
|
{
|
|
"epoch": 4.38,
|
|
"learning_rate": 9.841125737630505e-06,
|
|
"loss": 0.0773,
|
|
"step": 9660
|
|
},
|
|
{
|
|
"epoch": 4.39,
|
|
"learning_rate": 9.768497503404449e-06,
|
|
"loss": 0.0723,
|
|
"step": 9670
|
|
},
|
|
{
|
|
"epoch": 4.39,
|
|
"learning_rate": 9.695869269178394e-06,
|
|
"loss": 0.0899,
|
|
"step": 9680
|
|
},
|
|
{
|
|
"epoch": 4.4,
|
|
"learning_rate": 9.623241034952339e-06,
|
|
"loss": 0.081,
|
|
"step": 9690
|
|
},
|
|
{
|
|
"epoch": 4.4,
|
|
"learning_rate": 9.550612800726284e-06,
|
|
"loss": 0.0758,
|
|
"step": 9700
|
|
},
|
|
{
|
|
"epoch": 4.4,
|
|
"eval_loss": 0.1480877697467804,
|
|
"eval_mean_accuracy": 0.8042308847884129,
|
|
"eval_mean_iou": 0.6978584772456488,
|
|
"eval_overall_accuracy": 0.9620015005047402,
|
|
"eval_per_category_accuracy": [
|
|
0.9928457897854241,
|
|
0.7331401091338914,
|
|
0.911350005502366,
|
|
0.7341291372761801,
|
|
0.871964310341397,
|
|
0.766100362570052,
|
|
0.9007192892712107,
|
|
0.7903893293448436,
|
|
0.34749094671495084,
|
|
0.753992673992674,
|
|
0.7530248480198312,
|
|
0.9246559365523677,
|
|
0.891799902882938,
|
|
0.8920507540960526,
|
|
0.8524799534402483,
|
|
0.8266829393627955,
|
|
0.9091439274484453,
|
|
0.624195710455764
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855239967930458,
|
|
0.6758414807075155,
|
|
0.8167526035326061,
|
|
0.6312106368089573,
|
|
0.789750648624295,
|
|
0.6701058803987995,
|
|
0.8434524469007227,
|
|
0.5789336498221074,
|
|
0.2992648696814435,
|
|
0.5821793827495702,
|
|
0.6023558849912658,
|
|
0.8526471596170533,
|
|
0.8104330274329702,
|
|
0.8094245770357773,
|
|
0.7416351387462905,
|
|
0.7291547452451234,
|
|
0.8305421107675134,
|
|
0.31224435056661975
|
|
],
|
|
"eval_runtime": 76.325,
|
|
"eval_samples_per_second": 1.166,
|
|
"eval_steps_per_second": 0.157,
|
|
"step": 9700
|
|
},
|
|
{
|
|
"epoch": 4.41,
|
|
"learning_rate": 9.477984566500229e-06,
|
|
"loss": 0.0793,
|
|
"step": 9710
|
|
},
|
|
{
|
|
"epoch": 4.41,
|
|
"learning_rate": 9.405356332274172e-06,
|
|
"loss": 0.0692,
|
|
"step": 9720
|
|
},
|
|
{
|
|
"epoch": 4.42,
|
|
"learning_rate": 9.332728098048117e-06,
|
|
"loss": 0.0896,
|
|
"step": 9730
|
|
},
|
|
{
|
|
"epoch": 4.42,
|
|
"learning_rate": 9.260099863822062e-06,
|
|
"loss": 0.0738,
|
|
"step": 9740
|
|
},
|
|
{
|
|
"epoch": 4.43,
|
|
"learning_rate": 9.187471629596007e-06,
|
|
"loss": 0.0739,
|
|
"step": 9750
|
|
},
|
|
{
|
|
"epoch": 4.43,
|
|
"learning_rate": 9.11484339536995e-06,
|
|
"loss": 0.0746,
|
|
"step": 9760
|
|
},
|
|
{
|
|
"epoch": 4.43,
|
|
"learning_rate": 9.042215161143895e-06,
|
|
"loss": 0.0716,
|
|
"step": 9770
|
|
},
|
|
{
|
|
"epoch": 4.44,
|
|
"learning_rate": 8.96958692691784e-06,
|
|
"loss": 0.074,
|
|
"step": 9780
|
|
},
|
|
{
|
|
"epoch": 4.44,
|
|
"learning_rate": 8.896958692691786e-06,
|
|
"loss": 0.0761,
|
|
"step": 9790
|
|
},
|
|
{
|
|
"epoch": 4.45,
|
|
"learning_rate": 8.824330458465729e-06,
|
|
"loss": 0.12,
|
|
"step": 9800
|
|
},
|
|
{
|
|
"epoch": 4.45,
|
|
"eval_loss": 0.14720258116722107,
|
|
"eval_mean_accuracy": 0.8020529731503411,
|
|
"eval_mean_iou": 0.6962903154679329,
|
|
"eval_overall_accuracy": 0.9608089575606785,
|
|
"eval_per_category_accuracy": [
|
|
0.9929007048318473,
|
|
0.7357955925452161,
|
|
0.9132844310553538,
|
|
0.7322300596852958,
|
|
0.8734134136744032,
|
|
0.7652546601954611,
|
|
0.9059147362084025,
|
|
0.7283581981708583,
|
|
0.3609415416451112,
|
|
0.7462075702075702,
|
|
0.7708886659190619,
|
|
0.9268106573677615,
|
|
0.8765444310558783,
|
|
0.8904976857962444,
|
|
0.8645348335919987,
|
|
0.8191675231243577,
|
|
0.9104152459476723,
|
|
0.6237935656836461
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9854089528456684,
|
|
0.6768547894603572,
|
|
0.8160466209634548,
|
|
0.6293575842369127,
|
|
0.782235372556972,
|
|
0.6477415665026288,
|
|
0.8502126103058845,
|
|
0.5364941748761263,
|
|
0.3084711291891414,
|
|
0.5842900302114804,
|
|
0.6064477685860876,
|
|
0.8520062513629425,
|
|
0.8130943245136966,
|
|
0.8141156879859383,
|
|
0.7465680115407883,
|
|
0.7336109896684231,
|
|
0.8317550976683084,
|
|
0.3185147159479808
|
|
],
|
|
"eval_runtime": 75.2135,
|
|
"eval_samples_per_second": 1.183,
|
|
"eval_steps_per_second": 0.16,
|
|
"step": 9800
|
|
},
|
|
{
|
|
"epoch": 4.45,
|
|
"learning_rate": 8.751702224239674e-06,
|
|
"loss": 0.0803,
|
|
"step": 9810
|
|
},
|
|
{
|
|
"epoch": 4.46,
|
|
"learning_rate": 8.679073990013619e-06,
|
|
"loss": 0.0799,
|
|
"step": 9820
|
|
},
|
|
{
|
|
"epoch": 4.46,
|
|
"learning_rate": 8.606445755787564e-06,
|
|
"loss": 0.0719,
|
|
"step": 9830
|
|
},
|
|
{
|
|
"epoch": 4.47,
|
|
"learning_rate": 8.533817521561509e-06,
|
|
"loss": 0.0769,
|
|
"step": 9840
|
|
},
|
|
{
|
|
"epoch": 4.47,
|
|
"learning_rate": 8.461189287335452e-06,
|
|
"loss": 0.0745,
|
|
"step": 9850
|
|
},
|
|
{
|
|
"epoch": 4.48,
|
|
"learning_rate": 8.388561053109397e-06,
|
|
"loss": 0.0698,
|
|
"step": 9860
|
|
},
|
|
{
|
|
"epoch": 4.48,
|
|
"learning_rate": 8.31593281888334e-06,
|
|
"loss": 0.0881,
|
|
"step": 9870
|
|
},
|
|
{
|
|
"epoch": 4.48,
|
|
"learning_rate": 8.243304584657286e-06,
|
|
"loss": 0.0781,
|
|
"step": 9880
|
|
},
|
|
{
|
|
"epoch": 4.49,
|
|
"learning_rate": 8.17067635043123e-06,
|
|
"loss": 0.0815,
|
|
"step": 9890
|
|
},
|
|
{
|
|
"epoch": 4.49,
|
|
"learning_rate": 8.098048116205174e-06,
|
|
"loss": 0.0728,
|
|
"step": 9900
|
|
},
|
|
{
|
|
"epoch": 4.49,
|
|
"eval_loss": 0.14797358214855194,
|
|
"eval_mean_accuracy": 0.8018117289656665,
|
|
"eval_mean_iou": 0.6966242338337922,
|
|
"eval_overall_accuracy": 0.9613266419828608,
|
|
"eval_per_category_accuracy": [
|
|
0.9929076940195739,
|
|
0.7325248141971211,
|
|
0.9085902250467701,
|
|
0.7393516006511123,
|
|
0.8735046687611882,
|
|
0.7645370276321658,
|
|
0.8980763274836707,
|
|
0.7640017124493987,
|
|
0.33300569063631663,
|
|
0.758896214896215,
|
|
0.7537429420212871,
|
|
0.9265853012457844,
|
|
0.8966107943815981,
|
|
0.887404748077363,
|
|
0.8670137092602173,
|
|
0.819263874614594,
|
|
0.9001058403507851,
|
|
0.6164879356568365
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9854493616568202,
|
|
0.6770731999341485,
|
|
0.8165316100479652,
|
|
0.6310640268611787,
|
|
0.7888095041888148,
|
|
0.6502368137378259,
|
|
0.8400098948187279,
|
|
0.560748183964797,
|
|
0.2940746493672621,
|
|
0.5783322663728804,
|
|
0.6029350434748397,
|
|
0.8528126398678374,
|
|
0.814758371602739,
|
|
0.8124743206554578,
|
|
0.7471613996795691,
|
|
0.7346523817752434,
|
|
0.8230719743569486,
|
|
0.3290405666452028
|
|
],
|
|
"eval_runtime": 75.3594,
|
|
"eval_samples_per_second": 1.181,
|
|
"eval_steps_per_second": 0.159,
|
|
"step": 9900
|
|
},
|
|
{
|
|
"epoch": 4.5,
|
|
"learning_rate": 8.025419881979119e-06,
|
|
"loss": 0.0681,
|
|
"step": 9910
|
|
},
|
|
{
|
|
"epoch": 4.5,
|
|
"learning_rate": 7.952791647753064e-06,
|
|
"loss": 0.0963,
|
|
"step": 9920
|
|
},
|
|
{
|
|
"epoch": 4.51,
|
|
"learning_rate": 7.88016341352701e-06,
|
|
"loss": 0.0803,
|
|
"step": 9930
|
|
},
|
|
{
|
|
"epoch": 4.51,
|
|
"learning_rate": 7.807535179300954e-06,
|
|
"loss": 0.0691,
|
|
"step": 9940
|
|
},
|
|
{
|
|
"epoch": 4.52,
|
|
"learning_rate": 7.734906945074898e-06,
|
|
"loss": 0.0783,
|
|
"step": 9950
|
|
},
|
|
{
|
|
"epoch": 4.52,
|
|
"learning_rate": 7.662278710848843e-06,
|
|
"loss": 0.0832,
|
|
"step": 9960
|
|
},
|
|
{
|
|
"epoch": 4.53,
|
|
"learning_rate": 7.589650476622788e-06,
|
|
"loss": 0.0745,
|
|
"step": 9970
|
|
},
|
|
{
|
|
"epoch": 4.53,
|
|
"learning_rate": 7.517022242396732e-06,
|
|
"loss": 0.0848,
|
|
"step": 9980
|
|
},
|
|
{
|
|
"epoch": 4.53,
|
|
"learning_rate": 7.444394008170677e-06,
|
|
"loss": 0.0798,
|
|
"step": 9990
|
|
},
|
|
{
|
|
"epoch": 4.54,
|
|
"learning_rate": 7.371765773944621e-06,
|
|
"loss": 0.0903,
|
|
"step": 10000
|
|
},
|
|
{
|
|
"epoch": 4.54,
|
|
"eval_loss": 0.15016531944274902,
|
|
"eval_mean_accuracy": 0.8001159626297416,
|
|
"eval_mean_iou": 0.6937524225865573,
|
|
"eval_overall_accuracy": 0.9609915058264572,
|
|
"eval_per_category_accuracy": [
|
|
0.993084254214286,
|
|
0.7336420602665199,
|
|
0.9131640667987234,
|
|
0.738537710255019,
|
|
0.8735005519903558,
|
|
0.7647092594473567,
|
|
0.9077590676730453,
|
|
0.7338965629984483,
|
|
0.32948784273150544,
|
|
0.7564151404151405,
|
|
0.7450274449625214,
|
|
0.9219042196945436,
|
|
0.8985081740193874,
|
|
0.8839686394593739,
|
|
0.8614685721676151,
|
|
0.8171569886947585,
|
|
0.9040524820025118,
|
|
0.625804289544236
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9854932998596498,
|
|
0.674642644431209,
|
|
0.8165864534481433,
|
|
0.629786003470214,
|
|
0.7797532642688882,
|
|
0.6523827643758924,
|
|
0.8486701911346075,
|
|
0.5482705618329305,
|
|
0.2901462347956813,
|
|
0.5782321052552979,
|
|
0.5995203077629402,
|
|
0.8533952085727461,
|
|
0.8229923400049419,
|
|
0.8208716186678542,
|
|
0.7445229978203767,
|
|
0.7306587789328586,
|
|
0.8232707265533675,
|
|
0.28834810537043326
|
|
],
|
|
"eval_runtime": 72.1924,
|
|
"eval_samples_per_second": 1.233,
|
|
"eval_steps_per_second": 0.166,
|
|
"step": 10000
|
|
},
|
|
{
|
|
"epoch": 4.54,
|
|
"learning_rate": 7.299137539718566e-06,
|
|
"loss": 0.0784,
|
|
"step": 10010
|
|
},
|
|
{
|
|
"epoch": 4.55,
|
|
"learning_rate": 7.22650930549251e-06,
|
|
"loss": 0.0775,
|
|
"step": 10020
|
|
},
|
|
{
|
|
"epoch": 4.55,
|
|
"learning_rate": 7.153881071266455e-06,
|
|
"loss": 0.0812,
|
|
"step": 10030
|
|
},
|
|
{
|
|
"epoch": 4.56,
|
|
"learning_rate": 7.0812528370404e-06,
|
|
"loss": 0.0859,
|
|
"step": 10040
|
|
},
|
|
{
|
|
"epoch": 4.56,
|
|
"learning_rate": 7.0086246028143445e-06,
|
|
"loss": 0.0688,
|
|
"step": 10050
|
|
},
|
|
{
|
|
"epoch": 4.57,
|
|
"learning_rate": 6.9359963685882895e-06,
|
|
"loss": 0.0938,
|
|
"step": 10060
|
|
},
|
|
{
|
|
"epoch": 4.57,
|
|
"learning_rate": 6.863368134362234e-06,
|
|
"loss": 0.0843,
|
|
"step": 10070
|
|
},
|
|
{
|
|
"epoch": 4.58,
|
|
"learning_rate": 6.790739900136179e-06,
|
|
"loss": 0.0816,
|
|
"step": 10080
|
|
},
|
|
{
|
|
"epoch": 4.58,
|
|
"learning_rate": 6.718111665910123e-06,
|
|
"loss": 0.0735,
|
|
"step": 10090
|
|
},
|
|
{
|
|
"epoch": 4.58,
|
|
"learning_rate": 6.645483431684068e-06,
|
|
"loss": 0.0839,
|
|
"step": 10100
|
|
},
|
|
{
|
|
"epoch": 4.58,
|
|
"eval_loss": 0.14849814772605896,
|
|
"eval_mean_accuracy": 0.8015473405558757,
|
|
"eval_mean_iou": 0.6948947776696578,
|
|
"eval_overall_accuracy": 0.9613145121027914,
|
|
"eval_per_category_accuracy": [
|
|
0.9926776055220352,
|
|
0.7325086222251008,
|
|
0.918262352811709,
|
|
0.7271432447097125,
|
|
0.8698908300988094,
|
|
0.7637376953616647,
|
|
0.9021964943782689,
|
|
0.7724086106580105,
|
|
0.3287635799275737,
|
|
0.7487863247863248,
|
|
0.7622518640933326,
|
|
0.9257194593034519,
|
|
0.8912918367713972,
|
|
0.893115463808669,
|
|
0.8630528970512157,
|
|
0.8211973278520042,
|
|
0.8931321029521126,
|
|
0.62171581769437
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9856376280793812,
|
|
0.6757334050307702,
|
|
0.8159314047541198,
|
|
0.6285764540337712,
|
|
0.7863529090700465,
|
|
0.6570806000885276,
|
|
0.8424802428055216,
|
|
0.5606736475518387,
|
|
0.29176805472659656,
|
|
0.5810121497381326,
|
|
0.6037837290301468,
|
|
0.8526052078667846,
|
|
0.8174800201242093,
|
|
0.8123364306465941,
|
|
0.7446264430609864,
|
|
0.7302700726591418,
|
|
0.824526119679351,
|
|
0.297231479107921
|
|
],
|
|
"eval_runtime": 72.5017,
|
|
"eval_samples_per_second": 1.228,
|
|
"eval_steps_per_second": 0.166,
|
|
"step": 10100
|
|
},
|
|
{
|
|
"epoch": 4.59,
|
|
"learning_rate": 6.572855197458012e-06,
|
|
"loss": 0.0899,
|
|
"step": 10110
|
|
},
|
|
{
|
|
"epoch": 4.59,
|
|
"learning_rate": 6.500226963231957e-06,
|
|
"loss": 0.084,
|
|
"step": 10120
|
|
},
|
|
{
|
|
"epoch": 4.6,
|
|
"learning_rate": 6.427598729005901e-06,
|
|
"loss": 0.0784,
|
|
"step": 10130
|
|
},
|
|
{
|
|
"epoch": 4.6,
|
|
"learning_rate": 6.354970494779846e-06,
|
|
"loss": 0.078,
|
|
"step": 10140
|
|
},
|
|
{
|
|
"epoch": 4.61,
|
|
"learning_rate": 6.2823422605537905e-06,
|
|
"loss": 0.0816,
|
|
"step": 10150
|
|
},
|
|
{
|
|
"epoch": 4.61,
|
|
"learning_rate": 6.2097140263277356e-06,
|
|
"loss": 0.0701,
|
|
"step": 10160
|
|
},
|
|
{
|
|
"epoch": 4.62,
|
|
"learning_rate": 6.137085792101681e-06,
|
|
"loss": 0.0771,
|
|
"step": 10170
|
|
},
|
|
{
|
|
"epoch": 4.62,
|
|
"learning_rate": 6.064457557875625e-06,
|
|
"loss": 0.0777,
|
|
"step": 10180
|
|
},
|
|
{
|
|
"epoch": 4.63,
|
|
"learning_rate": 5.99182932364957e-06,
|
|
"loss": 0.092,
|
|
"step": 10190
|
|
},
|
|
{
|
|
"epoch": 4.63,
|
|
"learning_rate": 5.919201089423514e-06,
|
|
"loss": 0.0789,
|
|
"step": 10200
|
|
},
|
|
{
|
|
"epoch": 4.63,
|
|
"eval_loss": 0.14652346074581146,
|
|
"eval_mean_accuracy": 0.8004928725821632,
|
|
"eval_mean_iou": 0.6951126828621068,
|
|
"eval_overall_accuracy": 0.961037710811315,
|
|
"eval_per_category_accuracy": [
|
|
0.9928591579936948,
|
|
0.7318933272883305,
|
|
0.9138329481677121,
|
|
0.730127509495388,
|
|
0.8733523482403892,
|
|
0.7613772362534722,
|
|
0.8967261376567247,
|
|
0.7599229036494319,
|
|
0.3265907915157786,
|
|
0.7421343101343101,
|
|
0.7789746011135376,
|
|
0.9237070686703593,
|
|
0.8920471916982896,
|
|
0.895046900022878,
|
|
0.8626002327987584,
|
|
0.8174010791366907,
|
|
0.8885621449488232,
|
|
0.62171581769437
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9854643947814263,
|
|
0.6756098290087289,
|
|
0.8167550853552833,
|
|
0.6272213482491406,
|
|
0.7792095694014227,
|
|
0.6550510269990957,
|
|
0.841196133669782,
|
|
0.5596696431422216,
|
|
0.2894013019162006,
|
|
0.5838738434109003,
|
|
0.6083786607664178,
|
|
0.853801682490261,
|
|
0.8179704313266324,
|
|
0.8128701002916849,
|
|
0.7456156286245301,
|
|
0.7325895350109094,
|
|
0.8218969685379853,
|
|
0.30545310853530033
|
|
],
|
|
"eval_runtime": 73.6237,
|
|
"eval_samples_per_second": 1.209,
|
|
"eval_steps_per_second": 0.163,
|
|
"step": 10200
|
|
},
|
|
{
|
|
"epoch": 4.63,
|
|
"learning_rate": 5.846572855197459e-06,
|
|
"loss": 0.0784,
|
|
"step": 10210
|
|
},
|
|
{
|
|
"epoch": 4.64,
|
|
"learning_rate": 5.773944620971403e-06,
|
|
"loss": 0.0698,
|
|
"step": 10220
|
|
},
|
|
{
|
|
"epoch": 4.64,
|
|
"learning_rate": 5.701316386745348e-06,
|
|
"loss": 0.0752,
|
|
"step": 10230
|
|
},
|
|
{
|
|
"epoch": 4.65,
|
|
"learning_rate": 5.628688152519292e-06,
|
|
"loss": 0.08,
|
|
"step": 10240
|
|
},
|
|
{
|
|
"epoch": 4.65,
|
|
"learning_rate": 5.5560599182932374e-06,
|
|
"loss": 0.069,
|
|
"step": 10250
|
|
},
|
|
{
|
|
"epoch": 4.66,
|
|
"learning_rate": 5.483431684067182e-06,
|
|
"loss": 0.0816,
|
|
"step": 10260
|
|
},
|
|
{
|
|
"epoch": 4.66,
|
|
"learning_rate": 5.410803449841127e-06,
|
|
"loss": 0.0765,
|
|
"step": 10270
|
|
},
|
|
{
|
|
"epoch": 4.67,
|
|
"learning_rate": 5.338175215615071e-06,
|
|
"loss": 0.0703,
|
|
"step": 10280
|
|
},
|
|
{
|
|
"epoch": 4.67,
|
|
"learning_rate": 5.265546981389016e-06,
|
|
"loss": 0.0827,
|
|
"step": 10290
|
|
},
|
|
{
|
|
"epoch": 4.68,
|
|
"learning_rate": 5.192918747162961e-06,
|
|
"loss": 0.0692,
|
|
"step": 10300
|
|
},
|
|
{
|
|
"epoch": 4.68,
|
|
"eval_loss": 0.14556331932544708,
|
|
"eval_mean_accuracy": 0.7999499482492319,
|
|
"eval_mean_iou": 0.6912676735256027,
|
|
"eval_overall_accuracy": 0.9605477150906337,
|
|
"eval_per_category_accuracy": [
|
|
0.9925063704227338,
|
|
0.733771596042682,
|
|
0.914716765709255,
|
|
0.7370455778621813,
|
|
0.8745688540213646,
|
|
0.7683504166685068,
|
|
0.8941318940588168,
|
|
0.7407301031624316,
|
|
0.31345059493016036,
|
|
0.7543443223443224,
|
|
0.7648881544000472,
|
|
0.9230112322235533,
|
|
0.8903341546319443,
|
|
0.8931066645548458,
|
|
0.8511435161234696,
|
|
0.8249743062692703,
|
|
0.8998744110123337,
|
|
0.6281501340482574
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855265545887547,
|
|
0.6755362759566507,
|
|
0.8157647766027921,
|
|
0.6272076647812536,
|
|
0.7768120941162705,
|
|
0.6519911860926135,
|
|
0.8405735465867566,
|
|
0.5492412928493073,
|
|
0.2809645258520751,
|
|
0.579652925811391,
|
|
0.6032023086411133,
|
|
0.8529403168293681,
|
|
0.8117577610979704,
|
|
0.8102921100741651,
|
|
0.7390242418854488,
|
|
0.7305449850116323,
|
|
0.8277511282677188,
|
|
0.28403442841556553
|
|
],
|
|
"eval_runtime": 72.7214,
|
|
"eval_samples_per_second": 1.224,
|
|
"eval_steps_per_second": 0.165,
|
|
"step": 10300
|
|
},
|
|
{
|
|
"epoch": 4.68,
|
|
"learning_rate": 5.120290512936905e-06,
|
|
"loss": 0.0789,
|
|
"step": 10310
|
|
},
|
|
{
|
|
"epoch": 4.68,
|
|
"learning_rate": 5.04766227871085e-06,
|
|
"loss": 0.0821,
|
|
"step": 10320
|
|
},
|
|
{
|
|
"epoch": 4.69,
|
|
"learning_rate": 4.975034044484794e-06,
|
|
"loss": 0.0674,
|
|
"step": 10330
|
|
},
|
|
{
|
|
"epoch": 4.69,
|
|
"learning_rate": 4.9024058102587385e-06,
|
|
"loss": 0.0747,
|
|
"step": 10340
|
|
},
|
|
{
|
|
"epoch": 4.7,
|
|
"learning_rate": 4.8297775760326835e-06,
|
|
"loss": 0.0791,
|
|
"step": 10350
|
|
},
|
|
{
|
|
"epoch": 4.7,
|
|
"learning_rate": 4.757149341806628e-06,
|
|
"loss": 0.0746,
|
|
"step": 10360
|
|
},
|
|
{
|
|
"epoch": 4.71,
|
|
"learning_rate": 4.684521107580573e-06,
|
|
"loss": 0.0674,
|
|
"step": 10370
|
|
},
|
|
{
|
|
"epoch": 4.71,
|
|
"learning_rate": 4.611892873354517e-06,
|
|
"loss": 0.078,
|
|
"step": 10380
|
|
},
|
|
{
|
|
"epoch": 4.72,
|
|
"learning_rate": 4.539264639128462e-06,
|
|
"loss": 0.0716,
|
|
"step": 10390
|
|
},
|
|
{
|
|
"epoch": 4.72,
|
|
"learning_rate": 4.466636404902406e-06,
|
|
"loss": 0.0704,
|
|
"step": 10400
|
|
},
|
|
{
|
|
"epoch": 4.72,
|
|
"eval_loss": 0.14813955128192902,
|
|
"eval_mean_accuracy": 0.8019139040240622,
|
|
"eval_mean_iou": 0.6936253644894181,
|
|
"eval_overall_accuracy": 0.9608258022351212,
|
|
"eval_per_category_accuracy": [
|
|
0.9925674980804695,
|
|
0.735083145776324,
|
|
0.9134323071420711,
|
|
0.7302631578947368,
|
|
0.8735074132750764,
|
|
0.7665508150097818,
|
|
0.9013752448958997,
|
|
0.7417001298788098,
|
|
0.3291774443869633,
|
|
0.7587496947496948,
|
|
0.7666096126227154,
|
|
0.9250038547757706,
|
|
0.8937647249249141,
|
|
0.8924555197719234,
|
|
0.8613500172443525,
|
|
0.8203044707091469,
|
|
0.9015900738413876,
|
|
0.6309651474530831
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855797066964888,
|
|
0.6743512425543292,
|
|
0.815467107949539,
|
|
0.6274109900355457,
|
|
0.7761384385125846,
|
|
0.6540092575672228,
|
|
0.8437669999055335,
|
|
0.552212751192247,
|
|
0.29070723684210525,
|
|
0.5791302273218666,
|
|
0.6012065480690603,
|
|
0.853086167668164,
|
|
0.8174390058352078,
|
|
0.8141503413565159,
|
|
0.7431675950120421,
|
|
0.7328417307471594,
|
|
0.833127081527442,
|
|
0.2914641320164711
|
|
],
|
|
"eval_runtime": 74.519,
|
|
"eval_samples_per_second": 1.194,
|
|
"eval_steps_per_second": 0.161,
|
|
"step": 10400
|
|
},
|
|
{
|
|
"epoch": 4.73,
|
|
"learning_rate": 4.394008170676351e-06,
|
|
"loss": 0.0734,
|
|
"step": 10410
|
|
},
|
|
{
|
|
"epoch": 4.73,
|
|
"learning_rate": 4.321379936450295e-06,
|
|
"loss": 0.0759,
|
|
"step": 10420
|
|
},
|
|
{
|
|
"epoch": 4.73,
|
|
"learning_rate": 4.24875170222424e-06,
|
|
"loss": 0.0776,
|
|
"step": 10430
|
|
},
|
|
{
|
|
"epoch": 4.74,
|
|
"learning_rate": 4.1761234679981845e-06,
|
|
"loss": 0.0915,
|
|
"step": 10440
|
|
},
|
|
{
|
|
"epoch": 4.74,
|
|
"learning_rate": 4.103495233772129e-06,
|
|
"loss": 0.0803,
|
|
"step": 10450
|
|
},
|
|
{
|
|
"epoch": 4.75,
|
|
"learning_rate": 4.030866999546074e-06,
|
|
"loss": 0.0704,
|
|
"step": 10460
|
|
},
|
|
{
|
|
"epoch": 4.75,
|
|
"learning_rate": 3.958238765320018e-06,
|
|
"loss": 0.0696,
|
|
"step": 10470
|
|
},
|
|
{
|
|
"epoch": 4.76,
|
|
"learning_rate": 3.885610531093963e-06,
|
|
"loss": 0.0731,
|
|
"step": 10480
|
|
},
|
|
{
|
|
"epoch": 4.76,
|
|
"learning_rate": 3.8129822968679076e-06,
|
|
"loss": 0.0765,
|
|
"step": 10490
|
|
},
|
|
{
|
|
"epoch": 4.77,
|
|
"learning_rate": 3.740354062641852e-06,
|
|
"loss": 0.0833,
|
|
"step": 10500
|
|
},
|
|
{
|
|
"epoch": 4.77,
|
|
"eval_loss": 0.1479746252298355,
|
|
"eval_mean_accuracy": 0.799709068394295,
|
|
"eval_mean_iou": 0.6929739311875632,
|
|
"eval_overall_accuracy": 0.9607193336058198,
|
|
"eval_per_category_accuracy": [
|
|
0.9929768093204258,
|
|
0.7261289852491135,
|
|
0.9176381781666116,
|
|
0.7276180141074335,
|
|
0.8688746738316776,
|
|
0.7668378680351,
|
|
0.9009628802193711,
|
|
0.7443591230380442,
|
|
0.336316606311433,
|
|
0.7475164835164835,
|
|
0.7497786696570855,
|
|
0.925845975021053,
|
|
0.8911164865205115,
|
|
0.8976162821392746,
|
|
0.860536299361959,
|
|
0.8214285714285714,
|
|
0.8879110570766463,
|
|
0.6313002680965147
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9854424395099749,
|
|
0.6760280993728895,
|
|
0.8149795746955293,
|
|
0.6285814730181051,
|
|
0.7788548531433264,
|
|
0.6537910520816664,
|
|
0.8440724989445304,
|
|
0.54918256729976,
|
|
0.29528524709302323,
|
|
0.5783960274811237,
|
|
0.6000803035790203,
|
|
0.8532219396492047,
|
|
0.8131010207095737,
|
|
0.8117977081012255,
|
|
0.7413590592342583,
|
|
0.7304840570769213,
|
|
0.8240961371030564,
|
|
0.2947767032829468
|
|
],
|
|
"eval_runtime": 77.3271,
|
|
"eval_samples_per_second": 1.151,
|
|
"eval_steps_per_second": 0.155,
|
|
"step": 10500
|
|
},
|
|
{
|
|
"epoch": 4.77,
|
|
"learning_rate": 3.6677258284157968e-06,
|
|
"loss": 0.0725,
|
|
"step": 10510
|
|
},
|
|
{
|
|
"epoch": 4.78,
|
|
"learning_rate": 3.5950975941897414e-06,
|
|
"loss": 0.0795,
|
|
"step": 10520
|
|
},
|
|
{
|
|
"epoch": 4.78,
|
|
"learning_rate": 3.522469359963686e-06,
|
|
"loss": 0.0694,
|
|
"step": 10530
|
|
},
|
|
{
|
|
"epoch": 4.78,
|
|
"learning_rate": 3.4498411257376306e-06,
|
|
"loss": 0.0786,
|
|
"step": 10540
|
|
},
|
|
{
|
|
"epoch": 4.79,
|
|
"learning_rate": 3.377212891511575e-06,
|
|
"loss": 0.0776,
|
|
"step": 10550
|
|
},
|
|
{
|
|
"epoch": 4.79,
|
|
"learning_rate": 3.3045846572855202e-06,
|
|
"loss": 0.0772,
|
|
"step": 10560
|
|
},
|
|
{
|
|
"epoch": 4.8,
|
|
"learning_rate": 3.231956423059465e-06,
|
|
"loss": 0.0785,
|
|
"step": 10570
|
|
},
|
|
{
|
|
"epoch": 4.8,
|
|
"learning_rate": 3.1593281888334094e-06,
|
|
"loss": 0.0695,
|
|
"step": 10580
|
|
},
|
|
{
|
|
"epoch": 4.81,
|
|
"learning_rate": 3.086699954607354e-06,
|
|
"loss": 0.0665,
|
|
"step": 10590
|
|
},
|
|
{
|
|
"epoch": 4.81,
|
|
"learning_rate": 3.0140717203812986e-06,
|
|
"loss": 0.0666,
|
|
"step": 10600
|
|
},
|
|
{
|
|
"epoch": 4.81,
|
|
"eval_loss": 0.14828309416770935,
|
|
"eval_mean_accuracy": 0.8017485365453704,
|
|
"eval_mean_iou": 0.6961283653552043,
|
|
"eval_overall_accuracy": 0.9611695964684647,
|
|
"eval_per_category_accuracy": [
|
|
0.9927109428380961,
|
|
0.7328972295535874,
|
|
0.9128700341146693,
|
|
0.7346717308735757,
|
|
0.8750923700455521,
|
|
0.768122982348447,
|
|
0.8967122181739726,
|
|
0.7593141531775149,
|
|
0.33678220382824625,
|
|
0.7482881562881563,
|
|
0.7686065041610104,
|
|
0.9261741251635809,
|
|
0.8872318040393504,
|
|
0.8957376414480052,
|
|
0.8616032936713226,
|
|
0.8175488180883864,
|
|
0.8934499325769194,
|
|
0.6236595174262735
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855292226790823,
|
|
0.6760817936040867,
|
|
0.8162783331744533,
|
|
0.6303905022405867,
|
|
0.7806234430754729,
|
|
0.6596454367378207,
|
|
0.8434200857869009,
|
|
0.5597664260793139,
|
|
0.295882192527952,
|
|
0.579330268013794,
|
|
0.6043016906680692,
|
|
0.8535430087154225,
|
|
0.8135113185223052,
|
|
0.8111767190201844,
|
|
0.7431131685234503,
|
|
0.7315974018508938,
|
|
0.8272586234975129,
|
|
0.31886094167637585
|
|
],
|
|
"eval_runtime": 77.2989,
|
|
"eval_samples_per_second": 1.151,
|
|
"eval_steps_per_second": 0.155,
|
|
"step": 10600
|
|
},
|
|
{
|
|
"epoch": 4.82,
|
|
"learning_rate": 2.9414434861552432e-06,
|
|
"loss": 0.0955,
|
|
"step": 10610
|
|
},
|
|
{
|
|
"epoch": 4.82,
|
|
"learning_rate": 2.868815251929188e-06,
|
|
"loss": 0.0779,
|
|
"step": 10620
|
|
},
|
|
{
|
|
"epoch": 4.83,
|
|
"learning_rate": 2.7961870177031325e-06,
|
|
"loss": 0.0716,
|
|
"step": 10630
|
|
},
|
|
{
|
|
"epoch": 4.83,
|
|
"learning_rate": 2.723558783477077e-06,
|
|
"loss": 0.083,
|
|
"step": 10640
|
|
},
|
|
{
|
|
"epoch": 4.83,
|
|
"learning_rate": 2.6509305492510217e-06,
|
|
"loss": 0.0848,
|
|
"step": 10650
|
|
},
|
|
{
|
|
"epoch": 4.84,
|
|
"learning_rate": 2.5783023150249663e-06,
|
|
"loss": 0.0807,
|
|
"step": 10660
|
|
},
|
|
{
|
|
"epoch": 4.84,
|
|
"learning_rate": 2.505674080798911e-06,
|
|
"loss": 0.0868,
|
|
"step": 10670
|
|
},
|
|
{
|
|
"epoch": 4.85,
|
|
"learning_rate": 2.4330458465728555e-06,
|
|
"loss": 0.0808,
|
|
"step": 10680
|
|
},
|
|
{
|
|
"epoch": 4.85,
|
|
"learning_rate": 2.3604176123468e-06,
|
|
"loss": 0.0851,
|
|
"step": 10690
|
|
},
|
|
{
|
|
"epoch": 4.86,
|
|
"learning_rate": 2.2877893781207447e-06,
|
|
"loss": 0.071,
|
|
"step": 10700
|
|
},
|
|
{
|
|
"epoch": 4.86,
|
|
"eval_loss": 0.14659462869167328,
|
|
"eval_mean_accuracy": 0.8008380411850532,
|
|
"eval_mean_iou": 0.6962789687736569,
|
|
"eval_overall_accuracy": 0.9611141333419285,
|
|
"eval_per_category_accuracy": [
|
|
0.992941419623842,
|
|
0.7355689049369323,
|
|
0.9125588065368109,
|
|
0.735485621269669,
|
|
0.8745901240039987,
|
|
0.7635787121476424,
|
|
0.8998684608879934,
|
|
0.7497475582241827,
|
|
0.33652353854112776,
|
|
0.7513748473748474,
|
|
0.7542741348442819,
|
|
0.9236477644277338,
|
|
0.8949292304372066,
|
|
0.8923235309645742,
|
|
0.8587094757716848,
|
|
0.8168743576567318,
|
|
0.9007075566307591,
|
|
0.6213806970509383
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9854991646107671,
|
|
0.6764246042972647,
|
|
0.8162711290893151,
|
|
0.6299889618311741,
|
|
0.7792439461713443,
|
|
0.6550150680286927,
|
|
0.8416806760981469,
|
|
0.5569122436027313,
|
|
0.2950380986937591,
|
|
0.5779741376071651,
|
|
0.6005717642451537,
|
|
0.8539996563837682,
|
|
0.8179121037167923,
|
|
0.8140430023801019,
|
|
0.7433952405611305,
|
|
0.7322987446735,
|
|
0.8326487183436881,
|
|
0.3241041775913302
|
|
],
|
|
"eval_runtime": 72.8009,
|
|
"eval_samples_per_second": 1.223,
|
|
"eval_steps_per_second": 0.165,
|
|
"step": 10700
|
|
},
|
|
{
|
|
"epoch": 4.86,
|
|
"learning_rate": 2.2151611438946893e-06,
|
|
"loss": 0.0792,
|
|
"step": 10710
|
|
},
|
|
{
|
|
"epoch": 4.87,
|
|
"learning_rate": 2.142532909668634e-06,
|
|
"loss": 0.0867,
|
|
"step": 10720
|
|
},
|
|
{
|
|
"epoch": 4.87,
|
|
"learning_rate": 2.0699046754425785e-06,
|
|
"loss": 0.0793,
|
|
"step": 10730
|
|
},
|
|
{
|
|
"epoch": 4.88,
|
|
"learning_rate": 1.997276441216523e-06,
|
|
"loss": 0.0739,
|
|
"step": 10740
|
|
},
|
|
{
|
|
"epoch": 4.88,
|
|
"learning_rate": 1.9246482069904677e-06,
|
|
"loss": 0.0698,
|
|
"step": 10750
|
|
},
|
|
{
|
|
"epoch": 4.88,
|
|
"learning_rate": 1.8520199727644123e-06,
|
|
"loss": 0.0847,
|
|
"step": 10760
|
|
},
|
|
{
|
|
"epoch": 4.89,
|
|
"learning_rate": 1.779391738538357e-06,
|
|
"loss": 0.0827,
|
|
"step": 10770
|
|
},
|
|
{
|
|
"epoch": 4.89,
|
|
"learning_rate": 1.7067635043123015e-06,
|
|
"loss": 0.073,
|
|
"step": 10780
|
|
},
|
|
{
|
|
"epoch": 4.9,
|
|
"learning_rate": 1.6341352700862462e-06,
|
|
"loss": 0.0716,
|
|
"step": 10790
|
|
},
|
|
{
|
|
"epoch": 4.9,
|
|
"learning_rate": 1.5615070358601908e-06,
|
|
"loss": 0.0727,
|
|
"step": 10800
|
|
},
|
|
{
|
|
"epoch": 4.9,
|
|
"eval_loss": 0.14799156785011292,
|
|
"eval_mean_accuracy": 0.8013999895039033,
|
|
"eval_mean_iou": 0.6942882071400502,
|
|
"eval_overall_accuracy": 0.960845904403858,
|
|
"eval_per_category_accuracy": [
|
|
0.992796088894924,
|
|
0.7337392120986415,
|
|
0.9124745515571696,
|
|
0.7321622354856213,
|
|
0.875763403691234,
|
|
0.763996043084451,
|
|
0.8989167162548239,
|
|
0.733038531917853,
|
|
0.344593895499224,
|
|
0.749030525030525,
|
|
0.7541364181864684,
|
|
0.9257748099299024,
|
|
0.89742010323184,
|
|
0.8935730250074794,
|
|
0.8617056820141403,
|
|
0.8198805241521069,
|
|
0.9070425490553055,
|
|
0.6291554959785522
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855499280467153,
|
|
0.6765452373842938,
|
|
0.8160740126747936,
|
|
0.6306595781971139,
|
|
0.7764124356504543,
|
|
0.6527826516789489,
|
|
0.8405210000911066,
|
|
0.5506681756254308,
|
|
0.2954928577765948,
|
|
0.5789724036392465,
|
|
0.6016685109756001,
|
|
0.8536446656094493,
|
|
0.8161573784434712,
|
|
0.8133254310862653,
|
|
0.7438929646394395,
|
|
0.7333804483975133,
|
|
0.8340819644686327,
|
|
0.29735808413583376
|
|
],
|
|
"eval_runtime": 77.2539,
|
|
"eval_samples_per_second": 1.152,
|
|
"eval_steps_per_second": 0.155,
|
|
"step": 10800
|
|
},
|
|
{
|
|
"epoch": 4.91,
|
|
"learning_rate": 1.4888788016341354e-06,
|
|
"loss": 0.0762,
|
|
"step": 10810
|
|
},
|
|
{
|
|
"epoch": 4.91,
|
|
"learning_rate": 1.4162505674080802e-06,
|
|
"loss": 0.0863,
|
|
"step": 10820
|
|
},
|
|
{
|
|
"epoch": 4.92,
|
|
"learning_rate": 1.3436223331820248e-06,
|
|
"loss": 0.0763,
|
|
"step": 10830
|
|
},
|
|
{
|
|
"epoch": 4.92,
|
|
"learning_rate": 1.2709940989559694e-06,
|
|
"loss": 0.0731,
|
|
"step": 10840
|
|
},
|
|
{
|
|
"epoch": 4.93,
|
|
"learning_rate": 1.1983658647299138e-06,
|
|
"loss": 0.0808,
|
|
"step": 10850
|
|
},
|
|
{
|
|
"epoch": 4.93,
|
|
"learning_rate": 1.1257376305038586e-06,
|
|
"loss": 0.0837,
|
|
"step": 10860
|
|
},
|
|
{
|
|
"epoch": 4.93,
|
|
"learning_rate": 1.0531093962778032e-06,
|
|
"loss": 0.0825,
|
|
"step": 10870
|
|
},
|
|
{
|
|
"epoch": 4.94,
|
|
"learning_rate": 9.804811620517476e-07,
|
|
"loss": 0.0758,
|
|
"step": 10880
|
|
},
|
|
{
|
|
"epoch": 4.94,
|
|
"learning_rate": 9.078529278256923e-07,
|
|
"loss": 0.0887,
|
|
"step": 10890
|
|
},
|
|
{
|
|
"epoch": 4.95,
|
|
"learning_rate": 8.352246935996369e-07,
|
|
"loss": 0.0735,
|
|
"step": 10900
|
|
},
|
|
{
|
|
"epoch": 4.95,
|
|
"eval_loss": 0.14808058738708496,
|
|
"eval_mean_accuracy": 0.8014247245983838,
|
|
"eval_mean_iou": 0.6946340248335119,
|
|
"eval_overall_accuracy": 0.9609660030750746,
|
|
"eval_per_category_accuracy": [
|
|
0.9928440147536206,
|
|
0.731634255736006,
|
|
0.914145895234951,
|
|
0.7319587628865979,
|
|
0.8729063647335454,
|
|
0.7649786476711167,
|
|
0.9010411773098511,
|
|
0.7426069332524075,
|
|
0.34319710294878425,
|
|
0.7453089133089134,
|
|
0.7533199551437172,
|
|
0.9252252572815726,
|
|
0.8937917018865889,
|
|
0.8945321436742164,
|
|
0.8619050698396276,
|
|
0.8202081192189106,
|
|
0.9052096286947694,
|
|
0.6308310991957105
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855369404856755,
|
|
0.6766247379454927,
|
|
0.8163059116530241,
|
|
0.6302633884249256,
|
|
0.7772533853856423,
|
|
0.6539899119561442,
|
|
0.8418459327607872,
|
|
0.5540958550165581,
|
|
0.2964253798033959,
|
|
0.5797948328267477,
|
|
0.6018815429595398,
|
|
0.853853872115297,
|
|
0.8169328955317111,
|
|
0.814172332876833,
|
|
0.7441031700985363,
|
|
0.7332648057011927,
|
|
0.8343330403893027,
|
|
0.29273451107240606
|
|
],
|
|
"eval_runtime": 77.3655,
|
|
"eval_samples_per_second": 1.15,
|
|
"eval_steps_per_second": 0.155,
|
|
"step": 10900
|
|
},
|
|
{
|
|
"epoch": 4.95,
|
|
"learning_rate": 7.625964593735815e-07,
|
|
"loss": 0.0855,
|
|
"step": 10910
|
|
},
|
|
{
|
|
"epoch": 4.96,
|
|
"learning_rate": 6.899682251475262e-07,
|
|
"loss": 0.0784,
|
|
"step": 10920
|
|
},
|
|
{
|
|
"epoch": 4.96,
|
|
"learning_rate": 6.173399909214707e-07,
|
|
"loss": 0.0872,
|
|
"step": 10930
|
|
},
|
|
{
|
|
"epoch": 4.97,
|
|
"learning_rate": 5.447117566954155e-07,
|
|
"loss": 0.077,
|
|
"step": 10940
|
|
},
|
|
{
|
|
"epoch": 4.97,
|
|
"learning_rate": 4.7208352246936e-07,
|
|
"loss": 0.0806,
|
|
"step": 10950
|
|
},
|
|
{
|
|
"epoch": 4.98,
|
|
"learning_rate": 3.994552882433046e-07,
|
|
"loss": 0.0794,
|
|
"step": 10960
|
|
},
|
|
{
|
|
"epoch": 4.98,
|
|
"learning_rate": 3.2682705401724927e-07,
|
|
"loss": 0.0754,
|
|
"step": 10970
|
|
},
|
|
{
|
|
"epoch": 4.98,
|
|
"learning_rate": 2.541988197911938e-07,
|
|
"loss": 0.0779,
|
|
"step": 10980
|
|
},
|
|
{
|
|
"epoch": 4.99,
|
|
"learning_rate": 1.8157058556513846e-07,
|
|
"loss": 0.0931,
|
|
"step": 10990
|
|
},
|
|
{
|
|
"epoch": 4.99,
|
|
"learning_rate": 1.0894235133908308e-07,
|
|
"loss": 0.0853,
|
|
"step": 11000
|
|
},
|
|
{
|
|
"epoch": 4.99,
|
|
"eval_loss": 0.1483118236064911,
|
|
"eval_mean_accuracy": 0.8013773443925859,
|
|
"eval_mean_iou": 0.6942808214048504,
|
|
"eval_overall_accuracy": 0.9608961812565836,
|
|
"eval_per_category_accuracy": [
|
|
0.99286914254759,
|
|
0.7324762382810602,
|
|
0.914249064597777,
|
|
0.7332474226804123,
|
|
0.8738827255492974,
|
|
0.7648064158559259,
|
|
0.8971837406521974,
|
|
0.738701537049782,
|
|
0.34500775995861355,
|
|
0.743003663003663,
|
|
0.754480709831002,
|
|
0.9248298956640691,
|
|
0.8952934194198154,
|
|
0.896274395931225,
|
|
0.8613392395240559,
|
|
0.8215313463514903,
|
|
0.9057897449031546,
|
|
0.6298257372654156
|
|
],
|
|
"eval_per_category_iou": [
|
|
0.9855223246429033,
|
|
0.6765220512360357,
|
|
0.8160509554140127,
|
|
0.6297914482115811,
|
|
0.7767318840844862,
|
|
0.6542057486939079,
|
|
0.8407320751640647,
|
|
0.5527613058975714,
|
|
0.29650542415080916,
|
|
0.5793839404049175,
|
|
0.6019384711976142,
|
|
0.8538285772687122,
|
|
0.8154169345492815,
|
|
0.8131239148222803,
|
|
0.7433172737080993,
|
|
0.7330964117849363,
|
|
0.835010226344999,
|
|
0.2931158177110952
|
|
],
|
|
"eval_runtime": 72.1878,
|
|
"eval_samples_per_second": 1.233,
|
|
"eval_steps_per_second": 0.166,
|
|
"step": 11000
|
|
}
|
|
],
|
|
"max_steps": 11015,
|
|
"num_train_epochs": 5,
|
|
"total_flos": 1.1357438206053188e+19,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|