{
  "official_run_id": "official-run",
  "datasets": [
    "final-test-500"
  ],
  "results": [
    {
      "rank": 1,
      "team_id": "team_013",
      "team_alias": "Kaizen",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.8515625,
      "precision": 0.9121338912133892,
      "recall": 0.7985347985347986,
      "latency_ms_mean": 1051.166011,
      "latency_ms_total": 525583.0055
    },
    {
      "rank": 2,
      "team_id": "team_004",
      "team_alias": "Vita",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.8215767634854771,
      "precision": 0.9473684210526315,
      "recall": 0.7252747252747253,
      "latency_ms_mean": 2740.6195728000002,
      "latency_ms_total": 1370309.7864
    },
    {
      "rank": 3,
      "team_id": "team_042",
      "team_alias": "Mosaic Minds",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.8214971209213052,
      "precision": 0.8629032258064516,
      "recall": 0.7838827838827839,
      "latency_ms_mean": 1247.5215218,
      "latency_ms_total": 623760.7609
    },
    {
      "rank": 4,
      "team_id": "team_044",
      "team_alias": "Lowkey Critical",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.8121442125237192,
      "precision": 0.84251968503937,
      "recall": 0.7838827838827839,
      "latency_ms_mean": 1865.5007802,
      "latency_ms_total": 932750.3901
    },
    {
      "rank": 5,
      "team_id": "team_025",
      "team_alias": "SafeMind Red Team",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.8042704626334519,
      "precision": 0.7820069204152249,
      "recall": 0.8278388278388278,
      "latency_ms_mean": 1637.0578022,
      "latency_ms_total": 818528.9011
    },
    {
      "rank": 6,
      "team_id": "team_078",
      "team_alias": "MindBloom",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.8016528925619835,
      "precision": 0.919431279620853,
      "recall": 0.7106227106227107,
      "latency_ms_mean": 774.204456,
      "latency_ms_total": 387102.228
    },
    {
      "rank": 7,
      "team_id": "team_071",
      "team_alias": "Wingspan",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.8013937282229965,
      "precision": 0.7641196013289037,
      "recall": 0.8424908424908425,
      "latency_ms_mean": 1735.2816306,
      "latency_ms_total": 867640.8153
    },
    {
      "rank": 8,
      "team_id": "team_034",
      "team_alias": "Low Battery",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7747368421052632,
      "precision": 0.9108910891089109,
      "recall": 0.673992673992674,
      "latency_ms_mean": 5977.716546199999,
      "latency_ms_total": 2988858.2731
    },
    {
      "rank": 9,
      "team_id": "team_053",
      "team_alias": "Mindlift",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7623762376237624,
      "precision": 0.6936936936936937,
      "recall": 0.8461538461538461,
      "latency_ms_mean": 1165.1472296,
      "latency_ms_total": 582573.6148
    },
    {
      "rank": 10,
      "team_id": "team_045",
      "team_alias": "AI Guardians",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7401574803149606,
      "precision": 0.8,
      "recall": 0.6886446886446886,
      "latency_ms_mean": 1459.5694647999999,
      "latency_ms_total": 729784.7324
    },
    {
      "rank": 11,
      "team_id": "team_036",
      "team_alias": "S.A.I.F.E",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7325842696629213,
      "precision": 0.9476744186046512,
      "recall": 0.5970695970695971,
      "latency_ms_mean": 6652.9502088,
      "latency_ms_total": 3326475.1044
    },
    {
      "rank": 12,
      "team_id": "team_055",
      "team_alias": "UdeM AI",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7284768211920529,
      "precision": 0.9166666666666666,
      "recall": 0.6043956043956044,
      "latency_ms_mean": 6217.7779611999995,
      "latency_ms_total": 3108888.9806
    },
    {
      "rank": 13,
      "team_id": "team_016",
      "team_alias": "YouthGuard",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7283464566929134,
      "precision": 0.7872340425531915,
      "recall": 0.6776556776556777,
      "latency_ms_mean": 4008.0239986,
      "latency_ms_total": 2004011.9993
    },
    {
      "rank": 14,
      "team_id": "team_020",
      "team_alias": "MindCraft",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7219796215429404,
      "precision": 0.5990338164251208,
      "recall": 0.9084249084249084,
      "latency_ms_mean": 7820.5219116,
      "latency_ms_total": 3910260.9558
    },
    {
      "rank": 15,
      "team_id": "team_021",
      "team_alias": "404HarmNotFound",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7183673469387755,
      "precision": 0.8110599078341014,
      "recall": 0.6446886446886447,
      "latency_ms_mean": 2033.955513,
      "latency_ms_total": 1016977.7565
    },
    {
      "rank": 16,
      "team_id": "team_040",
      "team_alias": "MindGuardiansAI",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7170626349892009,
      "precision": 0.8736842105263158,
      "recall": 0.608058608058608,
      "latency_ms_mean": 4729.1521566,
      "latency_ms_total": 2364576.0783
    },
    {
      "rank": 17,
      "team_id": "team_072",
      "team_alias": "Bell Aura",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.7081081081081081,
      "precision": 0.5610278372591007,
      "recall": 0.9597069597069597,
      "latency_ms_mean": 19.895585200000003,
      "latency_ms_total": 9947.7926
    },
    {
      "rank": 18,
      "team_id": "team_064",
      "team_alias": "Bogatyri",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6926536731634183,
      "precision": 0.5862944162436549,
      "recall": 0.8461538461538461,
      "latency_ms_mean": 639.6123708,
      "latency_ms_total": 319806.1854
    },
    {
      "rank": 19,
      "team_id": "team_048",
      "team_alias": "Thrive Hacks",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6888888888888889,
      "precision": 0.6966292134831461,
      "recall": 0.6813186813186813,
      "latency_ms_mean": 5431.1664398,
      "latency_ms_total": 2715583.2199
    },
    {
      "rank": 20,
      "team_id": "team_007",
      "team_alias": "LES CURIEUX",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6822429906542056,
      "precision": 0.9419354838709677,
      "recall": 0.5347985347985348,
      "latency_ms_mean": 1951.8980896,
      "latency_ms_total": 975949.0448
    },
    {
      "rank": 21,
      "team_id": "team_067",
      "team_alias": "The Falcons",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6743119266055045,
      "precision": 0.901840490797546,
      "recall": 0.5384615384615384,
      "latency_ms_mean": 9360.7566006,
      "latency_ms_total": 4680378.3003
    },
    {
      "rank": 22,
      "team_id": "team_073",
      "team_alias": "Wink",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6713286713286714,
      "precision": 0.6421404682274248,
      "recall": 0.7032967032967034,
      "latency_ms_mean": 3795.5362148000004,
      "latency_ms_total": 1897768.1074
    },
    {
      "rank": 23,
      "team_id": "team_001",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6610169491525424,
      "precision": 0.7839195979899497,
      "recall": 0.5714285714285714,
      "latency_ms_mean": 36.151243799999996,
      "latency_ms_total": 18075.6219
    },
    {
      "rank": 24,
      "team_id": "team_019",
      "team_alias": "NeuraAegis",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6558044806517311,
      "precision": 0.7385321100917431,
      "recall": 0.5897435897435898,
      "latency_ms_mean": 6374.5199688,
      "latency_ms_total": 3187259.9844
    },
    {
      "rank": 25,
      "team_id": "team_018",
      "team_alias": "Mindful Machines",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6506024096385542,
      "precision": 0.9507042253521126,
      "recall": 0.4945054945054945,
      "latency_ms_mean": 4408.286542999999,
      "latency_ms_total": 2204143.2715
    },
    {
      "rank": 26,
      "team_id": "team_026",
      "team_alias": "Kindly Wired",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6426858513189448,
      "precision": 0.9305555555555556,
      "recall": 0.4908424908424908,
      "latency_ms_mean": 3822.770429,
      "latency_ms_total": 1911385.2145
    },
    {
      "rank": 27,
      "team_id": "team_037",
      "team_alias": "YoungAI",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6203473945409429,
      "precision": 0.9615384615384616,
      "recall": 0.45787545787545786,
      "latency_ms_mean": 2299.5310408,
      "latency_ms_total": 1149765.5204
    },
    {
      "rank": 28,
      "team_id": "team_027",
      "team_alias": "Code & Cope",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.6190476190476191,
      "precision": 0.8843537414965986,
      "recall": 0.47619047619047616,
      "latency_ms_mean": 156.58347260000002,
      "latency_ms_total": 78291.7363
    },
    {
      "rank": 29,
      "team_id": "team_076",
      "team_alias": "Unravel",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.606694560669456,
      "precision": 0.7073170731707317,
      "recall": 0.5311355311355311,
      "latency_ms_mean": 1043.1194346,
      "latency_ms_total": 521559.7173
    },
    {
      "rank": 30,
      "team_id": "team_058",
      "team_alias": "MTL",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.606694560669456,
      "precision": 0.7073170731707317,
      "recall": 0.5311355311355311,
      "latency_ms_mean": 4960.476866,
      "latency_ms_total": 2480238.433
    },
    {
      "rank": 31,
      "team_id": "team_014",
      "team_alias": "SafeMindAI",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.60625,
      "precision": 0.5286103542234333,
      "recall": 0.7106227106227107,
      "latency_ms_mean": 14.4392672,
      "latency_ms_total": 7219.6336
    },
    {
      "rank": 32,
      "team_id": "team_032",
      "team_alias": "The Cortex Guardians",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5954198473282443,
      "precision": 0.975,
      "recall": 0.42857142857142855,
      "latency_ms_mean": 3238.572004,
      "latency_ms_total": 1619286.002
    },
    {
      "rank": 33,
      "team_id": "team_010",
      "team_alias": "LUNA",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5918367346938775,
      "precision": 0.9747899159663865,
      "recall": 0.4249084249084249,
      "latency_ms_mean": 6258.534238,
      "latency_ms_total": 3129267.119
    },
    {
      "rank": 34,
      "team_id": "team_030",
      "team_alias": "GoldenMindTech",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5824742268041238,
      "precision": 0.9826086956521739,
      "recall": 0.4139194139194139,
      "latency_ms_mean": 6359.1331398,
      "latency_ms_total": 3179566.5699
    },
    {
      "rank": 35,
      "team_id": "team_009",
      "team_alias": "MindHack",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5731225296442688,
      "precision": 0.6223175965665236,
      "recall": 0.5311355311355311,
      "latency_ms_mean": 288.0393212,
      "latency_ms_total": 144019.6606
    },
    {
      "rank": 36,
      "team_id": "team_047",
      "team_alias": "Team YMH",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5688487584650113,
      "precision": 0.7411764705882353,
      "recall": 0.46153846153846156,
      "latency_ms_mean": 57.1459344,
      "latency_ms_total": 28572.9672
    },
    {
      "rank": 37,
      "team_id": "team_057",
      "team_alias": "Cauchy Enjoyers",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5644444444444444,
      "precision": 0.7175141242937854,
      "recall": 0.4652014652014652,
      "latency_ms_mean": 736.9854466,
      "latency_ms_total": 368492.7233
    },
    {
      "rank": 38,
      "team_id": "team_049",
      "team_alias": "MindSafe",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5435684647302904,
      "precision": 0.6267942583732058,
      "recall": 0.47985347985347987,
      "latency_ms_mean": 19.1166086,
      "latency_ms_total": 9558.3043
    },
    {
      "rank": 39,
      "team_id": "team_081",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5416666666666666,
      "precision": 0.9369369369369369,
      "recall": 0.38095238095238093,
      "latency_ms_mean": 5282.8925848,
      "latency_ms_total": 2641446.2924
    },
    {
      "rank": 40,
      "team_id": "team_051",
      "team_alias": "Galactic Pigeons",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5283018867924528,
      "precision": 0.5447470817120622,
      "recall": 0.5128205128205128,
      "latency_ms_mean": 349.6554162,
      "latency_ms_total": 174827.7081
    },
    {
      "rank": 41,
      "team_id": "team_043",
      "team_alias": "Ayida",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5183673469387755,
      "precision": 0.5852534562211982,
      "recall": 0.4652014652014652,
      "latency_ms_mean": 12.3193208,
      "latency_ms_total": 6159.6604
    },
    {
      "rank": 42,
      "team_id": "team_022",
      "team_alias": "CTRL + CARE",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.510752688172043,
      "precision": 0.9595959595959596,
      "recall": 0.34798534798534797,
      "latency_ms_mean": 6246.5008048,
      "latency_ms_total": 3123250.4024
    },
    {
      "rank": 43,
      "team_id": "team_068",
      "team_alias": "West_FL",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.5,
      "precision": 0.6903225806451613,
      "recall": 0.39194139194139194,
      "latency_ms_mean": 1.5571388,
      "latency_ms_total": 778.5694
    },
    {
      "rank": 44,
      "team_id": "team_006",
      "team_alias": "Liminal",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.49473684210526314,
      "precision": 0.8785046728971962,
      "recall": 0.3443223443223443,
      "latency_ms_mean": 162.92913579999998,
      "latency_ms_total": 81464.5679
    },
    {
      "rank": 45,
      "team_id": "team_046",
      "team_alias": "MindGuardians",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.48501362397820164,
      "precision": 0.9468085106382979,
      "recall": 0.326007326007326,
      "latency_ms_mean": 694.5574054,
      "latency_ms_total": 347278.7027
    },
    {
      "rank": 46,
      "team_id": "team_003",
      "team_alias": "Prefrontal Guard",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.4797843665768194,
      "precision": 0.9081632653061225,
      "recall": 0.326007326007326,
      "latency_ms_mean": 2867.7173974,
      "latency_ms_total": 1433858.6987
    },
    {
      "rank": 47,
      "team_id": "team_012",
      "team_alias": "Breathe Easy",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.4709141274238227,
      "precision": 0.9659090909090909,
      "recall": 0.31135531135531136,
      "latency_ms_mean": 6979.5679822,
      "latency_ms_total": 3489783.9911
    },
    {
      "rank": 48,
      "team_id": "team_029",
      "team_alias": "MindTrace",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.44562334217506633,
      "precision": 0.8076923076923077,
      "recall": 0.3076923076923077,
      "latency_ms_mean": 8.505867,
      "latency_ms_total": 4252.9335
    },
    {
      "rank": 49,
      "team_id": "team_070",
      "team_alias": "Touch-Starved",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.43454038997214484,
      "precision": 0.9069767441860465,
      "recall": 0.2857142857142857,
      "latency_ms_mean": 1656.537174,
      "latency_ms_total": 828268.587
    },
    {
      "rank": 50,
      "team_id": "team_062",
      "team_alias": "NewfoundAI",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.43304843304843305,
      "precision": 0.9743589743589743,
      "recall": 0.2783882783882784,
      "latency_ms_mean": 4968.0553722,
      "latency_ms_total": 2484027.6861
    },
    {
      "rank": 51,
      "team_id": "team_080",
      "team_alias": "Shakemyhead",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.43103448275862066,
      "precision": 1.0,
      "recall": 0.27472527472527475,
      "latency_ms_mean": 2202.7741194,
      "latency_ms_total": 1101387.0597
    },
    {
      "rank": 52,
      "team_id": "team_008",
      "team_alias": "Cyberpunk",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.4134078212290503,
      "precision": 0.8705882352941177,
      "recall": 0.27106227106227104,
      "latency_ms_mean": 458.68986359999997,
      "latency_ms_total": 229344.9318
    },
    {
      "rank": 53,
      "team_id": "team_039",
      "team_alias": "Haven",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.3409090909090909,
      "precision": 0.759493670886076,
      "recall": 0.21978021978021978,
      "latency_ms_mean": 7.9206692,
      "latency_ms_total": 3960.3346
    },
    {
      "rank": 54,
      "team_id": "team_059",
      "team_alias": "DNA",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.0989546,
      "latency_ms_total": 49.4773
    },
    {
      "rank": 55,
      "team_id": "team_065",
      "team_alias": "PulseOfMind",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.11745839999999999,
      "latency_ms_total": 58.7292
    },
    {
      "rank": 56,
      "team_id": "team_075",
      "team_alias": "QuadCore",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1176436,
      "latency_ms_total": 58.8218
    },
    {
      "rank": 57,
      "team_id": "team_069",
      "team_alias": "BDBM",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1178836,
      "latency_ms_total": 58.9418
    },
    {
      "rank": 58,
      "team_id": "team_066",
      "team_alias": "MARVIN-MINDPULSE",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1180446,
      "latency_ms_total": 59.0223
    },
    {
      "rank": 59,
      "team_id": "team_024",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.11807720000000001,
      "latency_ms_total": 59.0386
    },
    {
      "rank": 60,
      "team_id": "team_033",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1181054,
      "latency_ms_total": 59.0527
    },
    {
      "rank": 61,
      "team_id": "team_015",
      "team_alias": "Team Illuminati",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1182082,
      "latency_ms_total": 59.1041
    },
    {
      "rank": 62,
      "team_id": "team_077",
      "team_alias": "Law One",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.119659,
      "latency_ms_total": 59.8295
    },
    {
      "rank": 63,
      "team_id": "team_060",
      "team_alias": "Kangad",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1198038,
      "latency_ms_total": 59.9019
    },
    {
      "rank": 64,
      "team_id": "team_028",
      "team_alias": "SMH AI",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1199416,
      "latency_ms_total": 59.9708
    },
    {
      "rank": 65,
      "team_id": "team_038",
      "team_alias": "Sentinelle",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1211124,
      "latency_ms_total": 60.5562
    },
    {
      "rank": 66,
      "team_id": "team_005",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1213022,
      "latency_ms_total": 60.6511
    },
    {
      "rank": 67,
      "team_id": "team_011",
      "team_alias": "SaferYouth AI",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1271206,
      "latency_ms_total": 63.5603
    },
    {
      "rank": 68,
      "team_id": "team_054",
      "team_alias": "Soup for the Soul",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1275416,
      "latency_ms_total": 63.7708
    },
    {
      "rank": 69,
      "team_id": "team_041",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.1280268,
      "latency_ms_total": 64.0134
    },
    {
      "rank": 70,
      "team_id": "team_002",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.133152,
      "latency_ms_total": 66.576
    },
    {
      "rank": 71,
      "team_id": "team_017",
      "team_alias": "STEMinist",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.160662,
      "latency_ms_total": 80.331
    },
    {
      "rank": 72,
      "team_id": "team_035",
      "team_alias": "Safeguard",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.17023660000000002,
      "latency_ms_total": 85.1183
    },
    {
      "rank": 73,
      "team_id": "team_056",
      "team_alias": "404:NameNotFound",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.18369939999999998,
      "latency_ms_total": 91.8497
    },
    {
      "rank": 74,
      "team_id": "team_050",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.21829220000000002,
      "latency_ms_total": 109.1461
    },
    {
      "rank": 75,
      "team_id": "team_061",
      "team_alias": "3n1t",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.24245260000000002,
      "latency_ms_total": 121.22630000000001
    },
    {
      "rank": 76,
      "team_id": "team_031",
      "team_alias": null,
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.2655888,
      "latency_ms_total": 132.7944
    },
    {
      "rank": 77,
      "team_id": "team_079",
      "team_alias": "Serotonin Squad",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.2735136,
      "latency_ms_total": 136.7568
    },
    {
      "rank": 78,
      "team_id": "team_063",
      "team_alias": "The Bridge Builders",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.2988408,
      "latency_ms_total": 149.4204
    },
    {
      "rank": 79,
      "team_id": "team_023",
      "team_alias": "Sensorum",
      "status": "OK",
      "failed_stage": null,
      "f1": 0.08391608391608392,
      "precision": 0.9230769230769231,
      "recall": 0.04395604395604396,
      "latency_ms_mean": 0.3236228,
      "latency_ms_total": 161.8114
    },
    {
      "rank": null,
      "team_id": "team_052",
      "team_alias": "Kare+",
      "status": "FAILED",
      "failed_stage": "configure",
      "f1": null,
      "precision": null,
      "recall": null,
      "latency_ms_mean": null,
      "latency_ms_total": null
    },
    {
      "rank": null,
      "team_id": "team_074",
      "team_alias": "TSU Montreal",
      "status": "FAILED",
      "failed_stage": "configure",
      "f1": null,
      "precision": null,
      "recall": null,
      "latency_ms_mean": null,
      "latency_ms_total": null
    }
  ]
}
