{
  "nms_scores_th": 0.01,            // Filtering phase: Score threshold. All the bounding boxes with lower confidence score will be filtered
  "nms_iou_th": 1,                  // IoU Threshold. Use 1 here
  "image_dims": [640, 640],         // Input image size
  "max_proposals_per_class": 80,    // Maximum number of proposals per class
  "input_division_factor": 5,       // Post-processing is performed on 1/input_division_factor of the bboxes at a time (if >1, on-chip IoU can't be accurate)
  "classes": 80,                    // Overall number of classes (including background)
  "background_removal": false,      // Whether to remove the background channel
  "background_removal_index": 0,    // The index of the background channel
  "bbox_decoders": [                // List of bbox decoders for the NMS layer. Each model has its own number of boxes per anchor
      {
          "name": "bbox_decoder73", // Name of BBOX decoder to use (use any name you want)
          "w": [                    // List of Width coordinates (given as fraction of input size), defining each box dimensions around the anchor coordinates. CONSTANTS THAT WERE USED FOR TRAINING
              10,                   // Vertical size of first anchor out of three, going to first bbox decoder
              16,                   // Vertical size of second anchor out of three, going to first bbox decoder
              33                    // Vertical size of third anchor out of three, going to first bbox decoder
          ],
          "h": [                    // List of Height coordinates (given as fraction of input size), defining each box dimensions around the anchor coordinates. CONSTANTS THAT WERE USED FOR TRAINING
              13,                   // Horizontal size of first anchor out of three, going to first bbox decoder
              30,                   // Horizontal size of second anchor out of three, going to first bbox decoder
              23                    // Horizontal size of third anchor out of three, going to first bbox decoder
          ],
          "stride": 8,              // Pixels stride for given bbox. CONSTANT THAT IS USED FOR TRAINING
          "encoded_layer": "conv73" // Corresponding bbox encoding predictor layer from translated .hn file (inside the .har, or using ``hailo visualizer`` tool)
      },
      {
          "name": "bbox_decoder83",
          "w": [
              30,
              62,
              59
          ],
          "h": [
              61,
              45,
              119
          ],
          "stride": 16,
          "encoded_layer": "conv83"
      },
      {
          "name": "bbox_decoder93",
          "w": [
              116,
              156,
              373 
          ],
          "h": [
              90,
              198,
              326
          ],
          "stride": 32,
          "encoded_layer": "conv93"
      }
  ]
}
