Table Of Contents
Table Of Contents

Source code for gluoncv.data.mscoco.detection

"""MS COCO object detection dataset."""
from __future__ import absolute_import
from __future__ import division

import os

import mxnet as mx
import numpy as np
from PIL import Image

from .utils import try_import_pycocotools
from ..base import VisionDataset
from ...utils import try_import_dali
from ...utils.bbox import bbox_xywh_to_xyxy, bbox_clip_xyxy

dali = try_import_dali()

__all__ = ['COCODetection', 'COCODetectionDALI']


[docs]class COCODetection(VisionDataset): """MS COCO detection dataset. Parameters ---------- root : str, default '~/.mxnet/datasets/coco' Path to folder storing the dataset. splits : list of str, default ['instances_val2017'] Json annotations name. Candidates can be: instances_val2017, instances_train2017. transform : callable, default None A function that takes data and label and transforms them. Refer to :doc:`./transforms` for examples. A transform function for object detection should take label into consideration, because any geometric modification will require label to be modified. min_object_area : float Minimum accepted ground-truth area, if an object's area is smaller than this value, it will be ignored. skip_empty : bool, default is True Whether skip images with no valid object. This should be `True` in training, otherwise it will cause undefined behavior. use_crowd : bool, default is True Whether use boxes labeled as crowd instance. """ CLASSES = ['person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'traffic light', 'fire hydrant', 'stop sign', 'parking meter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sports ball', 'kite', 'baseball bat', 'baseball glove', 'skateboard', 'surfboard', 'tennis racket', 'bottle', 'wine glass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hot dog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'potted plant', 'bed', 'dining table', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cell phone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddy bear', 'hair drier', 'toothbrush'] def __init__(self, root=os.path.join('~', '.mxnet', 'datasets', 'coco'), splits=('instances_val2017',), transform=None, min_object_area=0, skip_empty=True, use_crowd=True): super(COCODetection, self).__init__(root) self._root = os.path.expanduser(root) self._transform = transform self._min_object_area = min_object_area self._skip_empty = skip_empty self._use_crowd = use_crowd if isinstance(splits, mx.base.string_types): splits = [splits] self._splits = splits # to avoid trouble, we always use contiguous IDs except dealing with cocoapi self.index_map = dict(zip(type(self).CLASSES, range(self.num_class))) self.json_id_to_contiguous = None self.contiguous_id_to_json = None self._coco = [] self._items, self._labels, self._im_aspect_ratios = self._load_jsons() def __str__(self): detail = ','.join([str(s) for s in self._splits]) return self.__class__.__name__ + '(' + detail + ')' @property def coco(self): """Return pycocotools object for evaluation purposes.""" if not self._coco: raise ValueError("No coco objects found, dataset not initialized.") if len(self._coco) > 1: raise NotImplementedError( "Currently we don't support evaluating {} JSON files. \ Please use single JSON dataset and evaluate one by one".format(len(self._coco))) return self._coco[0] @property def classes(self): """Category names.""" return type(self).CLASSES @property def annotation_dir(self): """ The subdir for annotations. Default is 'annotations'(coco default) For example, a coco format json file will be searched as 'root/annotation_dir/xxx.json' You can override if custom dataset don't follow the same pattern """ return 'annotations' def get_im_aspect_ratio(self): """Return the aspect ratio of each image in the order of the raw data.""" if self._im_aspect_ratios is not None: return self._im_aspect_ratios self._im_aspect_ratios = [None] * len(self._items) for i, img_path in enumerate(self._items): with Image.open(img_path) as im: w, h = im.size self._im_aspect_ratios[i] = 1.0 * w / h return self._im_aspect_ratios def _parse_image_path(self, entry): """How to parse image dir and path from entry. Parameters ---------- entry : dict COCO entry, e.g. including width, height, image path, etc.. Returns ------- abs_path : str Absolute path for corresponding image. """ dirname, filename = entry['coco_url'].split('/')[-2:] abs_path = os.path.join(self._root, dirname, filename) return abs_path def __len__(self): return len(self._items) def __getitem__(self, idx): img_path = self._items[idx] label = self._labels[idx] img = mx.image.imread(img_path, 1) if self._transform is not None: return self._transform(img, label) return img, np.array(label).copy() def _load_jsons(self): """Load all image paths and labels from JSON annotation files into buffer.""" items = [] labels = [] im_aspect_ratios = [] # lazy import pycocotools try_import_pycocotools() from pycocotools.coco import COCO for split in self._splits: anno = os.path.join(self._root, self.annotation_dir, split) + '.json' _coco = COCO(anno) self._coco.append(_coco) classes = [c['name'] for c in _coco.loadCats(_coco.getCatIds())] if not classes == self.classes: raise ValueError("Incompatible category names with COCO: ") assert classes == self.classes json_id_to_contiguous = { v: k for k, v in enumerate(_coco.getCatIds())} if self.json_id_to_contiguous is None: self.json_id_to_contiguous = json_id_to_contiguous self.contiguous_id_to_json = { v: k for k, v in self.json_id_to_contiguous.items()} else: assert self.json_id_to_contiguous == json_id_to_contiguous # iterate through the annotations image_ids = sorted(_coco.getImgIds()) for entry in _coco.loadImgs(image_ids): abs_path = self._parse_image_path(entry) if not os.path.exists(abs_path): raise IOError('Image: {} not exists.'.format(abs_path)) label = self._check_load_bbox(_coco, entry) if not label: continue im_aspect_ratios.append(float(entry['width']) / entry['height']) items.append(abs_path) labels.append(label) return items, labels, im_aspect_ratios def _check_load_bbox(self, coco, entry): """Check and load ground-truth labels""" entry_id = entry['id'] # fix pycocotools _isArrayLike which don't work for str in python3 entry_id = [entry_id] if not isinstance(entry_id, (list, tuple)) else entry_id ann_ids = coco.getAnnIds(imgIds=entry_id, iscrowd=None) objs = coco.loadAnns(ann_ids) # check valid bboxes valid_objs = [] width = entry['width'] height = entry['height'] for obj in objs: if obj['area'] < self._min_object_area: continue if obj.get('ignore', 0) == 1: continue if not self._use_crowd and obj.get('iscrowd', 0): continue # convert from (x, y, w, h) to (xmin, ymin, xmax, ymax) and clip bound xmin, ymin, xmax, ymax = bbox_clip_xyxy(bbox_xywh_to_xyxy(obj['bbox']), width, height) # require non-zero box area if obj['area'] > 0 and xmax > xmin and ymax > ymin: contiguous_cid = self.json_id_to_contiguous[obj['category_id']] valid_objs.append([xmin, ymin, xmax, ymax, contiguous_cid]) if not valid_objs: if not self._skip_empty: # dummy invalid labels if no valid objects are found valid_objs.append([-1, -1, -1, -1, -1]) return valid_objs
class COCODetectionDALI(object): """DALI partial pipeline with COCO Reader and loader. To be passed as a parameter of a DALI transform pipeline. Parameters ---------- num_shards: int DALI pipeline arg - Number of pipelines used, indicating to the reader how to split/shard the dataset. shard_id: int DALI pipeline arg - Shard id of the pipeline must be in [0, num_shards). file_root Directory containing the COCO dataset. annotations_file The COCO annotation file to read from. device_id: int GPU device used for the DALI pipeline. """ def __init__(self, num_shards, shard_id, file_root, annotations_file, device_id): self.input = dali.ops.COCOReader( file_root=file_root, annotations_file=annotations_file, skip_empty=True, shard_id=shard_id, num_shards=num_shards, ratio=True, ltrb=True, shuffle_after_epoch=True) self.decode = dali.ops.HostDecoder(device="cpu", output_type=dali.types.RGB) # We need to build the COCOReader ops to parse the annotations # and have acces to the dataset size. # TODO(spanev): Replace by DALI standalone ops when available class DummyMicroPipe(dali.Pipeline): """ Dummy pipeline which sole purpose is to build COCOReader and get the epoch size. To be replaced by DALI standalone op, when available. """ def __init__(self, device_id): super(DummyMicroPipe, self).__init__(batch_size=1, device_id=device_id, num_threads=1) self.input = dali.ops.COCOReader( file_root=file_root, annotations_file=annotations_file) def define_graph(self): inputs, bboxes, labels = self.input(name="Reader") return (inputs, bboxes, labels) micro_pipe = DummyMicroPipe(device_id=device_id) micro_pipe.build() self._size = micro_pipe.epoch_size(name="Reader") del micro_pipe def __call__(self): """Returns three DALI graph nodes: inputs, bboxes, labels. To be called in `define_graph`. """ inputs, bboxes, labels = self.input(name="Reader") images = self.decode(inputs) return (images, bboxes, labels) def size(self): """Returns size of COCO dataset """ return self._size