How does the current state of the art perform for detecting and recognizing objects in videos having different scales, illumination, noises, weather condition and occlusion? Can the application of image enhancement and restoration algorithms as a pre-processing step improve image interpretability for automatic visual recognition to classify scene content? The UG2+ Track 1 aims to advance the analysis of videos collected by small UAVs by applying image restoration and enhancement algorithms to improve recognition performance using the UG2 (UAV, Glider, Ground) dataset, which has been collected specifically for this purpose.
What should a software system that can interpret images from UAVs actually look like? It must incorporate a set of algorithms, drawn from the areas of computational photography and machine learning, into a processing pipeline that correct undesired visual degradation in UAV-based image collection and subsequently classifies images across time and space. Image restoration and enhancement algorithms that remove corruptions like blur, noise and mis-focus, or manipulate images to gain resolution, change perspective and compensate for lens distortion are now commonplace in photo editing tools. Such operations are necessary to improve the quality of images for recognition purposes. But they must be compatible with the recognition process itself, and not adversely affect feature extraction or decision making. Exploratory work is needed to find out which image pre-processing algorithms, in combination with the strongest features and supervised machine learning approaches, are promising candidates for UAV-related computer vision applications.
UG2+ Challenge 1 consists of two sub-challenges:
Object Detection Improvement on Video
- 1st Place: $15K
- 2nd Place: $10K
Object Classification Improvement on Video
- 1st Place: $15K
- 2nd Place: $10K
Sub-Challenge 1.1: Object Detection Improvement on Video
The goal of this challenge is to detect objects from a number of visual object classes in unconstrained environments (i.e., not pre-segmented objects). It is fundamentally a supervised learning problem in that a training set of labeled images will be provided. Participants are not expected to develop novel object detection models. They are expected to use a pre-processing step (super-resolution, de-noising, deblurring, etc. and any combinations of these algorithms are within scope here) in the detection pipeline. A list of detection algorithms that will be used for scoring will be made available to the participants in order to facilitate studies of the interaction between image restoration and enhancement algorithms and the detectors. During the evaluation, the selected object detection algorithms will be run on the sequestered test images. In-line with popular detectors, the metrics will be mAP@0.5 (mean average precision).
Sub-Challenge 1.2: Object Classification Improvement on Video
The goal of this challenge is to provide an improvement on the classification accuracy of a given object in a video captured from an unconstrained mobility platform. Participants will be tasked with the creation of a procesing pipeline to correct visual aberrations present in video in order to improve the classification results obtained with out-of-the-box classification algorithms.T he evaluation protocol will allow participants to make use of within dataset training data, and as much outside training data as they would like for training / validation purposes. Participants will not be tasked with the creation of novel classification algorithms. A list of classification algorithms will be provided to them in advance of the competition, in order to facilitate studies of the interaction between image restoration and enhancement algorithms and the classifiers. During the evaluation, the selected classification algorithms will be run on the sequestered test images. The metrics will be the mean classification accuracy (top-5 classification) over all the frames in each video segment.