MIC MM OBJECT DETECTION CHALLENGE 2022
MIC Object Detection Challenge
We offer a challenge, where models are to be trained in a data-deficient setting. The challenge is based on the MIC21 dataset. The main objective of the challenge is to detect and classify particular objects that are not usually included in CV tasks: cricketer [621 training instances], soccer player [ 571 training instances], chess player [733 training instances], van [307 training instances], sedan [312 training instances], police car [548 training instances], policeman [687 training instances], soldier [915 training instances], dancer [405 training instances], violinist [410 training instances].
Participants need to provide the output produced by their systems on the test data in the COCO JSON format over the target classes. They can use any of the available frameworks, such as YOLACT and Detectron2, using the COCO dataset as a benchmark dataset.
The MIC 21 training and validation dataset is available here. The test dataset will be released on 1st August 2022.
Using any of the available COCO-based models (which detect objects such as person, car, truck), the participants should provide annotations for objects representing cricketer, soccer player, chess player, van, sedan, police car, policeman, soldier, dancer, violinist. The training and evaluation datasets also contain annotations for objects associated with the objects selected for the task.
The output should be in the COCO JSON format and should contain information about images, categories and annotations (segmentations and bboxes). The evaluation and testing should be provided in the COCO testing framework.
The participants are strongly encouraged to share their code (for example in GitHub) for the purposes of a reproducibility study.
We allow the use of data other than the provided training data; all training data used, as well as the approach adopted by the participants, should be thoroughly documented in the Technical report.
We require the participants in the MIC MM Challenge to submit a technical report. Submissions of results without an associated technical report do not qualify for the competition.
The Organisers retain the right to disqualify any submissions that violate these rules.
The Organisers retain the right to extend the deadlines.
The technical reports submitted to the MIC MM Challenge should be between 3 and 6 pages long + up to 2 pages for references. The template for the submission is the CLIB 2022 template.
Upon acceptance, the technical reports will be published in the Proceedings of the International Conference Computational Linguistics in Bulgaria (CLIB 2022).
For more information about the dataset, you can check the following publication:
Koeva, S., I. Stoyanova, Y. Kralev. Multilingual Image Corpus – Towards a Multimodal and Multilingual Dataset. In: Proceedings of LREC 2022, pp. 1509-1518. (pdf)
Registration and submission
Preliminary registration for participation in the Challenge (by 30 June 2022), submission of output results (by 15 August 2022) and technical reports (by 20 August 2022) takes place via EasyChair.
Shared task announcement: 1 June 2022
Release of training and validation dataset: 1 June 2022
Registration deadline: 30 June 2022
Upload of outputs of training and validation: 31 July 2022
Release of test dataset: 1 August 2022
Shared task test output submission: 15 August 2022
Upload of technical reports: 20 August 2022
Notification of acceptance: 25 August 2022
Camera-ready of shared task technical reports: 5 September 2022
Notification of the results of the competition: 9 September 2022
Jose Manuel Gomez Perez
If you have any questions, please email the Organisers at email@example.com.