Auto-fit: A human-machine collaboration feature for fitting bounding box annotations
Large high-quality annotated datasets are essential in training deep learning models, but are expensive and time-consuming to create. A large chunk of time in the annotation process goes into adjusting bounding boxes to fit the desired object. In this paper, we propose the facilitation of human mach...
Saved in:
Main Authors: | , , , , , , , |
---|---|
Format: | text |
Published: |
Animo Repository
2020
|
Subjects: | |
Online Access: | https://animorepository.dlsu.edu.ph/faculty_research/12591 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | De La Salle University |
id |
oai:animorepository.dlsu.edu.ph:faculty_research-14588 |
---|---|
record_format |
eprints |
spelling |
oai:animorepository.dlsu.edu.ph:faculty_research-145882024-07-01T01:49:31Z Auto-fit: A human-machine collaboration feature for fitting bounding box annotations Cruz, Meygen Keh, Jefferson Velasco, Neil Oliver M. Jose, John Anthony Sybingco, Edwin Dadios, Elmer P. Madria, Wira F. Miguel, Angelimarie Large high-quality annotated datasets are essential in training deep learning models, but are expensive and time-consuming to create. A large chunk of time in the annotation process goes into adjusting bounding boxes to fit the desired object. In this paper, we propose the facilitation of human machine collaboration through the creation of an Auto- Fit feature which automatically tightens an initial bounding box around an object being annotated. The challenge lies in making this feature class agnostic in order to allow its usage regardless of the type of object being annotated. This is achieved through the use of various computer vision algorithms to extract the desired object as a foreground mask, determine the coordinates of its extremities, and redraw the bounding box based on these new coordinates. The best results were achieved with the Grabcut algorithm, which attained an accuracy of 84.69% on small boxes. The Pytorch implementation of ResNet-101 pre-trained on the COCO train2017 dataset is also used as a foreground extractor in one iteration of the implementation, in order to provide a baseline comparison between the performance of a computer vision- based solution versus one based on a standalone object detection model. This garnered an accuracy of 83.04% on small boxes, showing that the computer vision-based solution is able to surpass the accuracy of a standalone object detection model. 2020-01-01T08:00:00Z text https://animorepository.dlsu.edu.ph/faculty_research/12591 Faculty Research Work Animo Repository Computer vision Computer Engineering |
institution |
De La Salle University |
building |
De La Salle University Library |
continent |
Asia |
country |
Philippines Philippines |
content_provider |
De La Salle University Library |
collection |
DLSU Institutional Repository |
topic |
Computer vision Computer Engineering |
spellingShingle |
Computer vision Computer Engineering Cruz, Meygen Keh, Jefferson Velasco, Neil Oliver M. Jose, John Anthony Sybingco, Edwin Dadios, Elmer P. Madria, Wira F. Miguel, Angelimarie Auto-fit: A human-machine collaboration feature for fitting bounding box annotations |
description |
Large high-quality annotated datasets are essential in training deep learning models, but are expensive and time-consuming to create. A large chunk of time in the annotation process goes into adjusting bounding boxes to fit the desired object. In this paper, we propose the facilitation of human machine collaboration through the creation of an Auto- Fit feature which automatically tightens an initial bounding box around an object being annotated. The challenge lies in making this feature class agnostic in order to allow its usage regardless of the type of object being annotated. This is achieved through the use of various computer vision algorithms to extract the desired object as a foreground mask, determine the coordinates of its extremities, and redraw the bounding box based on these new coordinates. The best results were achieved with the Grabcut algorithm, which attained an accuracy of 84.69% on small boxes. The Pytorch implementation of ResNet-101 pre-trained on the COCO train2017 dataset is also used as a foreground extractor in one iteration of the implementation, in order to provide a baseline comparison between the performance of a computer vision- based solution versus one based on a standalone object detection model. This garnered an accuracy of 83.04% on small boxes, showing that the computer vision-based solution is able to surpass the accuracy of a standalone object detection model. |
format |
text |
author |
Cruz, Meygen Keh, Jefferson Velasco, Neil Oliver M. Jose, John Anthony Sybingco, Edwin Dadios, Elmer P. Madria, Wira F. Miguel, Angelimarie |
author_facet |
Cruz, Meygen Keh, Jefferson Velasco, Neil Oliver M. Jose, John Anthony Sybingco, Edwin Dadios, Elmer P. Madria, Wira F. Miguel, Angelimarie |
author_sort |
Cruz, Meygen |
title |
Auto-fit: A human-machine collaboration feature for fitting bounding box annotations |
title_short |
Auto-fit: A human-machine collaboration feature for fitting bounding box annotations |
title_full |
Auto-fit: A human-machine collaboration feature for fitting bounding box annotations |
title_fullStr |
Auto-fit: A human-machine collaboration feature for fitting bounding box annotations |
title_full_unstemmed |
Auto-fit: A human-machine collaboration feature for fitting bounding box annotations |
title_sort |
auto-fit: a human-machine collaboration feature for fitting bounding box annotations |
publisher |
Animo Repository |
publishDate |
2020 |
url |
https://animorepository.dlsu.edu.ph/faculty_research/12591 |
_version_ |
1806061243399667712 |