Auto-fit: A human-machine collaboration feature for fitting bounding box annotations

Large high-quality annotated datasets are essential in training deep learning models, but are expensive and time-consuming to create. A large chunk of time in the annotation process goes into adjusting bounding boxes to fit the desired object. In this paper, we propose the facilitation of human mach...

Full description

Saved in:
Bibliographic Details
Main Authors: Cruz, Meygen, Keh, Jefferson, Velasco, Neil Oliver M., Jose, John Anthony, Sybingco, Edwin, Dadios, Elmer P., Madria, Wira F., Miguel, Angelimarie
Format: text
Published: Animo Repository 2020
Subjects:
Online Access:https://animorepository.dlsu.edu.ph/faculty_research/12591
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: De La Salle University
id oai:animorepository.dlsu.edu.ph:faculty_research-14588
record_format eprints
spelling oai:animorepository.dlsu.edu.ph:faculty_research-145882024-07-01T01:49:31Z Auto-fit: A human-machine collaboration feature for fitting bounding box annotations Cruz, Meygen Keh, Jefferson Velasco, Neil Oliver M. Jose, John Anthony Sybingco, Edwin Dadios, Elmer P. Madria, Wira F. Miguel, Angelimarie Large high-quality annotated datasets are essential in training deep learning models, but are expensive and time-consuming to create. A large chunk of time in the annotation process goes into adjusting bounding boxes to fit the desired object. In this paper, we propose the facilitation of human machine collaboration through the creation of an Auto- Fit feature which automatically tightens an initial bounding box around an object being annotated. The challenge lies in making this feature class agnostic in order to allow its usage regardless of the type of object being annotated. This is achieved through the use of various computer vision algorithms to extract the desired object as a foreground mask, determine the coordinates of its extremities, and redraw the bounding box based on these new coordinates. The best results were achieved with the Grabcut algorithm, which attained an accuracy of 84.69% on small boxes. The Pytorch implementation of ResNet-101 pre-trained on the COCO train2017 dataset is also used as a foreground extractor in one iteration of the implementation, in order to provide a baseline comparison between the performance of a computer vision- based solution versus one based on a standalone object detection model. This garnered an accuracy of 83.04% on small boxes, showing that the computer vision-based solution is able to surpass the accuracy of a standalone object detection model. 2020-01-01T08:00:00Z text https://animorepository.dlsu.edu.ph/faculty_research/12591 Faculty Research Work Animo Repository Computer vision Computer Engineering
institution De La Salle University
building De La Salle University Library
continent Asia
country Philippines
Philippines
content_provider De La Salle University Library
collection DLSU Institutional Repository
topic Computer vision
Computer Engineering
spellingShingle Computer vision
Computer Engineering
Cruz, Meygen
Keh, Jefferson
Velasco, Neil Oliver M.
Jose, John Anthony
Sybingco, Edwin
Dadios, Elmer P.
Madria, Wira F.
Miguel, Angelimarie
Auto-fit: A human-machine collaboration feature for fitting bounding box annotations
description Large high-quality annotated datasets are essential in training deep learning models, but are expensive and time-consuming to create. A large chunk of time in the annotation process goes into adjusting bounding boxes to fit the desired object. In this paper, we propose the facilitation of human machine collaboration through the creation of an Auto- Fit feature which automatically tightens an initial bounding box around an object being annotated. The challenge lies in making this feature class agnostic in order to allow its usage regardless of the type of object being annotated. This is achieved through the use of various computer vision algorithms to extract the desired object as a foreground mask, determine the coordinates of its extremities, and redraw the bounding box based on these new coordinates. The best results were achieved with the Grabcut algorithm, which attained an accuracy of 84.69% on small boxes. The Pytorch implementation of ResNet-101 pre-trained on the COCO train2017 dataset is also used as a foreground extractor in one iteration of the implementation, in order to provide a baseline comparison between the performance of a computer vision- based solution versus one based on a standalone object detection model. This garnered an accuracy of 83.04% on small boxes, showing that the computer vision-based solution is able to surpass the accuracy of a standalone object detection model.
format text
author Cruz, Meygen
Keh, Jefferson
Velasco, Neil Oliver M.
Jose, John Anthony
Sybingco, Edwin
Dadios, Elmer P.
Madria, Wira F.
Miguel, Angelimarie
author_facet Cruz, Meygen
Keh, Jefferson
Velasco, Neil Oliver M.
Jose, John Anthony
Sybingco, Edwin
Dadios, Elmer P.
Madria, Wira F.
Miguel, Angelimarie
author_sort Cruz, Meygen
title Auto-fit: A human-machine collaboration feature for fitting bounding box annotations
title_short Auto-fit: A human-machine collaboration feature for fitting bounding box annotations
title_full Auto-fit: A human-machine collaboration feature for fitting bounding box annotations
title_fullStr Auto-fit: A human-machine collaboration feature for fitting bounding box annotations
title_full_unstemmed Auto-fit: A human-machine collaboration feature for fitting bounding box annotations
title_sort auto-fit: a human-machine collaboration feature for fitting bounding box annotations
publisher Animo Repository
publishDate 2020
url https://animorepository.dlsu.edu.ph/faculty_research/12591
_version_ 1806061243399667712