Birds eye view look-up table estimation with semantic segmentation

In this work, a study was carried out to estimate a look-up table (LUT) that converts a camera image plane to a birds eye view (BEV) plane using a single camera. The traditional camera pose estimation fields require high costs in researching and manufacturing autonomous vehicles for the future and m...

Full description

Saved in:
Bibliographic Details
Main Authors: Lee, Dongkyu, Tay, Wee Peng, Kee, Seok-Cheol
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/153942
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-153942
record_format dspace
spelling sg-ntu-dr.10356-1539422022-01-10T08:42:59Z Birds eye view look-up table estimation with semantic segmentation Lee, Dongkyu Tay, Wee Peng Kee, Seok-Cheol School of Electrical and Electronic Engineering Engineering::Computer science and engineering Birds Eye View Look-Up Table In this work, a study was carried out to estimate a look-up table (LUT) that converts a camera image plane to a birds eye view (BEV) plane using a single camera. The traditional camera pose estimation fields require high costs in researching and manufacturing autonomous vehicles for the future and may require pre-configured infra. This paper proposes an autonomous vehicle driving camera calibration system that is low cost and utilizes low infra. A network that outputs an image in the form of an LUT that converts the image into a BEV by estimating the camera pose under urban road driving conditions using a single camera was studied. We propose a network that predicts human-like poses from a single image. We collected synthetic data using a simulator, made BEV and LUT as ground truth, and utilized the proposed network and ground truth to train pose estimation function. In the progress, it predicts the pose by deciphering the semantic segmentation feature and increases its performance by attaching a layer that handles the overall direction of the network. The network outputs camera angle (roll/pitch/yaw) on the 3D coordinate system so that the user can monitor learning. Since the network's output is a LUT, there is no need for additional calculation, and real-time performance is improved. Published version This research was supported by the MOTIE (Ministry of Trade, Industry, and Energy) in Korea, under the Fostering Global Talents for Innovative Growth Program (P0008751) supervised by the Korea Institute for Advancement of Technology (KIAT). This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the Grand Information Technology Research Center support program (IITP-2021-2020-0-01462) supervised by the IITP (Institute for Information & communications Technology Planning & Evaluation. 2022-01-10T08:42:59Z 2022-01-10T08:42:59Z 2021 Journal Article Lee, D., Tay, W. P. & Kee, S. (2021). Birds eye view look-up table estimation with semantic segmentation. Applied Sciences, 11(17), 8047-. https://dx.doi.org/10.3390/app11178047 2076-3417 https://hdl.handle.net/10356/153942 10.3390/app11178047 2-s2.0-85114263699 17 11 8047 en Applied Sciences © 2021 The Author(s). Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Birds Eye View
Look-Up Table
spellingShingle Engineering::Computer science and engineering
Birds Eye View
Look-Up Table
Lee, Dongkyu
Tay, Wee Peng
Kee, Seok-Cheol
Birds eye view look-up table estimation with semantic segmentation
description In this work, a study was carried out to estimate a look-up table (LUT) that converts a camera image plane to a birds eye view (BEV) plane using a single camera. The traditional camera pose estimation fields require high costs in researching and manufacturing autonomous vehicles for the future and may require pre-configured infra. This paper proposes an autonomous vehicle driving camera calibration system that is low cost and utilizes low infra. A network that outputs an image in the form of an LUT that converts the image into a BEV by estimating the camera pose under urban road driving conditions using a single camera was studied. We propose a network that predicts human-like poses from a single image. We collected synthetic data using a simulator, made BEV and LUT as ground truth, and utilized the proposed network and ground truth to train pose estimation function. In the progress, it predicts the pose by deciphering the semantic segmentation feature and increases its performance by attaching a layer that handles the overall direction of the network. The network outputs camera angle (roll/pitch/yaw) on the 3D coordinate system so that the user can monitor learning. Since the network's output is a LUT, there is no need for additional calculation, and real-time performance is improved.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Lee, Dongkyu
Tay, Wee Peng
Kee, Seok-Cheol
format Article
author Lee, Dongkyu
Tay, Wee Peng
Kee, Seok-Cheol
author_sort Lee, Dongkyu
title Birds eye view look-up table estimation with semantic segmentation
title_short Birds eye view look-up table estimation with semantic segmentation
title_full Birds eye view look-up table estimation with semantic segmentation
title_fullStr Birds eye view look-up table estimation with semantic segmentation
title_full_unstemmed Birds eye view look-up table estimation with semantic segmentation
title_sort birds eye view look-up table estimation with semantic segmentation
publishDate 2022
url https://hdl.handle.net/10356/153942
_version_ 1722355345277321216