한국해양대학교

Detailed Information

Metadata Downloads

Depth Map Estimation of Focus Objects Using Vision Transformer

Title
Depth Map Estimation of Focus Objects Using Vision Transformer
Alternative Title
비전 트랜스포머를 사용한 관심 물체 깊이 맵 측정 방법
Author(s)
박채림
Keyword
Computer vision, Object detection, Vision Transformer, Attention, Depth map
Issued Date
2023
Publisher
한국해양대학교 대학원
URI
http://repository.kmou.ac.kr/handle/2014.oak/13113
http://kmou.dcollection.net/common/orgView/200000670220
Abstract
Techniques for estimating depth map from monocular images have long been studied. The depth maps are important to understand geometric relationships within the scene which can be used for object detection, 3D modeling, augmented reality and can potentially be inferred when obscuration occurs between objects. The estimation of the depth of an object is a key part of the field of computer vision and it is essential for numerous applications. Research on searching for image-level information and hierarchical characteristics has steadily been conducted using deep-learning. However, these methods have limitations in measuring depth and detecting forward objects at night and in shadowed environments.
In this paper, we propose a new method to overcome these limitations. The proposed method uses Vision Transformer (ViT) to automatically focus objects in images and measure depth maps through three different new modules: First, as Reconstitution module, the representation of the image is reconstructed, and Fusion module fuses and upsamples represented it for more detailed prediction. This can reduce the loss generated in the process of generating the depth map. In addition, it was confirmed through experiments that a cleaner and more accurate depth map was created by fine-tuning it by patch unit. This can be used in various environments and it has shown excellent results through quantitative and qualitative evaluation.
Appears in Collections:
기타 > 기타
Files in This Item:
There are no files associated with this item.

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.

Browse