한국해양대학교

Detailed Information

Metadata Downloads

Depth Map Estimation of Focus Objects Using Vision Transformer

DC Field Value Language
dc.contributor.advisor 조석제 -
dc.contributor.author 박채림 -
dc.date.accessioned 2024-01-03T17:28:34Z -
dc.date.available 2024-01-03T17:28:35Z -
dc.date.created 2023-03-03 -
dc.date.issued 2023 -
dc.identifier.uri http://repository.kmou.ac.kr/handle/2014.oak/13113 -
dc.identifier.uri http://kmou.dcollection.net/common/orgView/200000670220 -
dc.description.abstract Techniques for estimating depth map from monocular images have long been studied. The depth maps are important to understand geometric relationships within the scene which can be used for object detection, 3D modeling, augmented reality and can potentially be inferred when obscuration occurs between objects. The estimation of the depth of an object is a key part of the field of computer vision and it is essential for numerous applications. Research on searching for image-level information and hierarchical characteristics has steadily been conducted using deep-learning. However, these methods have limitations in measuring depth and detecting forward objects at night and in shadowed environments. In this paper, we propose a new method to overcome these limitations. The proposed method uses Vision Transformer (ViT) to automatically focus objects in images and measure depth maps through three different new modules: First, as Reconstitution module, the representation of the image is reconstructed, and Fusion module fuses and upsamples represented it for more detailed prediction. This can reduce the loss generated in the process of generating the depth map. In addition, it was confirmed through experiments that a cleaner and more accurate depth map was created by fine-tuning it by patch unit. This can be used in various environments and it has shown excellent results through quantitative and qualitative evaluation. -
dc.description.tableofcontents 1. Introduction 1 2. Related works 3 1) RiDAR 3 2) Stereo matching algorithm 3 3) Vision Transformer 5 3. Estimating depth map of focus object 7 1) Image correction 9 1.1) Retinex 9 1.2) Image feature extraction 10 1.3) Improver 12 2) Object depth estimation 13 2.1) Generating a depth map 13 2.2) SSIM loss 16 4. Experiment and Discussion 18 1) Correction image 19 2) Depth map 21 3) Quantitative analysis 32 5. Conclusion 34 References 35 -
dc.language eng -
dc.publisher 한국해양대학교 대학원 -
dc.rights 한국해양대학교 논문은 저작권에 의해 보호받습니다. -
dc.title Depth Map Estimation of Focus Objects Using Vision Transformer -
dc.title.alternative 비전 트랜스포머를 사용한 관심 물체 깊이 맵 측정 방법 -
dc.type Dissertation -
dc.date.awarded 2023-02 -
dc.embargo.terms 2023-03-03 -
dc.contributor.alternativeName Park Chae Rim -
dc.contributor.department 대학원 제어계측공학과 -
dc.contributor.affiliation 한국해양대학교 대학원 제어계측공학과 -
dc.description.degree Master -
dc.identifier.bibliographicCitation 박채림. (2023). Depth Map Estimation of Focus Objects Using Vision Transformer. -
dc.subject.keyword Computer vision, Object detection, Vision Transformer, Attention, Depth map -
dc.identifier.holdings 000000001979▲200000003272▲200000670220▲ -
Appears in Collections:
기타 > 기타
Files in This Item:
There are no files associated with this item.

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.

Browse