Depth Map Estimation of Focus Objects Using Vision Transformer
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | 조석제 | - |
dc.contributor.author | 박채림 | - |
dc.date.accessioned | 2024-01-03T17:28:34Z | - |
dc.date.available | 2024-01-03T17:28:35Z | - |
dc.date.created | 2023-03-03 | - |
dc.date.issued | 2023 | - |
dc.identifier.uri | http://repository.kmou.ac.kr/handle/2014.oak/13113 | - |
dc.identifier.uri | http://kmou.dcollection.net/common/orgView/200000670220 | - |
dc.description.abstract | Techniques for estimating depth map from monocular images have long been studied. The depth maps are important to understand geometric relationships within the scene which can be used for object detection, 3D modeling, augmented reality and can potentially be inferred when obscuration occurs between objects. The estimation of the depth of an object is a key part of the field of computer vision and it is essential for numerous applications. Research on searching for image-level information and hierarchical characteristics has steadily been conducted using deep-learning. However, these methods have limitations in measuring depth and detecting forward objects at night and in shadowed environments. In this paper, we propose a new method to overcome these limitations. The proposed method uses Vision Transformer (ViT) to automatically focus objects in images and measure depth maps through three different new modules: First, as Reconstitution module, the representation of the image is reconstructed, and Fusion module fuses and upsamples represented it for more detailed prediction. This can reduce the loss generated in the process of generating the depth map. In addition, it was confirmed through experiments that a cleaner and more accurate depth map was created by fine-tuning it by patch unit. This can be used in various environments and it has shown excellent results through quantitative and qualitative evaluation. | - |
dc.description.tableofcontents | 1. Introduction 1 2. Related works 3 1) RiDAR 3 2) Stereo matching algorithm 3 3) Vision Transformer 5 3. Estimating depth map of focus object 7 1) Image correction 9 1.1) Retinex 9 1.2) Image feature extraction 10 1.3) Improver 12 2) Object depth estimation 13 2.1) Generating a depth map 13 2.2) SSIM loss 16 4. Experiment and Discussion 18 1) Correction image 19 2) Depth map 21 3) Quantitative analysis 32 5. Conclusion 34 References 35 | - |
dc.language | eng | - |
dc.publisher | 한국해양대학교 대학원 | - |
dc.rights | 한국해양대학교 논문은 저작권에 의해 보호받습니다. | - |
dc.title | Depth Map Estimation of Focus Objects Using Vision Transformer | - |
dc.title.alternative | 비전 트랜스포머를 사용한 관심 물체 깊이 맵 측정 방법 | - |
dc.type | Dissertation | - |
dc.date.awarded | 2023-02 | - |
dc.embargo.terms | 2023-03-03 | - |
dc.contributor.alternativeName | Park Chae Rim | - |
dc.contributor.department | 대학원 제어계측공학과 | - |
dc.contributor.affiliation | 한국해양대학교 대학원 제어계측공학과 | - |
dc.description.degree | Master | - |
dc.identifier.bibliographicCitation | 박채림. (2023). Depth Map Estimation of Focus Objects Using Vision Transformer. | - |
dc.subject.keyword | Computer vision, Object detection, Vision Transformer, Attention, Depth map | - |
dc.identifier.holdings | 000000001979▲200000003272▲200000670220▲ | - |
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.