Browsing by Subject "Generative adversarial networks"
Now showing 1 - 2 of 2
- Results Per Page
- Sort Options
Item Generative Deep Learning Methods for Improving Few-Shot Segmentation of Infrared Images(2023-07) Yun, JunnoImage semantic segmentation is an essential topic in computer vision. However, most current deep learning-based networks primarily focus on large-scale visible spectrum datasets, specifically RGB images. Consequently, these models often struggle to perform well on segmentation in adverse environmental conditions such as total darkness, poor illumination, and smog, etc. On the other hand, a thermal infrared (IR) imaging system overcomes the limitations of visible spectrum imaging. Leveraging the advantages of thermal IR imaging, semantic segmentation using infrared images has extensive real-world applications, including defense applications, autonomous driving, and medical imaging. However, there remain challenges in harnessing the benefits of IR images. Firstly, accessing large-scale annotated IR images is challenging due to security considerations. Secondly, we can face data with unseen classes or rare categories in various IR imaging applications. Lastly, thermal IR images typically exhibit low-resolution, low-contrast, and obscure object boundaries due to the nature of thermal cameras. These images are composed of a single grayscale channel, which provides limited information compared to RGB images. Thus, applying only IR images to existing semantic segmentation models leads to inaccurate performance. In this thesis, our aim is to improve existing few-shot segmentation models to enable robust few-shot segmentation of IR images. To this end, we propose the use of generative methods to enhance our few-shot segmentation model by generating two types of synthesized data: one for augmenting the training data and the other for providing conditioning information. Results show that these strategies substantially improve few-shot segmentation of IR images.Item Monocular Depth Estimation using Adversarial Training(2020-07) Mitra, PallaviMonocular depth estimation is a fundamentally challenging problem in Computer Vision. It is useful for Robotics applications where design constraints prohibit the use of multiple cameras. It also finds widespread use in autonomous driving. Since the task is to estimate depth from a single image, rather than two or more, a global perspective of the scene is required. Pixel-wise losses like reconstruction loss, left-right consistency loss, capture local scene information. However, they do not take into account global scene consistency. Generative Adversarial Networks(GANs) effectively capture the global structure of the scene and produce real-looking images, so they have the potential of depth estimation from a single image. This work focuses on using adversarial training for a supervised monocular depth estimation task in combination with pixel-wise losses. We observe that with minimal depth-supervised training, there is a significant reduction of error in depth estimation in a number of GAN variants explored.