lat360.github.io

360 Image Reference-based Super-Resolution using Latitude Aware Convolution Learned from Synthetic to Real

Hee-Jae Kim, Je-Won Kang, Byung-Uk Lee

Ewha Womans University

[code]
[paper]

Abstract

High-resolution (HR) 360-degree images offer great advantages wherever an omnidirectional view is necessary such as in autonomous robot systems and virtual reality (VR) applications. One or more 360-degree images in adjacent views can be utilized to significantly improve the resolution of a target 360-degree image. In this paper, we propose an efficient reference-based 360-degree image super-resolution (RefSR) technique to exploit a wide field of view (FoV) among adjacent 360-degree cameras. Effective exploitation of spatial correlation is critical to achieving high quality even though the distortion inherent in the equi-rectangular projection (ERP) is a nontrivial problem. Accordingly, we develop a long-range 360 disparity estimator (DE360) to overcome a large and distorted disparity, particularly near the poles. Latitude-aware convolution (LatConv) is designed to generate more robust features to circumvent the distortion and keep the image quality. We also develop synthetic 360-degree image datasets and introduce a synthetic-to-real learning scheme that transfers knowledge learned from synthetic 360-degree images to a deep neural network conducting super-resolution (SR) of camera-captured images. The proposed network can learn useful features in the ERP-domain using a sufficient number of synthetic samples. The network is then adapted to camera-captured images through the transfer layer with a limited number of real-world datasets.

Dataset

Synthetic360 Dataset

We use Unity software to render 400 virtual indoor/outdoor scenes. In each scene, we record six reference videos from the virtual cameras positioned in 3D space around the target viewpoint. In half of the scenes, each camera is located 25 cm from the center. In the other half, the distance is set to 50 cm to test with various levels of similarity.

Target viewpoint
Refernce viewpoint 1
Refernce viewpoint 2
Refernce viewpoint 3

Real360 Dataset

We capture 251 groups of real-world 360-degree images formatted to a full ERP, using Samsung Gear360 in various environments. We categorize the images as S, M, and L, whose distances are less than 50 cm, 50 cm ~ 300 cm, and larger than 300 cm, respectively.

Target viewpoint
Disparity level ‘S’
Disparity level ‘M’
Disparity level ‘L’

Qualitative Result

Acknowledgement

This work was supported by the MSIT(Ministry of Science and ICT), Korea, under the ITRC(Information Technology Research Center) support program(IITP-2021-2020-0-01460) supervised by the IITP(Institute for Information & Communications Technology Planning & Evaluation).
This work was supported by the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No. 2020-0-00920, Development of Ultra High Resolution Unstructured Plenoptic Video Storage/Compression/Streaming Technology for Medium to Large Space).