- Full Conference Pass
- Full Conference 1-Day Pass
Date/Time: 28 November 2017, 09:00am - 10:45am
Venue: Amber 1
Location: Bangkok Int'l Trade & Exhibition Centre (BITEC)
Session Chair: Min H. Kim, Korea Advanced Institute of Science and Technology (KAIST), South Korea
Removal of Ghosting Artefacts in HDRI using Intensity Scaling Cue
Summary: Motion artefacts (ghosting) can be produced in HDR images due to moving objects in the input LDR dataset. This work aligns all LDR images with a reference image, and thus removes ghost artefacts in the HDR image. This is done without explicit detection of moving regions in LDR images. Intensity scaling factor between the reference image and target image is estimated by taking into account only non-moving and non-saturated pixels, and it is used as an index to determine moving pixels in target image. Finally, the values of the moving pixels in target image are corrected with their expected values. The method is more accurate than the existing methods with competitive computation speed.
Author(s): Seong-O Shim, University of Jeddah
Ishtiaq Rasool Khan, University of Jeddah
Speaker(s): Seong-O Shim, Faculty of Computing and IT, University of Jeddah, Saudi Arabia
Multi-video Object Synopsis Integrating Optimal View Switching
Summary: In this paper, we propose to integrate multi-video object synopsis with the display method of optimal multi-view switching. We group occurrences of the same object in different cameras together and operate them in the same way, based on which we propose a unified framework that optimizes object-based synopsis variables and multi-view switching variables simultaneously. The unified formulation encodes activity, occlusion, chronological order, and view switching costs. We develop an alternative optimization scheme composed of graph cuts and dynamic programming to minimize the combined costs efficiently. This integrated framework provides a more user convenient browsing approach for object-based synopsis results and allows to further reduce occlusion artifacts between synopsized objects. Experiments demonstrate the effectiveness and user convenience of our multi-video synopsis method.
Author(s): Zhensong Zhang, The Chinese University of Hong Kong, CUHK
Yongwei Nie, South China University of Technology
Hanqiu Sun, The Chinese University of Hong Kong
Qiuxia Lai, The Chinese University of Hong Kong
Guiqing Li, South China University of Technology
Speaker(s): Zhensong Zhang, The Chinese University of Hong Kong
Visual Enhancement via Reinforcement Parameter Learning for Low Backlighted Display
Summary: In this technical brief, we propose a system to enhance the image on low backlighted display in order to save electrical power for mobile devices. In addition to our brightness compensation,we also take human visual perception, such as just-noticeable difference (JND) and saliency, into account. To integrate these characteristics in a system, we need to adjust several parameters. Accordingly, we introduce our reinforcement parameter learning into the system. By taking actions and analyzing rewards, we can train these characteristic parameters off-line and test the image online. Experimental results show that our visual enhancement via reinforcement parameter learning outperforms the existing systems.
Author(s): Chih-Tsung Shen, Industrial Technology Research Institute
Ching-Hao Lai, Industrial Technology Research Institute
Yi-Ping Hung, Department of Computer Science and Information Engineering, National Taiwan University
Soo-Chang Pei, National Taiwan University
Speaker(s): Chih-Tsung Shen, Industrial Technology Research Institute
GPU Based Techniques for Deep Image Merging
Summary: We explore GPU based merging of deep images using different memory layouts for fragment lists - linked lists, linearised arrays, and a new interleaved arrays approach. We also report performance improvements for merging techniques which leverage GPU memory hierarchy and layout by processing blocks of fragment data using fast registers based on similar techniques used to improve performance of transparency rendering. Our results show a 2-5X improvement from combining these techniques.
Author(s): Jesse Archer, RMIT University
Geoff Leach, RMIT University
Ron van Schyndel, RMIT University
Speaker(s): Jesse Archer, RMIT University