Sorry, you need to enable JavaScript to visit this website.

High dynamic range (HDR) image formation from low dynamic range (LDR) images of different exposures is a well researched topic in the past two decades.
However, most of the developed techniques consider differently exposed LDR images that are acquired from the same camera view point, which assumes the scene to be static long enough to capture multiple images.
In this paper, we propose to address the problem of HDR imaging from differently exposed LDR stereo images using an encoder-decoder based convolutional neural network (CNN).


In this paper we address the problem of jointly retrieving a 3D dynamic shape, camera motion, and deformation grouping from partial 2D point trajectories in a monocular video. To this end, we introduce a union of piecewise Bézier subspaces with enforcing continuities to model 3D motion. We show that formulating the problem in terms of piecewise curves, allows for a better physical interpretation of the resulting priors and a more accurate representation of the motion.


Deep convolutional neural networks (CNNs), renowned for their consistent performance, are widely understood by practitioners that the stability of learning depends on the initialization of the model parameters in each layer. Kaiming initialization, the de facto standard, is derived from a much simpler CNN model which consists of only the convolution and fully connected layers. Compared to the current CNN models, the basis CNN model for the Kaiming initialization does not include the max pooling or global average pooling layers.


In the last few years, single image super-resolution (SISR) has benefited a lot from the rapid development of deep convolutional neural networks (CNNs), and the introduction of attention mechanisms further improves the performance of SISR. However, previous methods use one or more types of attention independently in multiple stages and ignore the correlations between different layers in the network.


Images captured in low light condition have a narrow dynamic range with a dark tone, which are seriously degraded by noise due to the low signal-to-noise ratio (SNR). Discrete wavelet transform (DWT) is invertible and thus is able to decompose an image into subbands without information loss minimizing redundancy. In this paper, we propose subband adaptive enhancement of low light images using wavelet-based convolutional neural networks. We adopt DWT to achieve joint contrast enhancement and noise reduction. We combine DWT with convolutional neural networks (CNNs), i.e.


Tone-mapping is one of the prevailing methods to overcome high dynamic range imaging limitations over low dynamic range display devices, but the tone-mapped output image may suffer from saturated regions with texture and color information loss. In this paper, a novel approach is proposed to solve the so-called clipping problem in tone-mapped high dynamic range images. A successful saturation correction framework, which relies on linear embeddings, difference of pixel intensities and gradient-guided block-search, is developed as a post-processing technique to tone-mapping.