Sorry, you need to enable JavaScript to visit this website.

Immersive Optical-See-Through Augmented Reality. Augmented Reality has been getting ready for the last 20 years, and is finally becoming real, powered by progress in enabling technologies such as graphics, vision, sensors, and displays. In this talk I’ll provide a personal retrospective on my journey, working on all those enablers, getting ready for the coming AR revolution. At Meta, we are working on immersive optical-see-through AR headset, as well as the full software stack. We’ll discuss the differences of optical vs.


This work introduces new method using the singular value decomposition (SVD) to recognise human activities from skeleton motion sequences. The primary focus was on different activity durations, inaccurate placement of the joints and loss of information about position of the joints. For that we needed to develop a robust model. At first, the pose features are created for description of skeleton pose per frame, that is created by directional vectors to alljoint pairwise combinations without repetition.


Priors play an important role of regularizers in image deblurring algorithms. Image priors are frequently studied and many forms were proposed in the literature. Blur priors are considered less important and the most common forms are simple uniform distributions with domain constraints. We propose a more informative blur prior based on the notion of atomic norm which favors blurs composed of line segments and is suitable for motion blur. The prior is formulated as a linear program that can be inserted into any optimization task.


Texture is an indispensable property to develop many vision
based autonomous applications. Compared to colour, feature
dimension in a local texture descriptor is quite large as dense
texture features need to represent the distribution of pixel intensities
in the neighbourhood of each pixel. Large dimensional
features require additional time for further processing
that often restrict real-time applications. In this paper, a robust
local texture descriptor is enhanced by reducing feature


3D scanners generate irregularly distributed cloud of points in
most of the cases. Dealing with such data, often in the form of
triangular meshes, requires a pre-processing step to regularize
the triangle facets shape and size. In this paper, we propose
CSIOR, a novel mesh regularization technique which is capable
of producing quasi-equilateral triangles, and distinguished
by two novel features, namely, its intrinsic ordered aspect and
its preservation of the geometric texture of the surface (relief


Advances in federated learning and edge computing advocate for deep learning models to run at edge devices for video analysis. However, the captured video frame rate is too high to be processed at the edge in real-time with a typical model such as CNN. Any approach to consecutively feed frames to the model compromises both the quality (by missing important frames) and the efficiency (by processing redundantly similar frames) of analysis.


Traditionally, the vision community has devised algorithms to estimate the distance between an original image and images that have been subject to perturbations. Inspiration was usually taken from the human visual perceptual system and how the system processes different perturbations in order to replicate to what extent it determines our ability to judge image quality. While recent works have presented deep neural networks trained to predict human perceptual quality, very few borrow any intuitions from the human visual system.