Image & Video Analysis / Processing

Query by Pictorial Example


Abstract: 
Appearance of real scenes is highly dependent on actual conditions as illumination and viewpoint, which significantly complicates automatic analysis of images of such scenes. In this thesis, we introduce novel textural features, which are suitable for robust recognition of natural and artificial materials (textures) present in real scenes. These features are based on efficient modelling of spatial relations by a type of Markov Random Field (MRF) model and we proved that they are invariant to illumination colour, cast shadows, and texture rotation. Moreover, the features are robust to illumination direction and degradation by Gaussian noise, they are also related to human perception of textures.

Image Retrieval Measures Based on Illumination Invariant Textural MRF Features


Abstract: 
Content-based image retrieval (CBIR) systems, target database images using feature similarities with respect to the query. We introduce fast and robust image retrieval measures that utilise novel illumination invariant features extracted from three different Markov random field (MRF) based texture representations. These measures allow retrieving images with similar scenes comprising colour-textured objects viewed with different illumination brightness or spectrum. The proposed illumination insensitive measures are compared favourably with the most frequently used features like the Local Binary Patterns, steerable pyramid and Gabor textural features, respectively. The superiority of these new illumination invariant measures and their robustness to added noise are empirically demonstrated in the illumination invariant recognition of textures from the Outex database.

A Fast Model-Based Restoration of Colour Movie Scratches


Abstract: 
This work presents a new type of scratch removal algorithm based on a causal adaptive multidimensional multitemporal prediction. The predictor use available information from the neighbourhood of a missing multispectral pixel due to spectral, temporal and spatial correlation of video data but not any information from the failed pixel itself. The model assumes white Gaussian noise in each spectral layer, but layers can be mutually correlated. A significant improvement of the 3D model performance is obtained if the temporal information is included, i.e., using the 3.5D causal AR model. Such information is natural to obtain from previous or/and following frame(s) for which we know all necessary data, due to high between-frame temporal correlation. Thanks to this we can treat data from different frames (specified by the contextual neighbourhood) in the same way, so we attach to each data information about its shift according to predicted pixel placement. The contextual neighbourhood has to be causal (in the reconstructed frame lattice subspace) . It means that the predictor can use only data from the model history. Then if we assume normal-Wishart parameter prior the predictor have analytical (not iterative) solution.

Range Image Segmentation by Curve Grouping


Abstract: 
A range image segmentation method based on a recursive adaptive regression model prediction for detecting range image step discontinuities which are present at object face borders. Detected face borders guides subsequent region growing step where neighbouring face curves are grouped together. Region growing based on curve segments instead of pixels like in classical approaches significantly speed up the algorithm. Curves to be grown are represented using the cubic spline model. Curves from the same region are required to have similar curvature and slope but they do not need to be of maximal length through the corresponding object face.

Fast Segmentation of Planar Surfaces in Range Images


Abstract: 
An algorithm for planar face segmentation in range images. The segmentation is based on a combination of recursive adaptive regression model prediction for detecting range image step discontinuities which are present at object face borders and of a region growing on surface lines. Border pixels are detected in two perpendicular directions and detection results are combined together. Two predictors in each direction use identical contextual information from the pixel's neighbourhood and they mutually compete for the most optimal discontinuity detection.
Syndicate content