In this section, we assess how the proposed super-ray construction method deals with occluded and dis-occluded parts, and to which extent the super-rays are consistent despite uncertainty
on the disparity information. The following figure shows examples of super-rays obtained with different synthetic light fields ( butterfly and Monasroom) and real light fields captured by
a Lytro Ilum camera (Flower 2, Rock, FountainVincent and StonePillarInside).In the first three columns, we have the original top left corner view, its corresponding disparity map and super
pixel segmentation using the SLIC algorithm [ref] respectively. In the fourth column, we show horizontal and vertical epipolar segments taken both from the 4D light field color information
and our final segmentation in specific regions of the image (the red blocks). We can see that we are following well the object borders, especially when the disparity map is reliable. Also
, we have always attained a high percentage of coherent super-rays across views (higher than $40\%$ as measured with Cons(%) in the fifth column). More precisely, Cons(%) gives the
percentage of coherent super-rays: A super-ray is coherent when it is made of super pixels having the same shape in all the views, with or without a displacement.
Animated segmentation maps
Fountain Vincent |
Flower 2 |
|
|
Local Graph Transforms
Local graph construction
In order to jointly capture spatial and angular correlations between pixels in the light field, we first consider a local non separable graph per super-ray.
More precisely, if we consider the luminance values in the whole light field and a segmentation map S, the super-ray K can be represented by
a signal defined on an undirected connected graph G = {V,E} which consists of a finite set V of vertices.
A set E of edges are built as follows.
- We first connect each pixel (m,n,x,y) in the set V and its 4-nearest neighbors in the spatial domain (i.e. the top, bottom, left and right neighbors)
- We then find the median disparity value d of the pixels inside the super-ray k in the top-left view. Using this disparity value, we project each pixel in super-ray k in the 4
nearest neighboring views (i.e. the top, bottom, left and right neighboring view). We end up with four projected pixels. If a projected pixel belongs to the set of vertices V, then we connect it to the original pixel.
Local Non Separable and Separable graph transform
The laplacian is computed as:
The non separable graph transform is defined as the projection of the graph signal onto the eigenvectors of the laplacian.
However, some high complexity remains. That's why we propose to use instead separable graph transforms where we perform a spatial transform followed by an angular transform.
The spatial laplacian is defined in each view. Using the eigenvectors of the spatial laplacian, we can compute a spatial transform inside each view.
We can then define a band vector out of all observations of a specific band across the views and construct an angular graph to capture inter-view correlations.
Using the eigenvectors of the angular laplacian, we compute the second angular transform.
For the shape varying super-rays, we end up with incompatible basis functions that will no longer preserve the angular correlations.