Next Article in Journal
Reliable Estimation of Deterioration Levels via Late Fusion Using Multi-View Distress Images for Practical Inspection
Next Article in Special Issue
Liouville Integrability in a Four-Dimensional Model of the Visual Cortex
Previous Article in Journal
Methodology for the Automated Visual Detection of Bird and Bat Collision Fatalities at Onshore Wind Turbines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Frequency Image Completion via a Biologically-Inspired Sub-Riemannian Model with Frequency and Phase

CNRS/NeuroPSI, Campus CEA, 91400 Saclay, France
J. Imaging 2021, 7(12), 271; https://doi.org/10.3390/jimaging7120271
Submission received: 27 October 2021 / Revised: 24 November 2021 / Accepted: 1 December 2021 / Published: 9 December 2021

Abstract

:
We present a novel cortically-inspired image completion algorithm. It uses five-dimensional sub-Riemannian cortical geometry, modeling the orientation, spatial frequency and phase-selective behavior of the cells in the visual cortex. The algorithm extracts the orientation, frequency and phase information existing in a given two-dimensional corrupted input image via a Gabor transform and represents those values in terms of cortical cell output responses in the model geometry. Then, it performs completion via a diffusion concentrated in a neighborhood along the neural connections within the model geometry. The diffusion models the activity propagation integrating orientation, frequency and phase features along the neural connections. Finally, the algorithm transforms the diffused and completed output responses back to the two-dimensional image plane.

1. Introduction

Visual perception has drawn the attention of experts from fields of philosophy, psychology and neuroscience, as well as the attention of mathematicians and physicists working on perceptual modeling. The question of how we perceive was studied by Edmund Husserl in their pioneering philosophical texts in phenomenology  [1,2,3]. With regard to psychology, we can think of the well-known Berlin school of experimental psychology, Gestalt psychology school, which formulated what is known today as Gestalt psychology of perception [4,5,6].
Gestalt psychology is a theory which attempts to provide the principles giving rise to a meaningful global perception of a visual scene by starting from the local properties of the objects within the scene. The main idea of Gestalt psychology is that the mind constructs the global whole by grouping similar fragments rather than purely summing the fragments as if they were indifferent. In terms of visual perception, those similar fragments can be thought of as point stimuli with the same (or closely) valued features of the same type. In the present paper, we consider orientation, spatial frequency and phase as features. We extract these features from a given two-dimensional grayscale input image which is partially occluded and reveal those occluded parts via an integration of the extracted features. By doing so, we follow the Gestalt principle, which is called the law of good continuity, to provide a feature integration mechanism reconstructing the occluded parts in the image.
The law of good continuity states that we group aligned pieces rather than those with sharp abrupt directional changes when we perceive an object as a whole which is formed by fragments; see Figure 1. This law was studied by Field, Hayes and Hess [7] from a psychophysical point of view. They studied how the visual system captures aligned fragments constituting lines or cocircular curves on a background of randomly oriented fragments; see Figure 2. They noticed that the aligned patterns captured by the visual system were locally overlapping with what they called association fields, which are shown in Figure 3. Those fields provide a geometric characterization of the law of good continuity, and they can be interpreted as the psychophysical representations of the neural connections which are biologically implemented in the visual cortex.
The visual system is capable of perceiving a lacunar curve as complete by capturing the whole curve pattern underlying the lacunar curve; see Figure 1. This is due to a general phenomena which is called perceptual completion. This phenomena provides the perception of contours and figures which are actually not present in the visual stimulus. We call such contours subjective contours.
Kanizsa [8,9] explains two categories of completion: modal completion and amodal completion. The first one refers to completion following the modality of vision. There is no direct stimulus corresponding to the object, yet the perceived object is indistinguishable from the real stimuli and we perceive it as a whole. In the second category, the completion does not make use of the modality of vision. In other words, the stimulus corresponding to the object is partial, but we still perceive the object as complete. In this case, the object is recognized as a whole, although only some specific fragments of the object evoke our sensory receptors. Examples of those two categories of completion resulting in subjective contours are the Kanizsa triangle and Ehrenstein’s illusion, which are illustrated in Figure 4. We focus on amodal completion in the present work, and provide an approach which reconstructs the occluded parts in a compatible way with the law of good continuity and association fields, where those two notions are considered in an extended way based on position, orientation, frequency and phase alignment.
The primary visual cortex (V1) is the main area of the cerebral cortex which is responsible for the first step processing of visual input so that a proper visual perception is achieved at a higher perceptual level. V1 contains a particular family of neurons, simple cells. These neurons are locally sensitive to visual features, such as orientation [10,11,12,13,14], spatial frequency [15,16,17,18,19,20,21,22], phase [23,24,25,26], scale [27] and ocular dominance [17,28,29]. The simple cells are organized in a hypercolumnar architecture, which was first discovered by Hubel and Wiesel [13]. In this architecture, a hypercolumn is assigned to each point ( x , y ) of the retinal plane M R 2 , and the hypercolumn contains all the simple cells sensitive to a particular value of the same feature type. The simple cells are able to locally exctract the feature values of the visual stimulus, and the activity propagation along the neural connections between the simple cells integrates those values to a coherent global unity. Those two mechanisms, the feature detection and the neural connectivity, comprise the functional geometry of V1.
A simple cell is identified by its receptive field, which is defined as the domain of the retina to which the cell is sensitive and connected through the retino-geniculo-cortical paths. Once a receptive field is stimulated, it evokes a spike transmitted to the corresponding simple cells. Each one of those simple cells produces a response to the spike. This response is what is known as receptive profile.
Hoffman proposed modeling the hypercolumnar architecture of V1 in terms of a contact bundle structure [30,31]. This framework was followed by Petitot and Tondut, where they improved the contact bundle structure and proposed a boundary completion method within the corresponding contact geometry [32]. Moreover, they geometrically interpreted the association fields as the integration along the vector fields generating the contact geometry. This setting was developed further by Citti and Sarti [33] to a framework in which they introduced a group-based approach to study the geometric modeling of V1 hypercolumnar architecture and the functional connectivity. They used the Gabor function as the receptive profile model and proposed the sub-Riemannian geometry of the group of rotations and translations ( SE ( 2 ) ) as the V1 model geometry. This framework was extended to higher dimensional geometries where scale [34] and velocity [35,36] were taken into account. Various biologically inspired models for optical illusions [37,38,39,40,41] and orientation preference maps [42], as well as frameworks for image processing [43,44,45,46,47,48,49,50], for pattern recognition [51] and for medical applications [52,53], were proposed in the sub-Riemannian geometry of SE ( 2 ) .
The model presented in [33] is an abstract geometric description of the orientation-sensitive V1 hypercolumnar architecture reported by Hubel and Wiesel [10,11,12]. This description provides a good phenomenological approximation of the biologically implemented V1 neural connections, which were reported by Bosking et al. [54]. In this model framework, the projections of a particular family of curves onto the two-dimensional image plane provide good approximations of the association fields. In other words, these curves model the V1 neural connections; see Figure 5. They are called horizontal integral curves and they are obtained by integration along the vector fields generating the SE ( 2 ) sub-Riemannian geometry. For this reason, the approach considered by Citti, Petitot and Sarti and used in our present paper is referred to as biologically inspired.
The sub-Riemannian model geometry proposed in [33] was extended in a recent work to multi-frequency and multi-phase setting [55]. The extended model corresponds to a natural geometry which is derived [55,56] from one of the very first perceptual mechanisms of vision: receptive profile. It is different from the classical approach, in which a suitable geometry is assigned to the neural responses represented in terms of receptive profiles. This extended model takes advantage of orientations, spatial frequencies and phases in a given 2D input image to encode the visual information. This is not the case in the sub-Riemannian model proposed in [33], which uses the SE ( 2 ) geometry, and in which only orientation can be represented as a visual feature. The extended model geometry was applied to enhancement on images in which several spatial frequencies were equally present [55], such as texture images. In the present work, we employ the same extended multi-frequency sub-Riemannian setting presented in [55] and propose an image completion algorithm within, which is aimed for use with grayscale texture images containing multiple spatial frequencies.
In Section 2, we explain the model framework which was introduced in [55] and its relation to our completion algorithm. In Section 3, we present a specific family of integral curves defined in the model geometry, which are the models of the neural connections in V1. Then, in Section 5, we introduce our completion algorithm, its discrete scheme and the corresponding pseudocode. Finally, we provide our simulation results and their comparison to some results obtained previously in [48,50]. At the end, we give the main conclusions and some perspectives for related future research.

2. Model Framework

In this section, we explain the cortical model geometry in which our completion algorithm is defined. This model framework is composed of two mechanisms: feature value extraction and neural connectivity [55].

2.1. Feature Value Extraction

Each simple cell is sensitive to a specific part of the retina, which is called the receptive field. Once the receptive field is stimulated by visual stimulus, the retinal cells in the receptive field produce spikes which are transmitted through retino-geniculo-cortical pathways to the related simple cells in V1. Each simple cell generates a response to those spikes, which is the receptive profile corresponding to the simple cell. In other words, the receptive profile is the impulse response function of the simple cell. The simple cell receptive profile which is sensitive to the stimulus located at q M on the image plane M and selective to the set of feature values z S 1 × R + × S 1 is denoted by Ψ ( q , z ) , where q = ( q 1 , q 2 ) and z = ( θ , f , ϕ ) together denote a fixed point ( q , z ) in Q R 2 × S 1 × R + × S 1 . Here, Q represents the five-dimensional sub-Riemannian V1 model geometry.
Simple cell receptive profiles can be modeled in terms of Gabor functions [33,41,55,57]. In the orientation, frequency and phase selective model framework, the receptive profile of a simple cell is a Gabor function of the following type:
Ψ ( q , z ) ( x , y , s ) : = 1 2 σ 2 e i r · ( x q 1 , y q 2 ) ( s ϕ ) e | x q 1 | 2 + | y q 2 | 2 2 σ 2 ,
where f > 0 represents the spatial frequency (spatial frequency is found via f = 1 λ , where λ > 0 denotes the wavelength), r = ( r 1 , r 2 ) = ( f sin θ , f cos θ ) and σ > 0 is the scale of the localizing Gaussian. The complex exponential is the wave content, and it is the main component capturing the orientation, frequency and phase information of the objects in a given two-dimensional image. The second exponential is Gaussian, which spatially localizes the receptive profile around the point ( q 1 , q 2 ) . Frequency f determines how many wave peaks are found within the localizing Gaussian window; see Figure 6. As the number of wave peaks increases, the Gabor function can detect higher frequencies. Orientation θ is the orientation angle to which the simple cell receptive profile is sensitive. Parameter ϕ is the reference phase and it creates a phase shift in the waves of the Gabor function as it changes.Here, s ϕ represents the phase centered at ϕ and parameter ϕ is the reference phase; it creates a phase shift in the waves of the Gabor function as it changes. In other words, the even and odd symmetricity of the receptive profile is modulated with phase shift value to take into account all possible receptive profile patterns which can occur at a given time instant. Note that q = ( q 1 , q 2 ) and ( x , y ) denote the spatial parameters and spatial variables, respectively. Therefore, the vector fields in the cortical space Q are written in terms of derivatives with respect to the variables x and y, as is explained further.
We always consider a two-dimensional grayscale static image as a visual input. Each pixel on the image is thought of as a point stimulus. Normally, once the subject is exposed to the image, each pixel is projected onto the retinal surface, from which it is mapped to the cortical surface via retino-cortical maps. For the sake of simplicity, we disregard the coordinate map between the image plane and the retinal surface, as well as the retino-cortical map between the retinal surface and the V1 surface. We assume that the image plane is identically mapped to the retinal and cortical surfaces. We assume the responses of the simple cells to be linear and we compute the output response of a simple cell located at ( q , z ) Q to a given two-dimensional grayscale image I : M [ 0 , 1 ] via the convolution with the Gabor filter:
O I ( q , z ) = M Ψ ( q , z ) ( x , y , 0 ) I ( x , y ) d x d y .
We apply the convolution for every feature value z and for every point q. Consequently, we obtain the output responses of all receptive profiles corresponding to the V1 simple cells. We sometimes call this set of output responses lifted image and the Gabor transform, lifting. It is equivalent to the result of a multi-frequency Gabor transform applied to the given two-dimensional input image. Those output responses are representations of the feature values in the five-dimensional V1 model geometry Q.
In general, static receptive profile models based on linear filter banks and static nonlinearities [33,55,58,59,60,61] provide good responses to simple stimuli. However, their responses to complicated stimuli, such as natural images, are approximate up to a certain level. Several mechanisms such as response normalization, gain controls, cross-orientation suppression and intra-cortical modulation can result in radical changes in the receptive profile shape. Therefore, the aforementioned Gabor filter bank model for the receptive profiles should be considered as a first approximation of highly complex real dynamic receptive profile.
We employ all frequency components of the Gabor transform during the lifting. Therefore, exact inverse Gabor transform is valid, and we use it to obtain the corresponding two-dimensional image to the output responses:
I ( q 1 , q 2 ) = f Ψ L 2 Q O I ( x , y , z ) Ψ ¯ ( x , y , z ) ( q 1 , q 2 , 0 ) d x d y d z ,
with Ψ ¯ denoting the complex conjugate of the corresponding Gabor function Ψ .

2.2. Horizontal Connectivity

Lifting provides the output responses, which are complex valued functions in the five-dimensional model geometry Q. Each output response encodes the feature values corresponding to the orientation, frequency and phase of a pixel defined on the two-dimensional image plane. The output responses, the simple cells, are isolated from each other once lifting from the image plane M to the model geometry Q takes place. Therefore, the model geometry Q should be endowed with an integration mechanism which provides activity propagation, and therefore interactivity, between simple cells. The activity propagation provides an integrated form of the local feature vectors associated with the lifted image. This propagation is concentrated along a specific family of integral curves, horizontal integral curves, corresponding to the model geometry Q. The horizontal integral curves can be thought of as the models of the long range lateral connections, which connect the simple cells residing in different hypercolumns but selective to the same (or close) feature values.
We may associate the following differential one-form to each receptive profile described by (1):
Θ ( θ , f ) = f sin ( θ ) d x + f cos ( θ ) d y d s .
The one-form naturally induces the horizontal vector fields corresponding to the model geometry Q. The horizontal vector fields are formally defined as the elements of
ker ( Θ ) : = { X T Q : Θ ( X ) = 0 } ,
where T Q denotes the tangent bundle of Q. The horizontal vector fields corresponding to Q are found from (5) as
X 1 = cos ( θ ) x + sin ( θ ) y , X 2 = θ , X 3 = sin ( θ ) x + cos ( θ ) y + f s , X 4 = f .
Those horizontal vector fields endow Q with a sub-Riemannian structure, as explained in [55]. The sub-Riemannian structure is composed of the manifold Q, the horizontal tangent bundle H Q , which is a subbundle of the tangent bundle T Q , and a scalar product g describing a notion of distance on Q. It is expressed as ( Q , H Q , g ) . The horizontal vector fields span the horizontal tangent space H ( q , z ) Q at each ( q , z ) Q . The horizontal tangent space can be thought of as the analog of the Euclidean tangent space. In other words, differential operators such as gradient and Laplacian are defined in terms of the horizontal vector fields in the sub-Riemannian geometry. We remark that the differential operators are degenerate since the horizontal tangent space is spanned by four vector fields although it corresponds to Q, which is a five-dimensional geometry. The horizontal integral curves are defined as the integrated curves along the horizontal vector fields given in (6). Despite the degenerate character of the horizontal tangent space H ( q , z ) Q , they provide full connectivity in Q due to the fact that X 1 and X 2 do not commute, as we see below.
Nonzero commutators of the horizontal vector fields are found as
[ X 1 , X 2 ] = sin ( θ ) x cos ( θ ) y , [ X 2 , X 3 ] = cos ( θ ) x sin ( θ ) y , [ X 3 , X 4 ] = s .
The horizontal vector fields are bracket generating since
T ( q , z ) Q = span ( X 1 , X 2 , X 3 , X 4 , [ X 1 , X 2 ] ) | ( q , z ) ,
for all ( q , z ) Q where T ( q , z ) Q denotes the tangent space at ( q , z ) Q . Indeed, (8) shows that the horizontal vector fields fulfill the Hörmander condition [62]. Consequently, they provide the connectivity of any two points in Q through the horizontal integral curves due to the Chow–Rashevskii theorem [63,64,65]. This connectivity property has particular importance since it ensures that any two points in the V1 sub-Riemannian model geometry Q can be connected via the horizontal integral curves, which are the models of the neural connections implemented biologically in V1 and are close approximations of the association fields at the psychophysical level.
We emphasize that the horizontal tangent space is a natural consequence of the choice of the receptive profile function, which is the Gabor function given in (1). The one-form given in (4) induces the horizontal vector fields, which characterize the horizontal tangent space at every point in the cortical space. The horizontal tangent space is endowed with the aformentioned scalar product g, which is defined only in the horizontal tangent space. This scalar product provides a measure of the length of a horizontal curve γ connecting not only two spatial points, but also the corresponding orientation, frequency and phase values to those two spatial points. Therefore, the notion of length defined on the horizontal tangent space H ( q , z ) takes into account how close feature values two simple cells are sensitive to. The length of a horizontal curve is described via ( γ ) = g ( γ ( t ) , γ ( t ) ) d t . The horizontal curve γ connecting two points in the cortical space Q, i.e., two simple cells, and minimizing the length is called geodesic. Geodesics can be used as models of the neural connectivity in which a principle of minimizing the energy of activated neurons can also be taken into account. However, the computation of geodesics and the geodesic flow in a high-dimensional sub-Riemannian geometry such as Q is cumbersome and too mathematically involved. Therefore, in the present paper, we restrict ourselves to horizontal curves in general, and consider establishing the neural connectivity based on geodesics in Q as the refinement of the presented framework here.

3. Horizontal Integral Curves

The association fields were proposed to be modeled by the horizontal integral curves of SE ( 2 ) in the classical orientation-sensitive framework [33]. A similar line of thought was followed in [55], and it was proposed to employ the horizontal integral curves corresponding to the the five-dimensional sub-Riemannian geometry Q as the cortical counterparts of the association fields. The projection of those horizontal integral curves of Q are the same as the projections of the horizontal integral curves of SE ( 2 ) , which are shown in Figure 5. It was conjectured in [55] that the horizontal integral curves of Q coincide with the long-range lateral connections between orientation, frequency and phase-selective simple cells in V1.
Let us denote a time interval by I = [ 0 , T ] with 0 < T < and consider a horizontal integral curve ( q 1 , q 2 , θ , f , ϕ ) = γ : I M associated with the horizontal vector fields given in (6). We denote the initial point of γ by α ^ = ( q ^ 1 , q ^ 2 , θ ^ , f ^ , ϕ ^ ) and its velocity by γ . At each time instant t I , the velocity is written as a vector γ ( t ) span ( X 1 , X 2 , X 3 , X 4 ) γ ( t ) at γ ( t ) = ( q 1 ( t ) , q 2 ( t ) , θ ( t ) , f ( t ) , ϕ ( t ) ) Q . One way to compute the horizontal integral curves starting from the initial point α ^ is to solve the following ODE system for all t I :
γ ( t ) = ( c 1 X 1 + c 2 X 2 + c 3 X 3 + c 4 X 4 ) | γ ( t ) ,
where c 2 , 3 , 4 denote the coefficients. In the activity propagation machinery which we propose here, the propagation is concentrated along a neighborhood of the horizontal integral curves with constant coefficients c 2 , 3 , 4 in the cortical space Q [55]; see Figure 7 and Figure 8. However, the coefficients need not necessarily be constants in the generic framework of horizontal integral curves.

4. Sub-Riemannian Diffusion in the Cortical Space

The activity propagation was proposed to be modeled in terms of a sub-Riemannian diffusion process in the classical orientation-sensitive SE ( 2 ) framework [33]. This sub-Riemannian diffusion process can be interpreted as the model of the interacting neural dynamics defined in terms of the corresponding horizontal vector fields, and it was applied to several biologically inspired image processing algorithms [43,44,45,50].
We follow a similar approach as in [33] and define a sub-Riemannian diffusion procedure in the five-dimensional model geometry Q. We denote by Σ Q the subspace in which all output responses O I s are defined. These are the output responses obtained as the lifting of the two-dimensional input image I. We denote by Π Σ the subspace in which we only find the output responses corresponding to the part to be completed in the image I. The sub-Riemannian diffusion operator for all ( q , z ) Σ is defined as
L : = X 1 2 + β 2 2 X 2 2 + β 3 2 X 3 2 + β 4 2 X 4 2 ,
where X i 2 denotes the ith second-order horizontal derivative, and β 2 , 3 , 4 are coefficients assuring the unit coherency in spatial, orientation, frequency and phase dimensions. The sub-Riemannian diffusion is described for all ( q , z ) Σ by
t u ( q , z , t ) = L u ( q , z , t ) , u ( q , z , 0 ) = O I ( q , z ) , t ( 0 , T ] , 0 < T < ,
with the corrupted region marked by the boundary conditions given by u ( q ˜ , z ˜ , t ) = O I ( q ˜ , z ˜ ) for all ( q ˜ , z ˜ ) Σ Π and t ( 0 , T ] . Here, T denotes a sufficiently large final time and u : Q × [ 0 , T ] C stands for the output responses evolving in time. We denote the number of orientations, frequencies and phases of the N × N image I by K, L and M, respectively. We consider the cortical space to be periodic in orientation, frequency and phase dimensions to guarantee the diffusion to remain in axes boundaries corresponding to those dimensions. Then, the coefficients are found as:
β 2 = K N 2 , β 3 = L N 2 , β 4 = M N 2 .
We note that span ( X 1 , X 2 ) and span ( X 3 , X 4 ) define two subspaces of the horizontal tangent space H ( q , z ) Q at each point ( q , z ) Q . This allows us to decompose the sub-Riemannian horizontal tangent space into two components of which each one is defined by the vector fields of those two subspaces of T ( q , z ) Q . Consequently, we may approximate the sub-Riemannian diffusion described by (10) as a diffusion applied in each frequency and phase channel separately. More precisely, we apply the classical sub-Riemannian diffusion procedure defined in SE ( 2 ) [33,45,50] for each frequency and phase channel separately by using the SE ( 2 ) sub-Riemannian diffusion operator
L ˜ = X 1 2 + β 2 2 X 2 2 ,
where β 2 provides the unit coherency between the diffusion components in spatial and orientation dimensions. The approximate sub-Riemannian procedure is described by
t u ( q , z , t ) = L ˜ u ( q , z , t ) , u ( q , z , 0 ) = O I ( q , z ) ,
with the boundary conditions given as u ( q ˜ , z ˜ , t ) = O I ( q ˜ , z ˜ ) . The advantage of such approximation is that we perform the diffusion procedure in L three-dimensional spaces instead of in a five-dimensional geometry by still taking advantage of the frequency information extracted from the input image. We implement both the exact and the approximate procedures by using a simple forward Euler scheme in which the derivatives are implemented in terms of B-spline interpolated central finite differences.

5. Algorithm

Our algorithm is based on three steps. Once a two-dimensional grayscale input image is given, the first step is lifting the image via the convolution with Gabor functions, as given in (2). This provides the output responses which encode the orientation, frequency, phase values and which are represented in the five-dimensional model geometry. The second step is the sub-Riemannian diffusion described by (10) or by (12) if the approximate setting is used. In time, we integrate (10) for the exact framework and (12) for the approximate framework by applying it on the output responses via iterating it with the time step Δ t until the final time T, at which the steady state is reached, i.e.,  t u = 0 . We employ an explicit method where we use B-splined interpolated finite differences to implement the horizontal vector fields given in (6). The final step is to transform back the evolved output responses to the two-dimensional image plane. This is achived by the inverse Gabor transform given by (3).

5.1. Discretization of the Output Responses

We employ a uniform spatial grid to discretize the image plane such that
I [ i , j ] = I ( i Δ x , j Δ y ) ,
where i , j { 1 , 2 , , N } , with N denoting the image size and Δ x , Δ y R + denoting the pixel width. In our case, the input images are square images and Δ x = Δ y = 1 in terms of pixel unit. We denote the number of samples in the orientation dimension by K, in the frequency dimension by L and in the phase dimension by M. We express the distance between two adjacent samples in the orientation dimension with Δ θ , in the frequency dimension with Δ f and in the phase dimension with Δ s . Discretized output response O I ( q 1 , i , q 2 , j , θ k , f l , ϕ m ) given to I [ i , j ] on uniform orientation, frequency and phase grids with points θ k = k Δ θ , f l = l Δ f and ϕ m = m Δ s ( k { 1 , 2 , , K } , l { 1 , 2 , , L } , m { 1 , 2 , , M } ) is denoted by
O I [ i , j , k , l , m ] = O I ( q 1 , i , q 2 , j , θ k , f l , ϕ m ) ,
where q 1 , i = i Δ x and q 2 , j = j Δ y .
The Gabor function given by (1) is written in the discrete setting as follows:
Ψ [ i , j , k , l , m ] [ i ˜ , j ˜ , n ˜ ] = Ψ ( q 1 , i , q 2 , j , θ k , f l , ϕ m ) ( x i ˜ , y j ˜ , s n ˜ ) ,
where i ˜ , j ˜ { 1 , 2 , , N } , k ˜ { 1 , 2 , , K } , n ˜ { 1 , 2 , , M } , for each orientation θ k , frequency f l and phase ϕ m . We fix s n ˜ = 0 and express the discretized cell response obtained from the input image I [ i , j ] via the lifting described by the discrete Gabor transform as follows:
O I [ i , j , k , l , m ] = i ˜ , j ˜ Ψ [ i , j , k , l , m ] [ i ˜ , j ˜ , 0 ] I [ i ˜ , j ˜ ] .
We discretize the time interval by V N + samples and denote it by h v . Here, h v is the time instant h v = v Δ t , with Δ t satisfying T = V Δ t and v { 1 , 2 , , V } . Discretized evolving output response is written as
U [ i , j , k , l , m , v ] = u ( q 1 , i , q 2 , j , θ k , f l , ϕ m , h v )
Finally, the discrete inverse transform applied via a normalized kernel Ψ ¯ on the evolved output responses until the final time T gives the completed two-dimensional image I T , which is found as follows:
I T [ i , j ] = i ˜ , j ˜ , k ˜ , m ˜ f l ˜ l ˜ U T [ i ˜ , j ˜ , k ˜ , l ˜ , m ˜ ] Ψ ¯ [ i ˜ , j ˜ , k ˜ , l ˜ , m ˜ ] [ i , j , 0 ] .

5.2. Explicit Scheme with Finite Differences

Here, we provide the explicit numerical scheme which we employ to iterate the exact and approximate frameworks given in (10) and (12), respectively. Our motivation for choosing an explicit scheme rather than an implicit scheme is that the latter requires large memory and computational power in our multidimensional framework.
We follow [44,66] to implement the horizontal vector fields given in (6) via B-spline interpolated central finite differences. The interpolation takes place on a uniform grid. It is needed since the horizontal vectors are not always aligned with the spatial grid point samples. B-spline interpolation is based on the coefficients b ( i , j )
sp ( x , y ) = i , j Z b ( i , j ) ρ ( x i , y j ) .
The coefficients are determined such that the spline polynomial sp ( x , y ) , together with the B-spline basis functions ρ ( x i , y j ) , coincides with the horizontal derivatives at the grid points. For example, the condition sp ( i Δ x , j Δ y ) = X 1 O I [ i , j , k , l , m ] must be satisfied once the correct coefficients b are determined; see [44,66,67] for more details.
We define
e ξ k : = ( Δ x cos ( θ k ) , Δ y sin ( θ k ) ) , e η k : = ( Δ x sin ( θ k ) , Δ y cos ( θ k ) ) ,
whose illustrations corresponding to Δ x = Δ y = 1 case are given in Figure 9. We abuse the notation to denote the evolving output responses:
U = U [ i , j , k , l , m , v ] = u ( q 1 , i , q 2 , j , θ k , f l , ϕ m , h v ) ,
and then, we write the central finite differences of the second-order horizontal derivatives as
X 1 X 1 U 1 ( Δ x ) 2 ( u ( q + e ξ k , θ k , f l , ϕ m ) 2 u ( q , θ k , f l , ϕ m ) + u ( q e ξ k , θ k , f l , ϕ m ) ) , X 2 X 2 U 1 ( Δ θ ) 2 ( u ( q , θ k + 1 , f l , ϕ m ) 2 u ( q , θ k , f l , ϕ m ) + u ( q , θ k 1 , f l , ϕ m ) ) , X 3 X 3 U 1 ( Δ x ) 2 u ( q + e η k , θ k , f l , ϕ m ) 2 u ( q , θ k , f l , ϕ m ) + u ( q e η k , θ k , f l , ϕ m ) + f l 2 ( Δ s ) 2 ( u ( q , θ k , f l , ϕ m + 1 ) 2 u ( q , θ k , f l , ϕ m ) + u ( q , θ k , f l , ϕ m 1 ) ) + f cos ( θ k ) 2 Δ s Δ x ( u ( q + e η k , θ k , f l , ϕ m + 1 ) u ( q e η k , θ k , f l , ϕ m + 1 ) u ( q + e η k , θ k , f l , ϕ m 1 ) + u ( q e η k , θ k , f l , ϕ m 1 ) ) f sin ( θ k ) 2 Δ s Δ x ( u ( q + e ξ k , θ k , f l , ϕ m + 1 ) u ( q e ξ k , θ k , f l , ϕ m + 1 ) u ( q + e ξ k , θ k , f l , ϕ m 1 ) + u ( q e ξ k , θ k , f l , ϕ m 1 ) ) + f 2 2 Δ s ( u ( q , θ k , f l , ϕ m + 1 ) 2 u ( q , θ k , f l , ϕ ) + u ( q , θ k , f l , ϕ m 1 ) ) , X 4 X 4 u [ i , j , k , l , m ] 1 ( Δ f ) 2 u ( q , θ k , f l + 1 , ϕ m ) 2 u ( q , θ k , f l , ϕ m ) + u ( q , θ k , f l 1 , ϕ m ) .
Finally, we write the discretized numerical iteration for (10) and (12) as follows:
U [ i , j , k , l , m , v ] = u ( q i , 1 , q j , 2 , θ k , f l , ϕ m , h v ) = u ( q i , 1 , q j , 2 , θ k , f l , ϕ m , h v 1 ) + Δ t L ¯ u ( q i , 1 , q j , 2 , θ k , f l , ϕ m , h v 1 ) ,
where L ¯ represents the discretized version of either L or L ˜ , depending on which one between the exact and approximate frameworks is chosen. The discretization is achieved by replacing the second-order horizontal derivatives with their discrete versions given in (22).

5.3. Pseudocode of the Algorithm

We denote the processed two-dimensional image at the final time T by I T and the evolving discrete output responses at the time instant v Δ t by U v . Then, we provide a general scheme of the exact completion Algorithm 1 as follows:
Algorithm 1: Completion algorithm pseudocode.
Jimaging 07 00271 i001
The detailed Matlab® code can be found in the following link: https://drive.google.com/file/d/1YET47AxYo5_FQs3ObREAqvQp2KjuKh7n/view (accessible starting from 27 October 2021).

6. Numerical Experiments

We choose σ = 2 as the scale of the localizing Gaussian given in (1) for all experiments. We use 128 × 128 grayscale images in the experiments related to Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15, Figure 16, Figure 17 and Figure 18 and 256 × 256 grayscale image in the experiment corresponding to Figure 19. We use θ { 0 , π 32 , 2 π 32 , 3 π 32 , , 31 π 32 } , ϕ { 0 , π 8 , 2 π 8 , 3 π 8 , 4 π 8 } and Δ t = 0.1 in all the experiments.
Our first results are obtained by using an artificial test image, as shown in Figure 10. In the test image, we have arcs of circles centered at the top left corner and with different radii, with a sinusoidal function whose frequency increases linearly as the radius increases. The arcs with sinusoidal patterns are occluded by zero valued arcs belonging to the circles which are centered at the top right corner. We apply our completion procedure with T = 10 and f { 2 , 2.5455 , 3.0909 , , 8 } . We observe that both the approximate and exact frameworks show a similar performance and they provide proper completion.
In Figure 11, we present two cases of our completion algorithm applied to a real texture image, which is occluded by arcs corresponding to circles with different radii. We perform our completion procedure with f { 2.00 , 2.55 , 3.09 , 3.64 , 4.18 , 4.73 , 5.27 , 5.82 , 6.36 , 6.91 , 7.45 , 8.00 } . Here, T = 10 for the upper row and T = 5 for the bottom row. We use the same parameters for T = 15 in Figure 12 on the same texture image but now occluded by vertical and horizontal bars. In such a set of test images, the challenge is that the bars cross each other. Our completion algorithms provide proper completion of such crossing areas. We observe that in all cases, both the approximate and the exact algorithms are able to complete the occluded parts in Figure 11 and Figure 12.
In order to see the impact of using multiple frequencies in the completion process via the sub-Riemannian diffusion, we perform the sub-Riemannian diffusion by using only one single frequency; see Figure 13 and Figure 14. In other words, f is fixed to a constant both in the lifting procedure and in the sub-Riemannian diffusion. In Figure 13, the same corrupted image given in the bottom row of Figure 11 is used. The paramaters are the same as in the bottom row results of Figure 11. We use f = 2 , 2.55 , 3.64 from left to right. We observe that the performance of single-frequency framework is visibly lower than the multi-frequency framework whose results are presented in Figure 11 due to the presence of a wide range of spatial frequencies in the image to be completed. We observe the same in the single-frequency results presented in Figure 14, which are obtained by using the bottom row corrupted image given in Figure 12. Finally, we note that the inverse Gabor transform in those single-frequency experiments is not well defined since the information corresponding to the other frequency components in the input image is lost in the lifting procedure. In other words, the Parseval formula associated with multi-frequency Gabor transform does not hold any longer [46] (Equation (2)). Therefore, we simply project the processed output responses onto the image plane via a sum over orientation and phase components in the single-frequency results shown in Figure 13 and Figure 14.
In Figure 15 and Figure 16, we perform the same type of experiments as in Figure 11 and Figure 12, respectively, but now with f { 1.00 , 1.27 , 1.55 , 1.82 , 2.09 , 2.36 , 2.64 , 2.91 , 3.18 , 3.45 , 3.73 , 4.00 } and by using a different type of texture image. Here, T = 10 for both rows in Figure 15, and T = 15 for both rows in Figure 16. In Figure 17, we display the single-frequency experiments by using the occluded image found in the bottom row of Figure 15. Similarly, in Figure 18, we show the same single-frequency experiments but now using the occluded image given in the bottom row of Figure 16. Similarly to the case of the previous real test image, single-frequency completion cannot perform a proper completion due to the loss of the information corresponding to the other frequencies present in the input image.
Finally, in Figure 19, we compare the result of our algorithm to the results obtained by applying the algorithms explained in [48,50]. The method proposed in [50] is defined in the classical model framework SE ( 2 ) and it combines the SE ( 2 ) sub-Riemannian diffusion with a concentration mechanism resulting in a diffusion driven motion by curvature in SE ( 2 ) . The algorithm explained in [48] uses a semidiscrete version of the classical model geometry SE ( 2 ) , and this allows us to perform completion via the integration of a parallelizable finite set of Mathieu-type diffusions combined with a dynamical restoring mechanism. The main difference between our method and those previously proposed algorithms is that our model uses a higher order sub-Riemannian geometry; this allows us to take into account frequency and phase information as well. Moreover, we do not combine in our framework the diffusion procedure with a concentration or dynamical restoring mechanism. We see in Figure 19 that our algorithm produces completion results comparable to the other two methods. We observe that our algorithm is able to preserve the contextual information better than the other two, especially the high-frequency structures, thanks to the use of multiple frequencies and the exact inverse Gabor transform. The trade off is that in the corresponding stationary state, especially in the low-frequency parts such as the below-eye region, the completion is weaker compared to the other two methods. This supports the idea that our algorithm is better adapted to the texture images, such as the ones in Figure 11 and Figure 12, than to the natural images, such as the one given in Figure 19. In our simulation in Figure 19, we use f { 1.50 , 2.09 , 2.68 , 3.27 , 3.86 , 4.45 , 5.05 , 5.64 , 6.23 , 6.82 , 7.41 , 8.00 } and T = 50 .

7. Conclusions

In this work, we presented a completion algorithm which uses multiple-frequency and phase channels to take advantage of the spatial frequency and phase information of a given two-dimensional grayscale image. The algorithm consists of three mechanisms: feature extraction, sub-Riemannian diffusion and inverse transform. The first one is a linear filtering of the given input image with the Gabor filter banks. The filtering encodes the visual feature values in the output responses, i.e., orientation, frequency and phase values, of each pixel in the two-dimensional input image. The oputput responses are represented in the five-dimensional sub-Riemannian model geometry. The second mechanism, which is the sub-Riemannian diffusion, models the activity propagation between the simple cells in V1. It is concentrated in a neighborhood along the horizontal integral curves corresponding to the sub-Riemannian model geometry. Those horizontal integral curves are conjectured to be good approximations of the neural connections in V1 [55]. Once they are projected onto the two-dimensional image plane, their projections overlap closely with the association fields as was shown in Figure 5. Resulting from the sub-Riemannian diffusion, subjective contours of amodal completion are reconstructed in the five-dimensional model goemetry. Finally, the inverse Gabor transform provides the representation of the evolved output responses; therefore, the reconstructed subjective contours, together with the rest of the lifted image, are shown on the two-dimensional image plane. This final result is the completed image in which the initially occluded parts are revealed.
One of the novelties of the algorithm is that it is not only an image completion algorithm but it takes into account neurophysiological and psychophysical orientation, frequency and phase constraints observed in the visual cortex. Therefore, it should not be considered as an highly specialized image processing algorithm such as those found in medical imaging, radar imaging, robotics and computer vision. It should be considered rather as an algorithm compatible with a natural geometry which was derived from one of the first step mechanisms of the mammalian visual perception, from receptive profile. In other words, it reflects the cortical architecture. Moreover, it uses multi-frequency and phase channels, which was not the case in the previously proposed completion algorithms using the classical orientation-sensitive SE ( 2 ) sub-Riemannian framework [33,50] and its variant [48]. This allows our algorithm to employ the inverse Gabor transform instead of projecting the processed output responses onto the two-dimensional image plane to provide the completed final image; this was not possible in the aforementioned previous methods [33,48,50]. This provides good preservation of the contextual information (contours, edges etc.) with different frequencies in the input image.
One of the interesting aspects for future work is to consider a concentration mechanism. In the proposed completion algorithm, it is possible to embed a similar concentration mechanism to the one presented in [33], but now with a concentration in each frequency and phase channel. Moreover, the proposed completion algorithm uses the same model geometry as the enhancement algorithm which was presented in [55]. Another interesting future work is to combine those two algorithms to perform both completion and enhancement at the same time by employing orientation, frequency and phase information existing in a two-dimensional input image. Finally, the study of an analytical solution in a similar way as was carried out in SE ( 2 ) [70,71], but this time, a study regarding the sub-Riemannian diffusion defined in the five-dimensional model geometry could provide new techniques to perform the completion task, as well as many other image processing applications in the model geometry. This would open up new interesting questions, especially at theoretical level.

Funding

This research was funded by the Human Brain Project (European Union grant H2020-945539).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The code can be found in the following link: https://drive.google.com/file/d/1YET47AxYo5_FQs3ObREAqvQp2KjuKh7n/view (accessible starting from 27 October 2021).

Acknowledgments

I would like to thank Giovanna Citti and Alessandro Sarti for their inspiring discussions. I hereby acknowledge that I was supported by the Human Brain Project (European Union grant H2020-945539).

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Husserl, E.; Holenstein, E. Logische Untersuchungen. Springer Netherlands: The Hague, The Netherlands, 1975. [Google Scholar]
  2. Husserl, E.; Claesges, U. Ding und Raum: Vorlesungen 1907; Springer Netherlands: The Hague, The Netherlands, 1973. [Google Scholar]
  3. Husserl, E.; Schumann, K. Ideen zu einer reinen Phänomenologie und phänomenologischen Philosophie; Springer Netherlands: The Hague, The Netherlands, 1976. [Google Scholar]
  4. Wertheimer, M. Laws of Organization in Perceptual Forms. In A Source Book of Gestalt Psychology; Ellis, W.D., Ed.; Trubner & Company: Kegan Paul, Trench, 1938; pp. 71–88. [Google Scholar] [CrossRef]
  5. Köhler, W. Gestalt Psychology: An Introduction to New Concepts in Modern Psychology; Liveright: New York, NY, USA, 1970. [Google Scholar]
  6. Koffka, K. Principles of Gestalt Psychology; Routledge: Abingdon, UK, 2013; Volume 44. [Google Scholar]
  7. Field, D.J.; Hayes, A.; Hess, R.F. Contour integration by the human visual system: Evidence for a local “association field”. Vis. Res. 1993, 33, 173–193. [Google Scholar] [CrossRef]
  8. Kanizsa, G. Organization in Vision: Essays on Gestalt Perception; Praeger Publishers: Westport, CT, USA, 1979. [Google Scholar]
  9. Kanizsa, G. Grammatica del Vedere: Saggi su Percezione e Gestalt; Il Mulino: Bologna, Italy, 1980. [Google Scholar]
  10. Hubel, D.H.; Wiesel, T.N. Receptive fields of single neurones in the cat’s striate cortex. J. Physiol. 1959, 148, 574. [Google Scholar] [CrossRef] [PubMed]
  11. Hubel, D.H.; Wiesel, T.N. Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J. Physiol. 1962, 160, 106. [Google Scholar] [CrossRef] [PubMed]
  12. Hubel, D.H.; Wiesel, T.N. Shape and arrangement of columns in cat’s striate cortex. J. Physiol. 1963, 165, 559. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Hubel, D.H.; Wiesel, T.N. Uniformity of monkey striate cortex: A parallel relationship between field size, scatter, and magnification factor. J. Comp. Neurol. 1974, 158, 295–305. [Google Scholar] [CrossRef]
  14. Hubel, D.H.; Wiesel, T.N. Ferrier lecture: Functional architecture of macaque monkey visual cortex. Proc. R. Soc. Lond. Biol. Sci. 1977, 198, 1–59. [Google Scholar]
  15. Maffei, L.; Fiorentini, A. Spatial frequency rows in the striate visual cortex. Vis. Res. 1977, 17, 257–264. [Google Scholar] [CrossRef]
  16. Hübener, M.; Shoham, D.; Grinvald, A.; Bonhoeffer, T. Spatial relationships among three columnar systems in cat area 17. J. Neurosci. 1997, 17, 9270–9284. [Google Scholar] [CrossRef] [Green Version]
  17. Issa, N.P.; Trepel, C.; Stryker, M.P. Spatial frequency maps in cat visual cortex. J. Neurosci. 2000, 20, 8504–8514. [Google Scholar] [CrossRef] [Green Version]
  18. Issa, N.P.; Rosenberg, A.; Husson, T.R. Models and measurements of functional maps in V1. J. Neurophysiol. 2008, 99, 2745–2754. [Google Scholar] [CrossRef] [Green Version]
  19. Sirovich, L.; Uglesich, R. The organization of orientation and spatial frequency in primary visual cortex. Proc. Natl. Acad. Sci. USA 2004, 101, 16941–16946. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Tani, T.; Ribot, J.; O’Hashi, K.; Tanaka, S. Parallel development of orientation maps and spatial frequency selectivity in cat visual cortex. Eur. J. Neurosci. 2012, 35, 44–55. [Google Scholar] [CrossRef] [PubMed]
  21. Ribot, J.; Aushana, Y.; Bui-Quoc, E.; Milleret, C. Organization and origin of spatial frequency maps in cat visual cortex. J. Neurosci. 2013, 33, 13326–13343. [Google Scholar] [CrossRef] [Green Version]
  22. Ribot, J.; Romagnoni, A.; Milleret, C.; Bennequin, D.; Touboul, J. Pinwheel-dipole configuration in cat early visual cortex. NeuroImage 2016, 128, 63–73. [Google Scholar] [CrossRef] [PubMed]
  23. De Valois, K.; Tootell, R.B. Spatial-frequency-specific inhibition in cat striate cortex cells. J. Physiol. 1983, 336, 359–376. [Google Scholar] [CrossRef] [Green Version]
  24. Pollen, D.A.; Gaska, J.P.; Jacobson, L.D. Responses of simple and complex cells to compound sine-wave gratings. Vis. Res. 1988, 28, 25–39. [Google Scholar] [CrossRef]
  25. Levitt, J.B.; Sanchez, R.M.; Smith, E.L.; Movshon, J.A. Spatio-temporal interactions and the spatial phase preferences of visual neurons. Exp. Brain Res. 1990, 80, 441–445. [Google Scholar] [CrossRef]
  26. Mechler, F.; Reich, D.S.; Victor, J.D. Detection and discrimination of relative spatial phase by V1 neurons. J. Neurosci. 2002, 22, 6129–6157. [Google Scholar] [CrossRef] [Green Version]
  27. Blakemore, C.T.; Campbell, F.W. On the existence of neurones in the human visual system selectively sensitive to the orientation and size of retinal images. J. Physiol. 1969, 203, 237. [Google Scholar] [CrossRef]
  28. Shatz, C.J.; Stryker, M.P. Ocular dominance in layer iv of the cat’s visual cortex and the effects of monocular deprivation. J. Physiol. 1978, 281, 267–283. [Google Scholar] [CrossRef] [Green Version]
  29. LeVay, S.; Stryker, M.P.; Shatz, C.J. Ocular dominance columns and their development in layer iv of the cat’s visual cortex: A quantitative study. J. Comp. Neurol. 1978, 179, 223–244. [Google Scholar] [CrossRef]
  30. Hoffman, W.C. Higher visual perception as prolongation of the basic lie transformation group. Math. Biosci. 1970, 6, 437–471. [Google Scholar] [CrossRef]
  31. Hoffman, W.C. The visual cortex is a contact bundle. Appl. Math. Comput. 1989, 32, 137–167. [Google Scholar] [CrossRef]
  32. Petitot, J.; Tondut, Y. Vers une neurogéométrie. fibrations corticales, structures de contact et contours subjectifs modaux. Math. Sci. Hum. 1999, 145, 5–101. [Google Scholar] [CrossRef] [Green Version]
  33. Citti, G.; Sarti, A. A cortical based model of perceptual completion in the roto-translation space. J. Math. Imaging Vis. 2006, 24, 307–326. [Google Scholar] [CrossRef]
  34. Sarti, A.; Citti, G.; Petitot, J. The symplectic structure of the primary visual cortex. Biol. Cybern. 2008, 98, 33–48. [Google Scholar] [CrossRef]
  35. Barbieri, D.; Citti, G.; Cocci, G.; Sarti, A. A cortical-inspired geometry for contour perception and motion integration. J. Math. Imaging Vis. 2014, 49, 511–529. [Google Scholar] [CrossRef] [Green Version]
  36. Cocci, G.; Barbieri, D.; Citti, G.; Sarti, A. Cortical spatiotemporal dimensionality reduction for visual grouping. Neural Comput. 2015, 27, 1252–1293. [Google Scholar] [CrossRef] [PubMed]
  37. Franceschiello, B.; Mashtakov, A.; Citti, G.; Sarti, A. Geometrical optical illusion via sub-Riemannian geodesics in the roto-translation group. Differ. Geom. Its Appl. 2019, 65, 55–77. [Google Scholar] [CrossRef] [Green Version]
  38. Franceschiello, B.; Sarti, A.; Citti, G. A neuromathematical model for geometrical optical illusions. J. Math. Imaging Vis. 2018, 60, 94–108. [Google Scholar] [CrossRef] [Green Version]
  39. Bertalmío, M.; Calatroni, L.; Franceschi, V.; Franceschiello, B.; Villa, A.G.; Prandi, D. Visual illusions via neural dynamics: Wilson–Cowan-type models and the efficient representation principle. J. Neurophysiol. 2020, 123, 1606–1618. [Google Scholar] [CrossRef] [Green Version]
  40. Bertalmio, M.; Calatroni, L.; Franceschi, V.; Franceschiello, B.; Prandi, D. Cortical-inspired Wilson-Cowan-type equations for orientation-dependent contrast perception modelling. J. Math. Imaging Vis. 2021, 63, 263–281. [Google Scholar] [CrossRef]
  41. Baspinar, E.; Calatroni, L.; Franceschi, V.; Prandi, D. A cortical-inspired sub-Riemannian model for Poggendorff-type visual illusions. J. Imaging 2021, 7, 41. [Google Scholar] [CrossRef]
  42. Baspinar, E.; Citti, G.; Sarti, A. A geometric model of multi-scale orientation preference maps via Gabor functions. J. Math. Imaging Vis. 2018, 60, 900–912. [Google Scholar] [CrossRef] [Green Version]
  43. Duits, R.; Franken, E. Left-invariant parabolic evolutions on SE(2) and contour enhancement via invertible orientation scores part I: Linear left-invariant diffusion equations on SE(2). Q. Appl. Math. 2010, 68, 255–292. [Google Scholar] [CrossRef] [Green Version]
  44. Duits, R.; Franken, E. Left-invariant parabolic evolutions on se(2) and contour enhancement via invertible orientation scores part ii: Nonlinear left-invariant diffusions on invertible orientation scores. Q. Appl. Math. 2010, 68, 293–331. [Google Scholar] [CrossRef] [Green Version]
  45. Boscain, U.; Duplaix, J.; Gauthier, J.P.; Rossi, F. Anthropomorphic image reconstruction via hypoelliptic diffusion. SIAM J. Control Optim. 2012, 50, 1309–1336. [Google Scholar] [CrossRef]
  46. Duits, R.; Führ, H.; Janssen, B.; Bruurmijn, M.; Florack, L.; van Assen, H. Evolution equations on Gabor transforms and their applications. Appl. Comput. Harmon. Anal. 2013, 35, 483–526. [Google Scholar] [CrossRef]
  47. Boscain, U.; Gauthier, J.P.; Prandi, D.; Remizov, A. Image reconstruction via non-isotropic diffusion in Dubins/Reed-Shepp-like control systems. In Proceedings of the 53rd IEEE Conference on Decision and Control, Los Angeles, CA, USA, 15–17 December 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 4278–4283. [Google Scholar]
  48. Boscain, U.; Chertovskih, R.A.; Gauthier, J.P.; Remizov, A.O. Hypoelliptic diffusion and human vision: A semidiscrete new twist. SIAM J. Imaging Sci. 2014, 7, 669–695. [Google Scholar] [CrossRef]
  49. Sharma, U.; Duits, R. Left-invariant evolutions of wavelet transforms on the similitude group. Appl. Comput. Harmon. Anal. 2015, 39, 110–137. [Google Scholar] [CrossRef]
  50. Citti, G.; Franceschiello, B.; Sanguinetti, G.; Sarti, A. Sub-Riemannian mean curvature flow for image processing. SIAM J. Imaging Sci. 2016, 9, 212–237. [Google Scholar] [CrossRef] [Green Version]
  51. Prandi, D.; Gauthier, J.-P. A Semidiscrete Version of the Citti-Petitot-Sarti Model as a Plausible Model for Anthropomorphic Image Reconstruction and Pattern Recognition; Springer Briefs in Mathematics; Springer: Cham, Switzerland, 2018. [Google Scholar] [CrossRef] [Green Version]
  52. ter Haar Romeny, B.M.; Bekkers, E.J.; Zhang, J.; Abbasi-Sureshjani, S.; Huang, F.; Duits, R.; Dashtbozorg, B.; Berendschot, T.T.J.M.; Smit-Ockeloen, I.; Eppenhof, K.A.J.; et al. Brain-inspired algorithms for retinal image analysis. Mach. Vis. Appl. 2016, 27, 1117–1135. [Google Scholar] [CrossRef] [Green Version]
  53. Bekkers, E.; Duits, R.; Berendschot, T.; Romeny, B.H. A multi-orientation analysis approach to retinal vessel tracking. J. Math. Imaging Vis. 2014, 49, 583–610. [Google Scholar] [CrossRef] [Green Version]
  54. Bosking, W.H.; Zhang, Y.; Schofield, B.; Fitzpatrick, D. Orientation selectivity and the arrangement of horizontal connections in tree shrew striate cortex. J. Neurosci. 1997, 17, 2112–2127. [Google Scholar] [CrossRef] [PubMed]
  55. Baspinar, E.; Sarti, A.; Citti, G. A sub-Riemannian model of the visual cortex with frequency and phase. J. Math. Neurosci. 2020, 10, 1–31. [Google Scholar] [CrossRef] [PubMed]
  56. Baspinar, E. Minimal Surfaces in Sub-Riemannian Structures and Functional Geometry of the Visual Cortex. Dissertation Thesis, Alma Mater Studiorum Università di Bologna, Bologna, Italy, 2018. [Google Scholar]
  57. Daugman, J.G. Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters. JOSA A 1985, 2, 1160–1169. [Google Scholar] [CrossRef]
  58. Bekkers, E.J.; Lafarge, M.W.; Veta, M.; Eppenhof, K.A.J.; Pluim, J.P.W.; Duits, R. Roto-translation covariant convolutional networks for medical image analysis. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2018; pp. 440–448. [Google Scholar]
  59. Duits, R.; Duits, M.; van Almsick, M.; Romeny, B.H. Invertible orientation scores as an application of generalized wavelet theory. Pattern Recognit. Image Anal. 2007, 17, 42–75. [Google Scholar] [CrossRef] [Green Version]
  60. Koenderink, J.J.; van Doorn, A.J. Representation of local geometry in the visual system. Biol. Cybern. 1987, 55, 367–375. [Google Scholar] [CrossRef] [PubMed]
  61. Lindeberg, T. A computational theory of visual receptive fields. Biol. Cybern. 2013, 107, 589–635. [Google Scholar] [CrossRef]
  62. Hörmander, L. Hypoelliptic second order differential equations. Acta Math. 1967, 119, 147–171. [Google Scholar] [CrossRef]
  63. Rashevskii, P.K. Any two points of a totally nonholonomic space may be connected by an admissible line. Uch. Zap. Ped. Inst. Liebknechta Ser. Phys. Math. 1938, 2, 83–94. [Google Scholar]
  64. Chow, W.-L. Über systeme von linearen partiellen differentialgleichungen erster ordnung. Math. Ann. 1940, 117, 98–105. [Google Scholar] [CrossRef]
  65. Agrachev, A.; Barilari, D.; Boscain, U. A Comprehensive Introduction to Sub-Riemannian Geometry; Cambridge University Press: Cambridge, UK, 2019; Volume 181. [Google Scholar]
  66. Unser, M. Splines: A perfect fit for signal and image processing. IEEE Signal Process. Mag. 1999, 16, 22–38. [Google Scholar] [CrossRef] [Green Version]
  67. Franken, E.M. Enhancement of Crossing Elongated Structures in Images; Eindhoven University of Technology: Eindhoven, The Netherlands, 2008. [Google Scholar]
  68. Kimmel, R.; Malladi, R.; Sochen, N. Images as embedded maps and minimal surfaces: Movies, color, texture, and volumetric medical images. Int. J. Comput. Vis. 2000, 39, 111–129. [Google Scholar] [CrossRef]
  69. Boscain, U.V.; Chertovskih, R.; Gauthier, J.P.; Prandi, D.; Remizov, A. Highly corrupted image inpainting through hypoelliptic diffusion. J. Math. Imaging Vis. 2018, 60, 1231–1245. [Google Scholar] [CrossRef] [Green Version]
  70. Duits, R.; Almsick, M.V. The explicit solutions of linear left-invariant second order stochastic evolution equations on the 2d euclidean motion group. Q. Appl. Math. 2008, 66, 27–67. [Google Scholar] [CrossRef] [Green Version]
  71. Zhang, J.; Duits, R.; Sanguinetti, G.; ter Haar Romeny, B.M. Numerical approaches for linear left-invariant diffusions on se (2), their comparison to exact solutions, and their applications in retinal imaging. Numer. Math. Theory Methods Appl. 2016, 9, 1–50. [Google Scholar] [CrossRef] [Green Version]
Figure 1. An example of the law of good continuity. We capture the curve on the bottom left as the curve underlying the aligned fragments on the top left; we do not capture any curve underlying the fragments on the top right due to the abruptly changing orientation angles of the fragments.
Figure 1. An example of the law of good continuity. We capture the curve on the bottom left as the curve underlying the aligned fragments on the top left; we do not capture any curve underlying the fragments on the top right due to the abruptly changing orientation angles of the fragments.
Jimaging 07 00271 g001
Figure 2. Two experimental settings from Field, Hayes and Hess [7]. A stimuli with aligned patches which we capture (left) and a stimuli plus the background with randomly oriented patches (right) are shown. Abrupt changes in the fragment orientations make it difficult to detect the aligned pattern in the bottom row.
Figure 2. Two experimental settings from Field, Hayes and Hess [7]. A stimuli with aligned patches which we capture (left) and a stimuli plus the background with randomly oriented patches (right) are shown. Abrupt changes in the fragment orientations make it difficult to detect the aligned pattern in the bottom row.
Jimaging 07 00271 g002
Figure 3. Top: association fields aligned with a horizontal patch as shown by Field, Hayes and Hess [7]. Bottom: solid curves represent the association fields between strongly associated fragments, and the dashed ones imply no fields between weakly associated fragments.
Figure 3. Top: association fields aligned with a horizontal patch as shown by Field, Hayes and Hess [7]. Bottom: solid curves represent the association fields between strongly associated fragments, and the dashed ones imply no fields between weakly associated fragments.
Jimaging 07 00271 g003
Figure 4. Left: Kanizsa triangle. There is no direct stimulus, yet we perceive a white triangle on top of the rest (modal completion). We perceive another triangular whose border is marked by the black lines on the bottom layer (amodal completion). Right: Ehrenstein illusion. We perceive a white disk around the center despite the absence of a direct stimulus (modal completion). We recognize that each vertical, horizontal or diagonal line fragment comprises a whole line which is occluded by the white disk (amodal completion).
Figure 4. Left: Kanizsa triangle. There is no direct stimulus, yet we perceive a white triangle on top of the rest (modal completion). We perceive another triangular whose border is marked by the black lines on the bottom layer (amodal completion). Right: Ehrenstein illusion. We perceive a white disk around the center despite the absence of a direct stimulus (modal completion). We recognize that each vertical, horizontal or diagonal line fragment comprises a whole line which is occluded by the white disk (amodal completion).
Jimaging 07 00271 g004
Figure 5. Left: real association fields. Right: projections of SE ( 2 ) horizontal integral curves. Figures are adapted from [7,33].
Figure 5. Left: real association fields. Right: projections of SE ( 2 ) horizontal integral curves. Figures are adapted from [7,33].
Jimaging 07 00271 g005
Figure 6. Two Gabor functions with low (left column) and high (right column) spatial frequencies. Top row: even (real) component of the Gabor functions. Bottom row: odd (imaginary) components.
Figure 6. Two Gabor functions with low (left column) and high (right column) spatial frequencies. Top row: even (real) component of the Gabor functions. Bottom row: odd (imaginary) components.
Jimaging 07 00271 g006
Figure 7. A horizontal integral curve along the vector field X 1 + X 2 . It represents an orientation fiber once f and ϕ are fixed. The tangent planes spanned by X 1 , X 2 (left) and X 3 , X 4 (right) are shown at six points on the curve.
Figure 7. A horizontal integral curve along the vector field X 1 + X 2 . It represents an orientation fiber once f and ϕ are fixed. The tangent planes spanned by X 1 , X 2 (left) and X 3 , X 4 (right) are shown at six points on the curve.
Jimaging 07 00271 g007
Figure 8. Horizontal integral curve fans corresponding to X 1 + c 2 X 2 (left) and X 3 + c 4 X 4 (right) where c 2 and c 4 are varied.
Figure 8. Horizontal integral curve fans corresponding to X 1 + c 2 X 2 (left) and X 3 + c 4 X 4 (right) where c 2 and c 4 are varied.
Jimaging 07 00271 g008
Figure 9. Illustration of the vectors e ξ k and e η k at ( 0 , 0 ) with Δ x = Δ y = 1 . The figure was modified and adapted from [44].
Figure 9. Illustration of the vectors e ξ k and e η k at ( 0 , 0 ) with Δ x = Δ y = 1 . The figure was modified and adapted from [44].
Jimaging 07 00271 g009
Figure 10. Completion of arcs with sinusoidal pattern. Left: original image. Middle left: corrupted image with occluding arcs. Middle right: completed image via the approximate method. Right: completion via the exact method.
Figure 10. Completion of arcs with sinusoidal pattern. Left: original image. Middle left: corrupted image with occluding arcs. Middle right: completed image via the approximate method. Right: completion via the exact method.
Jimaging 07 00271 g010
Figure 11. Completion of an occluded real texture image by two different arc patterns on the top and bottom rows. Left: Original image. Middle left: Image with occluding arcs. Middle right: Completed image via the approximate framework. Right: Completed image via the exact framework.
Figure 11. Completion of an occluded real texture image by two different arc patterns on the top and bottom rows. Left: Original image. Middle left: Image with occluding arcs. Middle right: Completed image via the approximate framework. Right: Completed image via the exact framework.
Jimaging 07 00271 g011
Figure 12. Completion of a real texture image occluded by two different line patterns on the top and bottom rows. Left: original image. Middle left: image with occluding vertical and horizontal lines. Middle right: completed image via the approximate framework. Right: completed image via the exact framework.
Figure 12. Completion of a real texture image occluded by two different line patterns on the top and bottom rows. Left: original image. Middle left: image with occluding vertical and horizontal lines. Middle right: completed image via the approximate framework. Right: completed image via the exact framework.
Jimaging 07 00271 g012
Figure 13. Single-frequency completion associated with the bottom row of Figure 11.
Figure 13. Single-frequency completion associated with the bottom row of Figure 11.
Jimaging 07 00271 g013
Figure 14. Single-frequency completion associated with the bottom row of Figure 12.
Figure 14. Single-frequency completion associated with the bottom row of Figure 12.
Jimaging 07 00271 g014
Figure 15. Completion of an occluded real texture image by two different arc patterns on the top and bottom rows. Left: original image taken from [68]. Middle left: image with occluding arcs. Middle right: completed image via the approximate framework. Right: completed image via the exact framework.
Figure 15. Completion of an occluded real texture image by two different arc patterns on the top and bottom rows. Left: original image taken from [68]. Middle left: image with occluding arcs. Middle right: completed image via the approximate framework. Right: completed image via the exact framework.
Jimaging 07 00271 g015
Figure 16. Completion of an occluded real texture image by two different line patterns on the top and bottom rows. Left: original image taken from [68]. Middle left: image with occluding arcs. Middle right: completed image via the approximate framework. Right: completed image via the exact framework.
Figure 16. Completion of an occluded real texture image by two different line patterns on the top and bottom rows. Left: original image taken from [68]. Middle left: image with occluding arcs. Middle right: completed image via the approximate framework. Right: completed image via the exact framework.
Jimaging 07 00271 g016
Figure 17. Single-frequency completion associated with the bottom row of Figure 15.
Figure 17. Single-frequency completion associated with the bottom row of Figure 15.
Jimaging 07 00271 g017
Figure 18. Single-frequency completion associated with the bottom row of Figure 16.
Figure 18. Single-frequency completion associated with the bottom row of Figure 16.
Jimaging 07 00271 g018
Figure 19. Top left: test image from [69]. Top right: image to be completed. Bottom left: completion via our method. Bottom middle: completion via the method in [50]. Bottom right: completion via the method in [48].
Figure 19. Top left: test image from [69]. Top right: image to be completed. Bottom left: completion via our method. Bottom middle: completion via the method in [50]. Bottom right: completion via the method in [48].
Jimaging 07 00271 g019
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Baspinar, E. Multi-Frequency Image Completion via a Biologically-Inspired Sub-Riemannian Model with Frequency and Phase. J. Imaging 2021, 7, 271. https://doi.org/10.3390/jimaging7120271

AMA Style

Baspinar E. Multi-Frequency Image Completion via a Biologically-Inspired Sub-Riemannian Model with Frequency and Phase. Journal of Imaging. 2021; 7(12):271. https://doi.org/10.3390/jimaging7120271

Chicago/Turabian Style

Baspinar, Emre. 2021. "Multi-Frequency Image Completion via a Biologically-Inspired Sub-Riemannian Model with Frequency and Phase" Journal of Imaging 7, no. 12: 271. https://doi.org/10.3390/jimaging7120271

APA Style

Baspinar, E. (2021). Multi-Frequency Image Completion via a Biologically-Inspired Sub-Riemannian Model with Frequency and Phase. Journal of Imaging, 7(12), 271. https://doi.org/10.3390/jimaging7120271

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop