Next Article in Journal
Modeling of the Atomic Diffusion Coefficient in Nanostructured Materials
Next Article in Special Issue
Lyapunov Exponents of a Discontinuous 4D Hyperchaotic System of Integer or Fractional Order
Previous Article in Journal
Fluid-Fluid Interfaces of Multi-Component Mixtures in Local Equilibrium
Previous Article in Special Issue
2D Tsallis Entropy for Image Segmentation Based on Modified Chaotic Bat Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Fractional-Order Chaotic Phase Synchronization Model for Visual Selection and Shifting

1
College of Computer Science, Chongqing University, Chongqing 400044, China
2
Chongqing/MII Key Lab. of Computer Network and Communication Technology, Chongqing 400044, China
3
College of Mathematics and Information Engineering, Chongqing University of Education, Chongqing 400065, China
*
Author to whom correspondence should be addressed.
Entropy 2018, 20(4), 251; https://doi.org/10.3390/e20040251
Submission received: 3 March 2018 / Revised: 1 April 2018 / Accepted: 2 April 2018 / Published: 4 April 2018
(This article belongs to the Special Issue Research Frontier in Chaos Theory and Complex Networks)

Abstract

:
Visual information processing is one of the fields of cognitive informatics. In this paper, a two-layer fractional-order chaotic network, which can simulate the mechanism of visual selection and shifting, is established. Unlike other object selection models, the proposed model introduces control units to select object. The first chaotic network layer of the model is used to implement image segmentation. A control layer is added as the second layer, consisting of a central neuron, which controls object selection and shifting. To implement visual selection and shifting, a strategy is proposed that can achieve different subnets corresponding to the objects in the first layer synchronizing with the central neuron at different time. The central unit acting as the central nervous system synchronizes with different subnets (hybrid systems), implementing the mechanism of visual selection and shifting in the human system. The proposed model corresponds better with the human visual system than the typical model of visual information encoding and transmission and provides new possibilities for further analysis of the mechanisms of the human cognitive system. The reasonability of the proposed model is verified by experiments using artificial and natural images.

1. Introduction

Cognitive informatics is an interdisciplinary research area of information sciences and cognitive sciences, which studies the internal information processing mechanisms and processes of natural intelligence in human brains and minds [1]. Visual information processing in the brain is the foundation of cognitive informatics [2]. The human visual system can select salient objects from a complex background and shift attention between different significant objects. Visual selection and shifting is a basic activity of the human visual system. The activity includes two stages: bottom-up and top-down. The first stage mainly depends on the unconscious behavior of retinal cells [3] and the second stage relies on control by the visual cortex. In these neural activities, synchronization plays an important role. Many physiological experiments have confirmed that synchronization occurs in the brain activity of humans, monkeys, mice and cats [4,5,6,7,8]. Synchronization is an important mechanism in image segmentation and feature binding. In the synchronization system, objects are represented by temporal correlation and spatially distributed neurons. Temporal correlation is encoded by the synchronization of oscillators that represent some features of an object and is termed “oscillatory correlation” [9]. Oscillatory correlation has been used to implement image segmentation and feature binding [10,11,12]. The oscillatory correlation mechanism is used to select objects and segment images in a network in which the oscillators oscillate in a limited cycle [13,14]. In other literature, chaotic synchronization is used to bind and segment images using complete and phase synchronization [15,16,17]. In real experiments, complete synchronization is rarely observed and the conditions required to achieve complete synchronization are strict. Phase synchronization is a basic mechanism for visual feature binding, as confirmed by multiple studies [18,19]. Therefore, we will use phase synchronization to construct a novel visual attention model.
There are two types of computer visual attention models. With the first type, location-based models, attention is only focused on a point or a small area and only one neuron is activated. With the second type, object-based models, visual attention can shift from one object to another. This type of model is supported by many behavioral and neurophysiological experiments [9,20,21]. The existing models discussed above are integer-order models and do not include any of the control units. The cognitive system, including the control unit, is more similar to the human cognitive system. The simulation system provides some help in the analysis of the human cognitive system. Some unique properties can be better represented by fractional-order models than by integer-order models [22] because a fractional-order model can exhibit more complex dynamic phenomena than an integer-order model [23]. The fractional time-delay systems exhibit superiority in their description of identification and the memory system [24,25]. Thus, in this paper, we depict visual attention using a fractional-order model with a control unit.
Phase synchronization in chaotic systems is that the phase difference of two systems is kept bounded over time while their amplitudes may be completely uncorrelated [26]. Phase synchronization in a chaotic system and network has been widely investigated [27,28,29]. Most of the literature focuses on phase synchronization among two or more coupled chaotic systems but a chaotic network synchronized with a single chaotic oscillator is still rarely reported. For synchronization among hybrid systems [30,31], we will propose a new synchronization strategy in which the chaotic network is considered a hybrid system composed of the linear superposition of phase synchronization oscillators. The hybrid system will achieve synchronization with the central neuron by active control.
For visual perceptions, we will propose a novel two-layer model. The first layer of the model, scene segmenting, is based on a model proposed by Zhao [17]. In Zhao et al.’s model, visual shift is based on the coupling strength changing over time. The control effect of a central neuron is not involved. In another study [32], the visual attention model is also two-layer, implementing visual selection and shift based on the resonance frequency. In our model, the second layer consists of a central unit, with which it implements visual selection and shift by chaotic control. The neuron in the second-layer can simulate the cerebral cortex neuron to control the first layer represented by the different objects by time division phase synchronization. The two-layer model designed in this way will be closer to the human visual system. This paper contributes three findings: Firstly, a synchronization strategy between a network and a single chaotic system is proposed; Secondly, chaotic control theory and fractional calculus are introduced into the model of visual attention; and finally, a fractional-order visual attention model with a control unit is established.
The remainder of this paper is organized as follows: in Section 2, the fractional-order chaotic system is presented and time division phase synchronization in the two-layer model is implemented; in Section 3, the visual attention model is described; the simulation results are shown in Section 4 and our conclusions are presented in Section 5.

2. Phase Synchronization of Fractional-Order Chaotic Systems

The role of synchronization in brain information processing has received support from neurobiologists [5]. The processing of oscillator synchronization is the foundation of temporal coding, called oscillatory correlation. The feature detectors are represented by oscillators and binding of the oscillators is represented by an assembly by synchrony. The oscillators representing different assemblies are desynchronous. Complete chaotic synchronization has been used to code in image segmentation. Phase synchronization has a less-restrictive condition than complete synchronization and is easier to implement. Thus, we use phase synchronization to realize the function of temporal coding.
Phase synchronization is when the phase difference of two systems is kept bounded while their amplitudes are uncorrelated. The unwrapped phase ψ is depicted as
ψ = Υ ( a r c t a n ( y x ) )
where x and y are the variables of the chaotic system; Υ is the unwrap operation and ψ is an increasing variable. For convenience, we also use phase to represent an unwrapped phase.

2.1. Fractional-Order Chaotic Network

A network coupled by fractional-order Rössler oscillators can achieve phase synchronization [27]. The coupled system can be modeled as
{ D α x j = o j y j z j + λ ( 2 x j x j 1 x j + 1 ) D α y j = ο j x j + a y j D α z j = b + z j ( x j c )
where λ and oj are the coupling strength and the system frequency, respectively; α = 0.9 is the fractional order of every oscillator and a = 0.48, b = 0.6, c = 6 are the system parameters. The state of synchronization is affected by the coupling strength λ. The phase standard deviations for different coupling strengths are shown as in Figure 1. As the coupling strength increases in absolute value, the phases of the coupled oscillators tend towards consistency. The phase standard deviation of the system is minimized at λ = −0.09.

2.2. Phase Synchronization between the Central System and the Chaotic Network

Time division synchronization is a means of communication that has been widely used [33,34]. Here, we use this method to control the central system and achieve phase synchronization with different subnets (groups) in the networks at different times. The response system (central system or central unit) is constructed as
{ D α x r = o r y r z r + u x r D α y r = o r x r + a y r + u y r D α z r = b + z r ( x r c ) + u z r
The driven system is a network composed of many oscillators (Rössler oscillators) as follows:
{ D α x j = o j y j z j + λ ( 2 x j x j 1 x j + 1 ) D α y j = o j x j + a y j D α z j = b + z j ( x j c )
The network is composed of two groups of oscillators whose frequencies are at different intervals. The frequencies of the two groups are within [1.01, 1.02] and [0.98, 0.99], respectively. The oscillators are synchronized with other neurons in the same group by the appropriate coupling strength. The central oscillator synchronizes firstly with the oscillators whose phase velocities are faster and then with the second group. The active control method described in Reference [29] is used to implement phase synchronization. If the central oscillator implements synchronization with every oscillator in the two groups, the control function u(x, y, z) will be extremely complicated. Liu et al. [30,31] have realized hybrid synchronization between heterogeneous chaotic systems. Inspired by this idea, we construct a hybrid system composed of a group of oscillators and carry out phase synchronization between the central system and the hybrid system. The constructed hybrid system is the mean of system variables of the oscillators in the network, depicted by Formula (5):
{ w x = x k + x k + 1 x k + n n w y = y k + y k + 1 x k + n n w z = z k + z k + 1 z k + n n
The error system is depicted as
{ e x = x r w x e y = y r w y e z = z r w z
Phase synchronization between the response system and the hybrid driven system is implemented by active control. Derivation of the control function u(x, y, z) can be seen presented in Appendix A.
The phase growth of the hybrid system is the same as the oscillators in the group, as shown in Figure 2. The hybrid system can represent the states of the oscillators in a group. This means that phase synchronization between a chaotic oscillator and a chaotic network can be translated to phase synchronization between a chaotic oscillator and a hybrid system. The response system is set up to synchronize with different groups of neurons at different times. When t = t1, the first hybrid system is controlled by the central unit to realize phase synchronization. Suppose the two systems maintain synchrony for ts. When t = t1 + ts + 0.01, the response system starts to synchronize with the second hybrid system. The phase time-division synchronization is depicted in Figure 2.
The phase growth of the central unit with two hybrid systems is shown in Figure 2a. The phase growth of the first hybrid system and the central system are the same from t = 30 to t = 50. This indicates that the two systems are in phase synchronization in this time span. The red line in Figure 2a stands for the phase of the central unit. The phase jumps at t = 30 and t = 50, as shown in Figure 2b. This means the phase of the central unit has a transition. After t = 80, the control unit does not control the second group of neurons and the phase growth of the central unit differs from the second group of neurons. Figure 3 shows that the central unit achieves phase synchronization with randomly selected oscillators in each group. The phase standard deviations of oscillators representing the same object are small, as shown in Figure 4. This also indicates that the oscillators representing the same object are synchronized. Figure 5 depicts the attractor of each of the three systems. Figure 4a,b shows that the phase standard deviations representing the same group obtained by our model changes are smaller than that obtained by the model proposed in [17]. It shows that the consistency of the oscillators in the same group is better. Comparing Figure 5a,b, it can be seen that the phase of the first hybrid system changes more quickly than the second one does and the central unit has an obvious transition. Thus, we have implemented phase synchronization between the single central unit and the network composed of coupled oscillators.

3. Model Description

In this section, the visual shift model is described by graph and text. The proposed model consists of two-layer Rössler oscillators and is shown in Figure 6. In the first layer, every oscillator couples with eight neighbors, except oscillators on the boundary. The mathematical model of the network is described by Equation (7). The main function of the first layer is to represent the formation of visual information. The structure of this layer is the same as the oscillator network in [17] and the way of connecting the network can encode topology and can separate different objects in the image. The number of oscillators is selected according to the size of the image. The second layer, composed of a central neuron oscillator, acts as the central neuron. The central neuron is coupled with different groups of neurons in the first layer at different times. The different groups of neurons represent different objects. The central neuron synchronizes with different groups of neurons by active control methods and realizes the communication with the imaging neurons. The model is described by Equations (7) and (8):
The first layer:
{ D α x j , k = o j , k y j , k z j , k + λ j , k + σ + x j , k + λ j , k σ x j , k D α y j , k = o j , k x j , k + a y j , k D α z j , k = b + z j , k ( x j , k c )
The second layer:
{ D α x r = o r y r z r + u x r D α y r = ο r x r + a y r + u y r D α z r = b + z r ( x r c ) + u z r
where (j, k) stands for the oscillator in the j-th row and k-th column of the two-dimensional topology, j ∈ [1, M], k ∈ [1, N]; M, N is the size of the input image; λ j , k + and λ j , k are the positive and negative coupling strength; and u x r ,   u y r ,   u z r are the active control functions of the second layer (response system). The details of the control functions can be seen in Appendix A. The positive and negative coupling terms are σ + x j , k and σ x j , k , with
σ ± x j , k = Δ j 1 , k 1 ; j , k ( x j 1 , k 1 x j , k ) + Δ j 1 , k ; j , k ( x j 1 , k x j , k )     + Δ j 1 , k + 1 ; j , k + Δ j , k 1 ; j , k ( x j , k 1 x j , k )     + ( x j 1 , k + 1 x j , k ) Δ j , k + 1 ; j , k ( x j , k + 1 x j , k )     + Δ j + 1 , k 1 ; j , k ( x j + 1 , k 1 x j , k ) + Δ j , k 1 ; j , k ( x j , k 1 x j , k )     + Δ j , k + 1 ; j , k ( x j , k + 1 x j , k )
where Δ j , k ; p , q = { 1 ,   if   oscillator   ( j , k )   is   coupled   to   ( p , q ) 0 ,   otherwise , o j , k is used to code the contrast of the pixel (j, k), because the contrast sensitivity is one of the information processing mechanisms in the human visual system. o j , k is determined by Δ o , which depends on the image and Cj,k, as follows:
o j , k = 1 0.5 Δ o + Δ o C j , k
Cj,k is the weighted average of the absolute difference:
C j , k = l w l | f j , k l f a v g l | l w l
where f j , k l , f a v g l are the values of pixel(j, k) and the average value for feature l, respectively. In this paper, feature refers to the gray intensity and the three R, G, B color channels. The four weights w l for the four features are set to be 0.5, 1/6, 1/6, 1/6. The positive and negative coupling strengths are described by Equations (11) and (12):
λ j , k + = λ m a x + × e x p ( ( 1 C j , k ) 2 2 σ 2 ) ,
λ j , k = λ m a x × ( 1 e x p ( ( 1 C j , k ) 2 2 σ 2 ) ) ,
where λ m a x + and λ m a x are the max positive and negative coupling strength. σ is set according to the image. In the second layer, o r is the mean value of all o j , k in the first layer. Other parameters are the same as for the first layer.
The block diagram of the proposed model is shown as in Figure 7. The blue blocks are the functional modules and the yellow blocks are the technical measures to implement the next functional modules.

4. Experiments

In this section, simulation results obtained by the proposed model are demonstrated. Artificial and natural images are used as the experimental example to evaluate the performance of the model. In the experiment, the max positive and negative coupling strengths are 0.05 and 0.02, respectively. In the experiments with natural images, a Gaussian filter is used for image denoising.
The first artificial image has three objects, as shown in Figure 8. To improve computing speed, the image is resized to 210 × 151 pixels. The oscillators representing the same object achieve phase synchronization due to the coupling strength and homologous frequency. On the other hand, the neurons representing different objects are desynchronized. Figure 9 shows the differences in phase among oscillators representing different objects from different perspectives. In the (i, j) axis, a random selection of oscillators is shown corresponding to the blue object, background, yellow object, the background and the green object one after the other from right to left, respectively. The oscillators representing yellow objects have the fastest phase growth, closely followed by the oscillators representing green objects, then oscillators representing blue objects and finally oscillators representing the background. The attractors of the hybrid systems corresponding to different objects are shown in Figure 10. The moving trajectories of oscillators for different objects are distinctive. In the experiment, the central unit and the first layer oscillators begin to oscillate at the same time. When the central unit finds the most salient object whose phase velocity is the fastest, it will couple with these oscillators and achieve synchronization with them by the active control method. The duration of synchronization is attention span, denoted by Ft (in this experiment Ft = 10). When the fixation time is over, attention will shift to the second salient object. The shift strategies are the same until the attention scans all objects in the scene. The attention shift process is shown as in Figure 11. When t = 20, the central neuron will start phase synchronization with the oscillators corresponding to the yellow object and will then maintain synchronization with them for Ft. At t = 30.01, the central unit will couple with the neurons representing the green object. During the entire process, the phase of the central neuron transitions four times. The control unit’s attractor is shown as in Figure 9e and the transition is obvious. The phase growth between different objects is distinctive. The phase standard deviation of each object is shown in Figure 12. The standard deviation is low between the same objects and high between different objects. In Figure 9, Figure 10 and Figure 11, it can be seen that for more salient objects, the phase growth is faster. This result of this experiment is consistent with the conclusion in [17]. It suggests that the fractional-order network layer can be used for image segmentation.
We compare our model with the model proposed in [17]. If the phases of oscillators representing same object are similar and the phases of oscillators representing different objects are very different, the segmentation results may be satisfactory. In this paper, we do three comparison tests to compare the segmentation results, as shown in Figure 12. In Figure 12, the phase standard deviations representing the same object obtained by our model changes less than that obtained by the model proposed in [17] and the phase standard deviations representing different objects obtained by our model grow more rapidly than that obtained by the model proposed in [17]. We also calculate the entropy of the segmentation results of Figure 8, which represents the probability of a particular information occurrence. If the entropy is smaller, the consistency of the segmentation results is better. Using our model and Zhao’s model, the segmentation results showed the local entropy to be 5.8986 and 6.4393, respectively. In general, for segmentation, the performance of our proposed model is better than that of the model in [17].
The proposed model can also be used in scenes with more objects, such as those shown in Figure 13, Figure 14, Figure 15, Figure 16 and Figure 17. There are six different phase types in Figure 14: yellow object, purple object, green object, azure object, red object and background. The attractors corresponding to different-colored objects are different. The central unit, represented by the pink line, transitions six times. The central unit’s attractor also reflects the phase transitions. The standard deviation is low between the same objects and high between different objects. The results confirm that the proposed model is reasonable.
Furthermore, we employ the natural images to test the proposed model. The natural images are shown in Figure 18 and Figure 19 and the experimental results are shown in Figure 20, Figure 21, Figure 22, Figure 23, Figure 24, Figure 25, Figure 26 and Figure 27. It can be seen that the phase growth is distinct for different objects. The control unit is also synchronized with the “attention object” during the control time. More detailed views are included as Figure 22b and Figure 26b, in which the phase transition is obvious. Note in Figure 23 and Figure 27 that the phase standard deviation between different objects is relatively small. This is because the contrasts between pixels derived from Equation (9) are small. The small contrasts lead to small phase differences between oscillators representing different objects. However, the central neuron can still synchronize with different oscillators at different times. The phase standard deviations representing the different objects obtained by our model grows more rapidly than that obtained by the model proposed in [17], as shown in Figure 23b Thus, the proposed model is also effective for visual selection and shifting in natural images.
We also selected the images from the BSDS500 (Berkley Segmentation Data Set) [35] to test the performance of the proposed model. The results for Figure 28 and Figure 29 are shown in Figure 30 and Figure 31. The central unit has 4, 3 times transitioning to synchronize with the “attention object” during the control time in Figure 30 and Figure 31. We also calculate the mean entropy of the segmented object. For Figure 28, the mean entropies obtained by our model and the model in [17] are 4.3606 and 4.39965, respectively. For Figure 29, the mean entropies obtained by the mentioned models are 5.1878 and 5.2262, respectively. For the two images, the mean entropy of segmented objects obtained by our model are small. This means that the performance of our proposed model is better than that of the model in [17] for image segmentation.

5. Conclusions

In this paper, a two-layer model is presented for attention shift. The first layer of the model is used for image segmentation. The second layer, acting as a control function of cognitive systems, implements attention shift by time division phase synchronization between the network and an oscillator. In this processing, a strategy is designed for time division phase synchronization between the network and an oscillator. The proposed model consists of visual and central nervous system cells. The introduction of a control function can make the proposed model closer to the human visual system than previous models and the fractional-order oscillators can better portray the mechanism of visual shift than the model with integer order ones.
The main current work presented in this paper is a novel model for visual selection and shifting. In future research, we will try to introduce a competition mechanism to the model to make the attention shift more consistent with visual search theory [36,37].

Acknowledgments

This work was supported by the Major Project of Fundamental Science and Frontier Technology Research of Chongqing CSTC (Grant No. cstc2015jcyjBX0124), National High-tech R&D Program of China (No. 2015AA015308), Chongqing People’s Livelihood Science and technology innovation projects (Grant No. cstc2015shmszx120010) and the Open Funds from Chongqing/MII Key Lab. of Computer Network and Communication Technology (No. CY-CNCL-2017-02).

Author Contributions

Ying Qi and Shangbo Zhou conceived and designed the experiments; Hongbin Tang and Xianzhong Xie performed the experiments; Xiaoran Lin analyzed the data and wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

To determine the control functions u, we subtract (4) from (3) and obtain
{ D α e x = o r y r z r ( o j y j z j + Γ j ) + ( o n y n z n + Γ n ) n + u x D α e y = o r x r a y r ( o j x j a y j ) + ( o n x n a y n ) n + u y D α e z = z r x r c z r ( z j x j c z j ) + ( z n x n c z n ) n + u z ,
where Γ j = λ ( 2 x j x j 1 x j + 1 ) . Every neuron is phase synchronized with the central unit. The Formula (7) can be rewritten as
{ D α e x = ( ( o r y r z r ( o j y j z j + Γ j ) ) + ( o r y r z r ( o n y n z n + Γ n ) ) ) / n + u x D α e y = ( ( o r x r a y r ( o j x j a y j ) ) + ( o r x r a y r ( o n x n a y n ) ) ) / n + u y D α e z = ( ( z r x r c z r ( z j x j c z j ) ) + ( z r x r c z r ( z n x n c z n ) ) ) / n + u z
{ u x = ( ( o r y j o j y j + Γ j ) + ( o r y n o n y n + Γ n ) ) / n + V x u y = ( o r x j o j x j ) +   ( o r x n o n x n ) n + V y u z = z r x r + z j x j + z n x n n + V z ,
the Formula (8) becomes
{ D α e x = o r e y e z + V x D α e y = o r e x a e y + V y D α e z = c e z + V z .
The linear functions can be written as
{ V x = o r e y + e z   + θ x e x V y = o r e x + a e y   + θ y e y V z =   c e z + θ z e z ,
where θ x ,   θ y ,   θ z are the eigenvalues of the system (A4). To satisfy the condition | arg ( θ x , y , z ) | > α π / 2 , θ x ,   θ y   and θ z are set to −1, −1, −1, respectively. Then, the error system (A4) is stable and the two systems can synchronize.
This completes the derivation of the control function.

References

  1. Wang, Y. Congnitive informatics: A New Transdiciplinary Research Field. Brain Mind 2003, 4, 115–127. [Google Scholar] [CrossRef]
  2. Wang, Y. A Cognitive Informatics Theory for Visual Information Processing. In Proceedings of the IEEE International Conference on Cognitive Informatics, Stanford, CA, USA, 14–16 August 2008; pp. 317–323. [Google Scholar]
  3. Wallace, R. Congnition and Biology: Perspectives from Information Theory. Cognit. Process 2014, 15, 1–12. [Google Scholar] [CrossRef] [PubMed]
  4. Fries, P.; Reynolds, J.H.; Rorie, A.E.; Desimone, R. Modulation of Oscillatory Neuronal Synchronization by Selective Visual Attention. Science 2001, 291, 1560. [Google Scholar] [CrossRef] [PubMed]
  5. Saleem, A.B.; Lien, A.D.; Krumin, M.; Haider, B.; Rosón, M.R.; Ayaz, A.; Reinhold, K.; Busse, L.; Carandini, M.; Harris, K.D. Subcortical Source and Modulation of the Narrowband Gamma Oscillation in Mouse Visual Cortex. Neuron 2017, 93, 315–322. [Google Scholar] [CrossRef] [PubMed]
  6. Gray, C.M.; Konig, P.; Engel, A.K.; Singer, W. Oscillatory Responses in Cat Visual Cortex Exhibit Inter-Columnar Synchronization Which Reflects Global Stimulus Properties. Nature 1989, 338, 334–337. [Google Scholar] [CrossRef] [PubMed]
  7. Quentin, R.; Chanes, L.; Vernet, M.; Valero-Cabré, A. Fronto-Parietal Anatomical Connections Influence the Modulation of Conscious Visual Perception by High-Beta Frontal Oscillatory Activity. Cereb. Cortex 2015, 25, 2095–2101. [Google Scholar] [CrossRef] [PubMed]
  8. Kim, Y.J.; Grabowecky, M.; Paller, K.A.; Muthu, K.; Suzuki, S. Attention Induces Synchronization-Based Response Gain in Steady-State Visual Evoked Potentials. Nat. Neurosci. 2007, 10, 117–125. [Google Scholar] [PubMed]
  9. Wang, D. The Time Dimension for Scene Analysis. IEEE Trans. Neural Netw. 2005, 16, 1401–1426. [Google Scholar] [CrossRef] [PubMed]
  10. Finger, H.; König, P. Phase Synchrony Facilitates Binding and Segmentation of Natural Images in a Coupled Neural Oscillator Network. Front. Comput. Neurosci. 2013, 7, 195. [Google Scholar] [CrossRef] [PubMed]
  11. Li, C.; Li, Y. Fast and Robust Image Segmentation by Small-World Neural Oscillator Networks. Cognit. Neurodyn. 2011, 5, 209–220. [Google Scholar] [CrossRef] [PubMed]
  12. Steyn, J.S.; Andras, P. Analysis of the Dynamics of Temporal Relationships of Neural Activities Using Optical Imaging Data. J. Comput. Neurosci. 2017, 42, 107–121. [Google Scholar] [CrossRef] [PubMed]
  13. Quiles, M.G.; Wang, D.; Zhao, L.; Romero, R.A.F.; Huang, D.-S. Selecting Salient Objects in Real Scenes: An Oscillatory Correlation Model. Neural Netw. 2011, 24, 54–64. [Google Scholar] [CrossRef] [PubMed]
  14. Chen, K.; Wang, D. A Dynamically Coupled Neural Oscillator Network for Image Segmentation. Neural Netw. 2002, 15, 423–439. [Google Scholar] [CrossRef]
  15. Zhao, L.; Breve, F.A. Chaotic Synchronization in 2d Lattice for Scene Segmentation. Neurocomputing 2008, 71, 2761–2771. [Google Scholar] [CrossRef]
  16. Zhao, L.; Cupertino, T.H.; Bertini, J.R., Jr. Chaotic Synchronization in General Network Topology for Scene Segmentation. Neurocomputing 2008, 71, 3360–3366. [Google Scholar] [CrossRef]
  17. Breve, F.A.; Zhao, L.; Quiles, M.G.; Macau, E.E.N. Chaotic Phase Synchronization and Desynchronization in an Oscillator Network for Object Selection. Neural Netw. 2009, 22, 728–737. [Google Scholar] [CrossRef] [PubMed]
  18. Doesburg, S.M.; Ward, L.M. Synchronization between Sources: Emerging Methods for Understanding Large-Scale Functional Networks in the Human Brain. In Coordinated Activity in the Brain: Measurements and Relevance to Brain Function and Behavior; Wennberg, V.R., Ed.; Jose luis Perez of Alcoa Inc.: New York, NY, USA, 2009; pp. 25–42. [Google Scholar]
  19. Varela, F.; Lachaux, J.P.; Rodriguez, E.; Martinerie, J. The Brainweb: Phase Synchronization and Large-Scale Integration. Nat. Rev. Neurosci. 2001, 2, 229–239. [Google Scholar] [CrossRef] [PubMed]
  20. Roelfsema, P.R.; Lamme, V.A.F.; Spekreijse, H. Object-Based Attention in the Primary Visual Cortex of the Macaque Monkey. Nature 1998, 395, 376–381. [Google Scholar] [CrossRef] [PubMed]
  21. Richard, A.M.; Lee, H.; Vecera, S.P. Attentional Spreading in Object-Based Attention. J. Exp. Psychol. Hum. Percept. Perform. 2008, 34, 842–853. [Google Scholar] [CrossRef] [PubMed]
  22. Li, L.; Wang, Z.; Lu, J.; Li, Y. Adaptive Synchronization of Fractional-order Complex-Valued Neural Networks with Discrete and Distributed Delays. Entropy 2018, 20, 124. [Google Scholar] [CrossRef]
  23. Siddique, U.; Hasan, O.; Tahar, S. Towards the Formalization of Fractional Calculus in Higher-Order Logic. In Proceedings of the International Conferences on Intelligent Computer Mathematics, Washington, DC, USA, 13–17 July 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 316–324. [Google Scholar]
  24. Alofi, A.; Cao, J.; Elaiw, A.; AL-Mazrooei, A. Delay-Dependent Stability Criterion of Caputo Fractional Neural Networks with Distributed Delay. Discret. Dyn. Nat. Soc 2014, 2014, 529358. [Google Scholar] [CrossRef]
  25. Wang, H.; Yu, Y.; Wen, G.; Zhang, S. Stability Analysis of Fractional-Order Neural Networks with Time Delay. Neural Process. Lett. 2015, 42, 479–500. [Google Scholar] [CrossRef]
  26. De Carlo, L.; Gentile, G.; Giuliani, A. Construction of the Lyapunov Spectrum in a Chaotic System Displaying Phase Synchronization. Math. Phys. Anal. Geom. 2016, 19, 10. [Google Scholar] [CrossRef]
  27. Chen, F.; Shao, X.; Sheng, L. Phase Synchronization in Small-World Network Composed of Fractional-Order Chaotic Oscillator. In Proceedings of the International Conference on Advances in Energy, Environment and Chemical Science, Guangzhou, China, 23–24 April 2016; CEC Press: London, UK, 2016; pp. 363–368. [Google Scholar]
  28. Odibat, Z. A Note on Phase Synchronization in Coupled Chaotic Fractional Order Systems. Nonlinear Anal. Real 2012, 13, 779–789. [Google Scholar] [CrossRef]
  29. Taghvafard, H.; Erjaee, G.H. Phase and Anti-Phase Synchronization of Fractional Order Chaotic Systems via Active Control. Commun. Nonlinear Sci. 2011, 16, 4079–4088. [Google Scholar] [CrossRef]
  30. Nian, F.; Liu, W. Hybrid Synchronization of Two Independent Chaotic Systems on Complex Network. Pram. J. Phys. 2016, 86, 1209–1222. [Google Scholar] [CrossRef]
  31. Nian, F.; Liu, W. Hybrid Synchronization of Heterogeneous Chaotic Systems on Complex Network. Chaos Solition Fract. 2016, 91, 554–561. [Google Scholar] [CrossRef]
  32. Qiao, Y.; Liu, X.; Miao, J.; Duan, L. A Neural Network Model for Visual Selection and Shifting. J. Integr. Neurosci. 2016, 15, 321–335. [Google Scholar] [CrossRef] [PubMed]
  33. Vincent, U.E.; Saseyi, A.O.; McClintock, P.V.E. Multi-Switching Combination Synchronization of Chaotic Systems. Nonlinear Dyn. 2015, 80, 845–854. [Google Scholar] [CrossRef] [Green Version]
  34. Uçar, A.; Lonngren, K.E.; Bai, E.-W. Multi-Switching Synchronization of Chaotic Systems with Active Controllers. Chaos Solition Fract. 2008, 38, 254–262. [Google Scholar] [CrossRef]
  35. Arbelaez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour Detection and Hierarchical Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 898–916. [Google Scholar] [CrossRef] [PubMed]
  36. Kazanovich, Y.; Borisyuk, R. Reaction Times in Visual Search Can Be Explained by a Simple Model of Neural Synchronization. Neural Netw. 2017, 87, 1–7. [Google Scholar] [CrossRef] [PubMed]
  37. Hulleman, J.; Olivers, C.N.L. The Impending Demise of the Item in Visual Search. Behav. Brain Sci. 2017, 40, e132. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Phase standard deviation with different coupling strengths.
Figure 1. Phase standard deviation with different coupling strengths.
Entropy 20 00251 g001
Figure 2. Phase growth of the proposed fractional-order systems. (a) The central unit and two hybrid systems; (b) Detail view of phase jump in the visual shift.
Figure 2. Phase growth of the proposed fractional-order systems. (a) The central unit and two hybrid systems; (b) Detail view of phase jump in the visual shift.
Entropy 20 00251 g002
Figure 3. Phase growth of the central unit and randomly selected oscillators from each group.
Figure 3. Phase growth of the central unit and randomly selected oscillators from each group.
Entropy 20 00251 g003
Figure 4. Phase standard deviation of groups over time. (a) Our proposed model; (b) The segmentation model in [17].
Figure 4. Phase standard deviation of groups over time. (a) Our proposed model; (b) The segmentation model in [17].
Entropy 20 00251 g004
Figure 5. Attractors of different systems. (a) The second hybrid system; (b) The first hybrid system; (c) The control system.
Figure 5. Attractors of different systems. (a) The second hybrid system; (b) The first hybrid system; (c) The control system.
Entropy 20 00251 g005
Figure 6. Proposed two-layer model.
Figure 6. Proposed two-layer model.
Entropy 20 00251 g006
Figure 7. The block diagram of the proposed model.
Figure 7. The block diagram of the proposed model.
Entropy 20 00251 g007
Figure 8. Artificial image with three objects.
Figure 8. Artificial image with three objects.
Entropy 20 00251 g008
Figure 9. Different perspectives for phase diagrams. The left describes the entire phase change, and the right describes the final phase profile.
Figure 9. Different perspectives for phase diagrams. The left describes the entire phase change, and the right describes the final phase profile.
Entropy 20 00251 g009
Figure 10. Attractors corresponding to the objects in Figure 8 and the central unit. (a) Background; (b) Blue object; (c) Green object; (d) Yellow object; (e) Central unit.
Figure 10. Attractors corresponding to the objects in Figure 8 and the central unit. (a) Background; (b) Blue object; (c) Green object; (d) Yellow object; (e) Central unit.
Entropy 20 00251 g010
Figure 11. (a) Phase growth corresponding to the objects in Figure 8; (b) Detail view of phase jump in the visual shift.
Figure 11. (a) Phase growth corresponding to the objects in Figure 8; (b) Detail view of phase jump in the visual shift.
Entropy 20 00251 g011
Figure 12. Phase standard deviation of objects in Figure 8 over time. (a) Our proposed model; (b) The segmentation model in [17].
Figure 12. Phase standard deviation of objects in Figure 8 over time. (a) Our proposed model; (b) The segmentation model in [17].
Entropy 20 00251 g012
Figure 13. Artificial image with five objects.
Figure 13. Artificial image with five objects.
Entropy 20 00251 g013
Figure 14. Different perspectives for phase diagrams. The left describes the entire phase change, and the right describes the final phase profile.
Figure 14. Different perspectives for phase diagrams. The left describes the entire phase change, and the right describes the final phase profile.
Entropy 20 00251 g014
Figure 15. Attractors corresponding to the objects in Figure 13 and the central unit. (a) Background; (b) Green object; (c) Red object; (d) Purple object; (e) Yellow object; (f) Azure object; (g) Central unit.
Figure 15. Attractors corresponding to the objects in Figure 13 and the central unit. (a) Background; (b) Green object; (c) Red object; (d) Purple object; (e) Yellow object; (f) Azure object; (g) Central unit.
Entropy 20 00251 g015
Figure 16. (a) Phase growth corresponding to the objects in Figure 13; (b) Detail view of phase jump in the visual shift.
Figure 16. (a) Phase growth corresponding to the objects in Figure 13; (b) Detail view of phase jump in the visual shift.
Entropy 20 00251 g016
Figure 17. Phase standard deviation for objects for Figure 13.
Figure 17. Phase standard deviation for objects for Figure 13.
Entropy 20 00251 g017
Figure 18. Natural image: “flower”.
Figure 18. Natural image: “flower”.
Entropy 20 00251 g018
Figure 19. Natural image: “bird”.
Figure 19. Natural image: “bird”.
Entropy 20 00251 g019
Figure 20. Different perspectives for phase diagrams. The left describes the entire phase change, and the right describes the final phase profile.
Figure 20. Different perspectives for phase diagrams. The left describes the entire phase change, and the right describes the final phase profile.
Entropy 20 00251 g020
Figure 21. Attractors corresponding to the objects in Figure 18 and the central unit. (a) Flower; (b) Leaves; (c) Central unit.
Figure 21. Attractors corresponding to the objects in Figure 18 and the central unit. (a) Flower; (b) Leaves; (c) Central unit.
Entropy 20 00251 g021
Figure 22. (a) Phase growth corresponding to the objects in Figure 18; (b) Detail view of phase jump in the visual shift.
Figure 22. (a) Phase growth corresponding to the objects in Figure 18; (b) Detail view of phase jump in the visual shift.
Entropy 20 00251 g022
Figure 23. Phase standard deviation for objects in Figure 18 over time. (a) Our proposed model; (b) The segmentation model in [17].
Figure 23. Phase standard deviation for objects in Figure 18 over time. (a) Our proposed model; (b) The segmentation model in [17].
Entropy 20 00251 g023
Figure 24. Different perspectives for phase diagrams. The left describes the entire phase change, and the right describes the final phase profile.
Figure 24. Different perspectives for phase diagrams. The left describes the entire phase change, and the right describes the final phase profile.
Entropy 20 00251 g024
Figure 25. Attractors corresponding to the objects in Figure 19 and the central unit. (a) Bird; (b) Grass; (c) Central unit.
Figure 25. Attractors corresponding to the objects in Figure 19 and the central unit. (a) Bird; (b) Grass; (c) Central unit.
Entropy 20 00251 g025
Figure 26. (a) Phase growth corresponding to the objects in Figure 19; (b) Detail view of phase jump in the visual shift.
Figure 26. (a) Phase growth corresponding to the objects in Figure 19; (b) Detail view of phase jump in the visual shift.
Entropy 20 00251 g026
Figure 27. Phase standard deviation for objects in Figure 19.
Figure 27. Phase standard deviation for objects in Figure 19.
Entropy 20 00251 g027
Figure 28. “Rhinoceros” form BSDS500 dataset.
Figure 28. “Rhinoceros” form BSDS500 dataset.
Entropy 20 00251 g028
Figure 29. “Parachuting” form BSDS500 dataset.
Figure 29. “Parachuting” form BSDS500 dataset.
Entropy 20 00251 g029
Figure 30. (a) Phase growth corresponding to the objects in Figure 28; (b) Detail view of phase jump in the visual shift.
Figure 30. (a) Phase growth corresponding to the objects in Figure 28; (b) Detail view of phase jump in the visual shift.
Entropy 20 00251 g030
Figure 31. (a) Phase growth corresponding to the objects in Figure 29; (b) Detail view of phase jump in the visual shift.
Figure 31. (a) Phase growth corresponding to the objects in Figure 29; (b) Detail view of phase jump in the visual shift.
Entropy 20 00251 g031

Share and Cite

MDPI and ACS Style

Lin, X.; Zhou, S.; Tang, H.; Qi, Y.; Xie, X. A Novel Fractional-Order Chaotic Phase Synchronization Model for Visual Selection and Shifting. Entropy 2018, 20, 251. https://doi.org/10.3390/e20040251

AMA Style

Lin X, Zhou S, Tang H, Qi Y, Xie X. A Novel Fractional-Order Chaotic Phase Synchronization Model for Visual Selection and Shifting. Entropy. 2018; 20(4):251. https://doi.org/10.3390/e20040251

Chicago/Turabian Style

Lin, Xiaoran, Shangbo Zhou, Hongbin Tang, Ying Qi, and Xianzhong Xie. 2018. "A Novel Fractional-Order Chaotic Phase Synchronization Model for Visual Selection and Shifting" Entropy 20, no. 4: 251. https://doi.org/10.3390/e20040251

APA Style

Lin, X., Zhou, S., Tang, H., Qi, Y., & Xie, X. (2018). A Novel Fractional-Order Chaotic Phase Synchronization Model for Visual Selection and Shifting. Entropy, 20(4), 251. https://doi.org/10.3390/e20040251

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop