Next Article in Journal
Automated Detection of Sleep Apnea-Hypopnea Events Based on 60 GHz Frequency-Modulated Continuous-Wave Radar Using Convolutional Recurrent Neural Networks: A Preliminary Report of a Prospective Cohort Study
Next Article in Special Issue
High-Speed Extraction of Regions of Interest in Optical Camera Communication Enabled by Grid Virtual Division
Previous Article in Journal
Intelligent Reflecting Surface-Based Non-LOS Human Activity Recognition for Next-Generation 6G-Enabled Healthcare System
Previous Article in Special Issue
High-Accuracy Height-Independent 3D VLP Based on Received Signal Strength Ratio
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep-Learning-Based Adaptive Symbol Decision for Visual MIMO System with Variable Channel Modeling

Department of Electronic Engineering, Kookmin University, Seongbuk-gu, Seoul 136-702, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(19), 7176; https://doi.org/10.3390/s22197176
Submission received: 18 August 2022 / Revised: 14 September 2022 / Accepted: 15 September 2022 / Published: 21 September 2022
(This article belongs to the Collection Visible Light Communication (VLC))

Abstract

:
A channel modeling method and deep-learning-based symbol decision method are proposed to improve the performance of a visual MIMO system for communication between a variable-color LED array and camera. Although image processing algorithms using color clustering are available to correct distorted color information in a channel, color-similarity-based approaches are limited by real-world distortions; to overcome such limitations, symbol decision is defined as a multiclass classification problem. Further, to learn a robust classifier against channel distortion, a deep neural network learning technique is applied to adaptively determine symbols from channel distortion. The network designed herein comprises the channel identification and symbol decision modules; the channel identification module extracts a channel identification vector for symbol determination from an input image using a two-dimensional deep convolutional neural network (CNN); the symbol decision module then generates a feature map by combining the channel identification vector and information on adjacent symbols to determine the symbol via learning correlations between adjacent symbols using a one-dimensional CNN. The two modules are connected together and learned simultaneously in an end-to-end manner. We also propose a new channel modeling method that intuitively reflects real-world distortion factors rather than the conventional additive white Gaussian noise channel to efficiently train deep-learning networks. Lastly, in the proposed channel distortion environment, the proposed method shows performance improvement by an average of about 41.8% (up to about 54.8%) compared to the existing Euclidean distance method, and about 6.3% (up to about 9.2%) on average compared to the SVM method.

1. Introduction

Visible light communication (VLC) is a wireless technology that transmits information by modulating data through the fast on–off switching of a visible light source. As the information and communication technology (ICT) industry’s demand for high-speed wireless access increases with an increase in data, the competition for occupying low-frequency bands increased, and many studies on VLC were performed to address this problem [1,2,3,4]. In addition, high-power, light-emitting diodes (LEDs) have developed significantly, enabling high-data-rate VLC networks. As a result, VLC has been highlighted as a new paradigm of wireless communication for innovations of the future.
Since most of the currently commercialized VLC-based services are modeled on lighting infrastructure inside buildings, it is clear that there are limits to realizing a hyperconnected internet, which is the ultimate goal of the Internet of Things (IoT) technology [5,6]. For VLC-based IoT technologies to be applied in various fields, such as the automobile, entertainment, and advertising industries, more advanced forms of VLC with improved flexibility and scalability are needed. To overcome the limitations of existing VLC technologies, the generalized color modulation (GCM) technique that secures the flexibility of LED colors and visual multiple-input multiple-output (visual-MIMO) technology that suggests infrastructure scalability through communication between an LED array and camera were developed [7,8,9,10,11,12,13]. In the visual MIMO concept, multiple transmit elements of a light-emitting array (LEA) are used as transmitters to communicate to the individual pixel elements of the camera that act as multiple receive elements to create the visual MIMO channel [11].
Color-independent visual MIMO technology, which combines the GCM method with visual MIMO technology, is expected to be applicable in more diverse fields because it enables continuous communication regardless of changes to the color or brightness of the LEDs during communication [14,15,16]. However, analyzing a signal received through a camera image sensor is not an easy task in the real world; this is because currently available commercial image sensors are very sensitive, and even small changes in external factors such as natural/ambient light, or internal factors such as exposure and focus of the camera, can greatly distort the transmission signals.
In previous studies based on the GCM technique, the Euclidean distance method that calculates the color similarity between the transmitted and received symbols was mainly used to determine the received symbol [10]. The Euclidean distance method was most often used in early studies of the GCM technique, where a single photodiode (PD) was used as the receiver; on the other hand, in the color-independent visual MIMO system, wherein an image sensor is used as the receiver, a color clustering algorithm is applied by taking advantage of the multiple inputs that somewhat guarantee the performance of the Euclidean distance method [17,18]. However, the color clustering algorithm is a classical image processing algorithm that is too simple compared to modern image recognition and computer vision solutions that combine numerous machine-learning and deep-learning algorithms [19]. As the position of the transmitted symbol for calculating the color similarity is fixed, the Euclidean-distance-based symbol decision method cannot respond adaptively when the position of the received symbol deviates greatly from that of the transmitted symbol in situations where the channel distortions change.
Although machine-learning (deep-learning) algorithms in the image processing field have recently developed significantly, classical methods (e.g., color clustering, Euclidean distance) have been mainly used for visual MIMO systems that have to determine color symbols. These methods not only have poor performance, but also do not flexibly respond to channel distortion in various environments. For this reason, in order to improve the SER performance of existing symbol decision methods in visual MIMO, in this study, we propose a method to adaptively determine symbols without additional information, using a deep-learning neural network through the output information of the image sensor, which is the receiver of the visual MIMO system. The whole network to be learned is divided into two modules. The first module is a channel identification module (CIM) that identifies the channel environment from the input image, and the second module is a symbol decision module (SDM) that classifies each symbol using information from its adjacent symbols. These two modules are combined into a single network and learned in an end-to-end manner. In addition, this study presents a new channel modeling method that intuitively reflects the real-world channel distortion characteristics, so that the proposed deep-learning neural network can learn robustly against various real-world channel distortions.
The remainder of this paper is organized as follows. Section 2 presents the proposed deep-learning neural network architecture designed for the adaptive SDM and explains the network training. Section 3 presents the limitations of the existing channel modeling methods and describes the approach that reflects the real-world distortions when performing simulations and experiments. Section 4 compares the symbol error rate (SER) performances of the proposed and existing methods and examines the experimental results. Finally, Section 5 presents the conclusions of this work.

2. Deep-Learning Neural Networks for Adaptive Symbol Decision

2.1. Deep-Learning Neural Network Architecture

Figure 1 shows the overall structure of the deep-learning neural network designed in this study; the network uses the CIM to predict the channel environment from the input image, and the SDM classifies these symbols by learning the correlations between adjacent symbols. Both modules are designed based on the convolutional neural network (CNN) algorithm, which has been mainly used for object recognition, semantic segmentation, and image reconstruction [20,21,22,23,24,25]. These two modules are combined into a single network and are learned in an end-to-end manner.
In the symbol feature map of Figure 1, the bold represents the target symbol (color) to be determined and the different colors mean the different channel identification vectors from CIM.

2.2. Channel Identification Module

Figure 2 shows the detailed structure of the CIM; the CIM detects the LED array from the image acquired through the image sensor and uses only this image as the input. In a previous study on color-independent visual MIMO systems [26], a technique was proposed for detecting the LED array region in an image, and it was proven that robust and continuous tracking of the LED array is possible through the Kalman filter algorithm.
CIM does not analyze the temporal information of continuous images on the receiver side to identify the channel state, but rather makes use of the spatial information of images obtained through LED array detection at a specific time. The detailed procedure is as follows. The detected LED array region is down-sampled through a two-dimensional CNN, and the compressed information is represented as a vector. In the CIM, the compressed information (feature maps) is transformed to a flattened one-dimensional vector and transferred to the hidden layer through the fully connected layer. This is different from that delivered to the decoder through a fully connected neural network in existing CNN encoder–decoder frameworks. Finally, information is extracted as a short-length implied vector at the output layer, which will henceforth be referred to as the channel identification vector.
The CIM is learned by an implicit method, so that the extracted channel identification vector becomes a feature for minimizing the loss function value of the entire deep-learning model. The final output predicted by the deep-learning neural network in this study is a multi-symbol index, and the approach here is to adaptively solve the multi-symbol classification problem under various channel environments. That is, the CIM is trained to extract additional features that enable classification of multiple symbols robustly based on changes in the channel environment.

2.3. Symbol Decision Module

Figure 3 shows the detailed structure of the SDM; the SDM determines the symbol for each LED of the input LED array through signal analysis. First, the channel information of the RGB, HSV, and CIE 1931 xyY models for the corresponding LED position is extracted from the image sensor and combined with the channel identification vector to generate its symbol representation. In the color-independent visual MIMO system, since the LED array is received simultaneously in a single frame from the image sensor, the spatial correlations between adjacent symbols can be important feature information. Therefore, the information of the inter-symbol area is used to decide each symbol, as shown in Figure 4.
The purpose of the Euclidean distance method, which is mainly used as the symbol decision method in existing color-independent visual MIMO systems, is to distinguish the similarity (or difference) between the color of the symbol sent from the transmitter and that of the symbol received at the receiver; thus, the symbol decision method proposed herein determines the symbol with the highest probability by learning its correlations with the adjacent symbols, without calculating the similarities between the two colors in the transmission/reception process.
At the boundary of the LED array, the padding method shown in Figure 5 is used. Although zero padding that fills with zero values or copying the values at the boundary can be used, the degree of distortion received by the LED array from the channel in this study is learned through the symbol correlation information within the LED array. To allow this, the adjacent symbol area is determined by copying the symbol information from the opposite side of the boundary.
Finally, the feature map required to learn the SDM is formed by combining the representations (RGB, HSV, and CIE 1931 xyY model channel values with channel identification vector) of the symbols included in the adjacent symbol area. Figure 6 shows the symbol feature map generated by the proposed method. The bold part in the figure indicates a symbol to be determined and the corresponding channel information of the RGB, HSV, and CIE 1931 xyY models.
The symbol feature map is used as the input to a one-dimensional CNN based on each feature channel. The network in Figure 3 is a CNN with N filters of size 5 × 12 symbol feature maps and a kernel size of 3. When the symbol feature map passes through one filter, a vector of length 3 is generated; when four filters are used, a vector of length 12 is obtained at the output. Unlike the CNN of the CIM, the CNN of the SDM proposed in this study is not designed as a deep CNN. In [27], the validity of using each channel of the RGB, HSV, and CIE 1931 xyY models as symbol decision features is demonstrated through the SER performance comparison experiment according to the color model. In the experiments, the three color models show almost similar SER performances. In other words, a deep CNN structure is not required because the symbol feature map generated through the adjacent symbol domain can be considered as already feature engineered with domain expertise from the color-independent visual MIMO system and color modeling perspective. Instead, N one-dimensional convolutional filters are applied with different kernel sizes and concatenated to finally extract a one-dimensional vector. The extracted vector passes through a fully connected layer and a softmax function to produce the final symbol. The one-dimensional CNN of the SDM is an independent network that does not share parameters with the CNN of the CIM; however, in the process of learning the entire deep-learning neural network by gradient descent method, it plays an important role in connecting the error backpropagation to the CNN of the CIM. In other words, the CNNs of the two modules are organically coupled, so that the error of the SDM is transmitted to the CIM by error backpropagation, and the output of the CIM affects the final output performance of the SDM. Through this process, the SDM learns to determine symbols in various channel distortion environments.

3. Variable Channel Modeling

3.1. Limitations of Existing Channel Modeling Methods

In previous studies, simulations were performed by adding the values of random variables of a Gaussian distribution defined according to the variance size as noise to each channel of the RGB model [10]. Figure 7 shows the actually observed representative distortions of the real world on the CIE 1931 xy chromaticity coordinate system [17]. In the figure, the symbol ‘⊙’inside the triangle indicates the white color position in the color space, and the symbols ‘⊙’ at the vertices of the triangle indicate the positions of red, green, and blue colors. It is seen that all four types of received symbols are clustered around white. This is a case in which the image captured by the image sensor suddenly becomes bright or dark, and occurs when the saturation of the LED array part of the image is lowered. When the saturation is low, as the brightness increases, the color approaches white, and as the brightness decreases, the color approaches black.
In general, it is difficult to intuitively understand the high and low levels of brightness and saturation in images represented by RGB models. Hence, the HSV model color space, designed to easily analyze the brightness and saturation levels in images, is widely used [28]. Since this study proposes an adaptive symbol decision method based on channel state changes, a channel modeling technique that is more reasonable to the real world is needed to effectively reflect the real-world distortions in deep learning.

3.2. Proposed Channel Modeling Method

Figure 8 shows the real-world distortion factors that may occur during the transmission/reception process of the visual MIMO system by classifying them into stages. By classifying these factors of distortion that occur simultaneously in the communication process as shown in Figure 8, it is assumed that the distortion at each step can be limited as follows. Transmission distortion is a factor where the transmitter LED emits light of the modulated symbol color, and it is assumed that it mainly causes color distortion of the symbol, or a distortion phenomenon in which the signal strength of the symbol is attenuated. Next, it is assumed that ambient distortion is a factor that occurs when natural or ambient light is strongly irradiated; this mainly causes color changes to an image and color distortion of the symbols. Reception distortion is a factor that occurs when the camera settings, such as exposure and gain, are adjusted. Finally, additive white Gaussian noise (AWGN) is considered as a channel noise factor of the image sensor output in the RGB model.
We cannot say that these limited assumed distortions reflect all the real-world distortions. In reality, all of the distortions assumed above may occur at the same time during the transmission/reception process. However, since the distortion phenomena assumed above are the biggest challenges for successful commercialization of visual MIMO systems, it is absolutely necessary to reflect them in the channel modeling.
In this study, the channel state according to time that reflects the major distortion factors is modeled as Equations (1)–(3):
S r x = C t S t x
where C t · denotes a channel state function at time t and S t x and S r x denote the transmit and receive symbol vectors in the color space, respectively. The symbol vector S can be represented as Equation (2), according to the color space in which it is expressed:
S s p a c e = R , G , B T H , S , V T x , y , Y T
where R, G, and B represent the three channel values of the RGB color space; H, S, and V represent the three channel values of the HSV color space; x, y, and Y represent the three channel values of the CIE 1931 xyY color space. The three color spaces in Equation (2) are modeled to be mutually transformable based on the RGB model [28,29]. The formulas for conversion between these color spaces is detailed elsewhere [28,29].
The channel state function C t · can be recursively represented as in Equation (3), so that the distortion factor is linked according to the transmission/reception flow shown in Figure 8:
C t S t x = D R x D L i g h t A × D T x S t x + n a w g n
where D T x · is a transmittance distortion function, D L i g h t · is an ambient distortion function, D R x · is a reception distortion function, A is a weight by transmitter attenuation distortion, and n a w g n denotes the conventional AWGN for the RGB model.
As described above, the transmission and ambient distortions cause changes in the chromaticity of the symbol color, and the reception distortion can cause changes in the saturation and brightness of the symbol color. Accordingly, Equation (3) can be represented as Equation (4):
C t S t x = D H S V D C I E S t x + n a w g n
where n a w g n can be represented as in Equations (5) and (6).
n a w g n = n r , n g , n b T
n r ~ N 0 , σ r 2 ,   n g ~ N 0 , σ g 2 ,   n r ~ N 0 , σ r 2
In Equation (4), D C I E · is a function that distorts the chromaticity of the input signal, and is given by Equations (7)–(9).
D C I E S t x C I E = S t x C I E + n x y C I E
n x y C I E = n x , n y , 0 T
n x ~ N 0 , σ x 2 ,   n y ~ N 0 , σ y 2
D H S V · is a function that distorts the brightness and saturation of the input signal, and is expressed by Equations (10)–(12).
D H S V S t x H S V = S t x H S V + n x y H S V
n s v H S V = n s , n v , 0 T
n s ~ N 0 , σ s 2 ,   n v ~ N 0 , σ v 2
Therefore, the proposed channel modeling technique includes a total of seven variables ( n x , n y , n s , n v , n r , n g , n b ), and each can be variably controlled by a Gaussian distribution with a different variance ( σ 2 ) value. The transmission/reception process shown in Figure 8 is only one scenario for a new channel modeling, and Equation (13) is also just one method defined by this scenario. The key here is that more intuitive and detailed distortion variables should be reflected in the channel model through multiple color spaces, to account for various channel distortions. Therefore, the channel state function through the proposed channel modeling can be represented as a comprehensive expression, in which the total distortion is determined by the seven variables defined above.
C t S t x = S t x + f ( n r , n g , n b , n s , n v , n x , n y )

3.3. Channel Modeling Verification

To validate the proposed channel modeling method, we attempt to reproduce the real-world distortions by adjusting the seven variables in Equation (13). Figure 9 shows the differences between the symbol images obtained by adjusting the exposure of the camera in the real world. In the figure, the upper and lower rows are examples of blue and red symbols, respectively. The dotted circle in the figure, represent the shape of a light source (LED). With respect to a camera, exposure refers to the amount of light that enters the image sensor through the lens; a high exposure means that a large amount of light enters the image sensor and a low exposure indicates that a small amount of light enters the image sensor through the lens. Exposure is determined by the camera’s aperture (F) and shutter speed. The smaller the aperture value, the larger the opening of the aperture and amount of light received. The lower the shutter speed, the longer the shutter is open; the longer the image sensor is exposed to light, the greater the amount of light received.
Table 1 shows the observed information in the images of Figure 9 through the RGB, HSV, and CIE 1931 xyY models.
When the camera exposure is changed, the channel-changing pattern of the RGB model differs depending on the symbol color. However, in the case of the HSV and CIE 1931 models, it is seen that the channel changes have similar trends, regardless of the symbol colors. Thus, it can be considered that the HSV and CIE 1931 models are suitable for channel modeling. A simulation experiment of the transceiving symbols of a 4 × 4 LED array was performed to reproduce the real-world distortions using the proposed channel modeling method. If the seven variables of Equation (13) are set as shown in Table 2, and the coordinates of the symbols received for 100 frames (total of 1600 symbols) are plotted in the CIE 1931 xy chromaticity coordinate system, the results are as shown in Figure 10.
In Figure 10, when the values of the variable n s are less than −0.2, the results are shown in Figure 11; these results represent the case where the symbol saturation is lowered and it can be considered as distortion wherein the camera exposure is largely adjusted. Therefore, it is proven that the proposed channel modeling method correctly reflects the real-world distortion factors and reproduces specific distortion phenomena similarly through the parameter settings.

4. Simulation and Experimental Results

4.1. Experimental Environment

Table 3 shows the GCM information used in the simulations. In the simulations, the LED array of the transmitter was set to a size of 4 × 4, and the 16 symbols were arranged and transmitted in parallel. The constellation diagram comprised four types of symbols through 2-bit encoding modulation. Table 4 shows the color space information for the target colors and symbols used in the simulations.
For the training of the proposed deep-learning neural network, as shown in Table 5, a total of 21 random channel environments (ID 0–20) were defined using the channel modeled in Section 3.2, and a total of 4200 frames (67,200 symbol samples) were generated as training data via 200 frames in each environment. In addition, to increase the reliability of the experimental results, the same amount of verification data as the training data were generated in an environment that was slightly worse than the environment used to generate each step of the training data. Table 5 shows the training data used in the simulation experiment, and Table 6 shows the verification data.

4.2. SER Performance

The comparison target for the performance evaluation in this study is the Euclidean distance method used in existing color-independent visual MIMO systems [27]. Additionally, the performance is compared with that of the support vector machine (SVM) method [30]. The SVM, which is already proven in numerous applications requiring machine learning, is the supervised learning model with associated learning algorithm that analyzes data for classification and regression analysis. In the case of SVM as a comparison target, radial basis function (RBF) kernel was adopted, and the values of gamma and C parameters with the highest performance were used experimentally. Here, C is a hypermeter that is set before the training model and used to control error, and gamma is also a hypermeter that is set before the training model and used to give curvature weight of the decision boundary. As a performance comparison metric, the SER is used, and SER performance is calculated as the reciprocal of accuracy.
The performances of the Euclidean distance method, which is the existing symbol decision method, and the SVM method, which is a representative machine-learning algorithm, are compared. Then, we compare the SER performance of the proposed deep-learning-based adaptive symbol decision method with that of the SVM method, which shows the best performance before deep learning. Figure 12 shows the SER performance comparisons between the Euclidean distance and SVM, SDM, and CIM + SDM (proposed) methods. The vertical axis indicates the SER value, and the symbol decision performance increases as the SER decreases. The horizontal axis indicates the channel environment ID value specified in Table 6, and the larger the value, the more severe the transmission symbol distortion due to the channel environment. Inside the figure, “w/intersymbol” means that learning with the information of adjacent symbols is considered, while “w/o intersymbol” means that learning with the information of adjacent symbols is not considered.
As seen in Figure 12, in the Euclidean distance method, the symbol decision performance deteriorates rapidly if the channel condition deteriorates even a little. Through this, it is clearly seen that the Euclidean distance method has obvious limitations in the proposed new channel modeling environment. In the case of the SVM method, on average, the performance improves by about 33.4% (up to about 43.6%) compared to the Euclidean method when learning with the information of adjacent symbols proposed in this study. Through this, it is proven that learning the information of adjacent symbols together has a positive effect on the symbol decision performance in situations where the channel is greatly distorted. Next, we compare the SER performance of the SVM and proposed CNN-based SDM methods. Both methods learn the correlations between symbols by using them as information features of the adjacent symbols. It can be seen that the CNN-based SDM shows a better overall performance than the SVM method. Although there is no dramatic performance improvement, the performance is improved by about 2.6% (up to about 5.7%). Considering the total amount of verification data, about 3830 symbol errors are reduced compared to the SVM method. From this, we can conclude that the deep-learning-based method is more suitable for learning various channel distortion factors than the classical machine-learning method. Figure 12 also presents the SER performance of the deep-learning-based adaptive symbol decision method (SDM + CIM) proposed in this study.
When the CIM is combined with the SDM, the performance improves on average by about 3.7% (up to about 6.2%). Among the comparison groups shown in Figure 12, only the proposed method maintains the SER below 1% until the channel environment ID #6. Through this, it is proven that the proposed method adapts to the channel changes and correctly determines the symbol, even in an environment where the channel condition deteriorates rapidly. Overall, it can be seen that the performance of the proposed deep-learning-based adaptive symbol decision method is improved on average by about 41.8% (up to about 54.8%) compared to the existing Euclidean distance method, and by about 6.3% (up to 9.2%) compared to the SVM method in the entire channel state used for evaluation.

5. Conclusions

This study proposed an adaptive symbol decision method for a visual MIMO system to which a deep-learning neural network was applied. The proposed deep-learning neural network comprises the SDM and CIM, both of which use the CNN algorithm. The CIM is designed with a deep CNN structure, and the SDM is designed with a one-dimensional convolutional multiple-filter structure. In addition, unlike the existing symbol decision methods that calculate the color similarities between the transmitted and received symbols to distinguish the received symbols, the proposed SDM uses the information of the adjacent symbols as the feature map, and determines the symbols by learning the correlations with the adjacent symbols.
To prove the superiority of the proposed adaptive symbol decision method over extant methods, a new channel modeling technique of a multicolor model mixing method that reflects various real-world distortion factors is presented; based on this, we succeed in reproducing the distortions caused by natural and ambient light and camera exposure variations in the real world. In addition, by directly comparing the performances of the learning methods in the proposed channel modeling environment and in the existing AWGN channel modeling environment, the proposed method is shown to not only include the AWGN environment, but also reflect more realistic channel distortion factors than the AWGN.
Lastly, in the simulations and experiments performed in the proposed channel distortion environment, the proposed method shows performance improvement by an average of about 41.8% (up to about 54.8%) compared to the existing Euclidean distance method, and about 6.3% (up to about 9.2%) on average compared to the SVM method. Based on the presented SER performances, it is proven that the deep-learning-based adaptive symbol decision method proposed in this study can robustly maintain SER performance, even when the channel environment deteriorates rapidly.

Author Contributions

J.-E.K. generated the idea and primarily simulated it, and he wrote and organized the manuscript. T.-H.K. helped in algorithm development, and also organized the final simulation and results. K.-D.K., as corresponding author, initiated the idea and supervised the whole process of this research and wrote the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation (NRF) of Korea funded by the Ministry of Science and ICT (2022R1A5A7000765) and was also supported by Basic Science Research Program through the National Research Foundation (NRF) of Korea funded by the Ministry of Education (NRF-2022R1A2C2010298).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this research is available upon a valid request to any of the authors of this research paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pathak, P.H.; Feng, X.; Hu, P.; Mohapatra, P. Visible Light Communication, Networking, and Sensing: A Survey, Potential and Challenges. IEEE Commun. Surv. Tutor. 2015, 17, 2047–2077. [Google Scholar] [CrossRef]
  2. Sugiyama, H.; Haruyama, S.; Nakagawa, M. Experimental Investigation of Modulation Method for Visible-Light Communications. IEICE Trans. Commun. 2006, E89-B, 3393–3400. [Google Scholar] [CrossRef]
  3. Yokoi, A. Color multiplex coding for VLC. In IEEE P802.15 Working Group for Wireless Personal Area Networks (WPANs); Samsung Yokohama Research Institute: Yokohama, Japan, 2008. [Google Scholar]
  4. Cevik, T.; Yilmaz, S. An Overview of Visible Light Communication Systems. Int. J. Comput. Netw. Commun. 2015, 7, 139–150. [Google Scholar] [CrossRef]
  5. Katz, M.D.; O’Brien, D. Exploiting novel concepts for visible light communications: From light-based IoT to living surfaces. Optik 2019, 195, 163176. [Google Scholar] [CrossRef]
  6. Warmerdam, K.; Pandharipande, A.; Caicedo, D. Connectivity in IoT indoor lighting systems with visible light communications. In Proceedings of the 2015 IEEE Online Conference on Green Communications (OnlineGreenComm), Piscataway, NJ, USA, 10–12 November 2015; pp. 47–52. [Google Scholar]
  7. Das, P.; Kim, B.-Y.; Park, Y.; Kim, K.-D. Color-independent VLC based on a color space without sending target color information. Opt. Commun. 2013, 286, 69–73. [Google Scholar] [CrossRef]
  8. Yuan, W.; Dana, K.; Varga, M.; Ashok, A.; Gruteser, M.; Mandayam, N. Computer vision methods for visual MIMO optical system. In Proceedings of the CVPR 2011 Workshops, Colorado Springs, CO, USA, 20–25 June 2011; pp. 37–43. [Google Scholar]
  9. Das, P.; Park, Y.; Kim, K.-D. Performance analysis of color-independent visible light communication using a color-space-based constellation diagram and modulation scheme. Wirel. Pers. Commun. 2014, 74, 665–682. [Google Scholar] [CrossRef]
  10. Das, P.; Park, Y.; Kim, K.-D. Performance improvement of color space based VLC modulation schemes under color and intensity variation. Opt. Commun. 2013, 303, 1–7. [Google Scholar] [CrossRef]
  11. Ashokz, A.; Gruteserz, M.; Mandayamz, N.; Dana, K. Characterizing Multiplexing and Diversity in Visual MIMO. In Proceedings of the 2011 45th Annual Conference on Information Sciences and Systems (CISS), Baltimore, MD, USA, 23–25 March 2011; pp. 1–6. [Google Scholar]
  12. Hu, W.; Gu, H.; Pu, Q. Lightsync: Unsynchronizedvisual communication over screen-camera links. In Proceedings of the 19th Annual International Conference on Mobile Computing Networking, MobiCom’13, New York, NY, USA, 30 September–4 October 2013; pp. 15–26. [Google Scholar]
  13. Hu, W.; Mao, J.; Huang, Z.; Xue, Y.; She, J.; Bian, K.; Shen, G. Strata: Layered coding for scalablevisual communication. In Proceedings of the 20th Annual International Conference on Mobile Computing and Networking, MobiCom’14, New York, NY, USA, 7–11 September 2014; pp. 79–90. [Google Scholar]
  14. Kim, J.-E.; Kim, J.-W.; Park, Y.; Kim, K.-D. Color-Space-Based Visual-MIMO for V2X Communication. Sensors 2016, 16, 591. [Google Scholar] [CrossRef] [PubMed]
  15. Kwon, T.-H.; Kim, J.-E.; Kim, Y.-H.; Kim, K.-D. Color-Independent Visible Light Communications Based on Color Space: State of the Art and Potentials. Electronics 2018, 7, 190. [Google Scholar] [CrossRef]
  16. Kim, J.-E.; Kim, J.-W.; Park, Y.; Kim, K.-D. Applicability of color-independent visual-MIMO for V2X communication. In Proceedings of the 2015 Seventh International Conference on Ubiquitous and Future Networks (ICUFN), Sapporo, Japan, 7–10 July 2015. [Google Scholar]
  17. Kim, J.-E.; Kim, K.-D. Symbol decision method of color-independent visual-MIMO system using a dynamic palette. In Proceedings of the 2017 23rd Asia-Pacific Conference on Communications (APCC), Perth, Australia, 11–13 December 2017; pp. 1–4. [Google Scholar]
  18. Aldhahi, W.; Kim, J.-E.; Kim, K.-D. Performance Improvement of Color-Independent Visual-MIMO System Using Color Quantization Based Palette. JKICS 2017, 42, 2418–2424. [Google Scholar]
  19. Dollar, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian Detection: A Benchmark. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  20. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  21. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  22. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  23. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  24. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  25. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  26. Kim, J.-E.; Kim, J.-W.; Kim, K.-D. LEA Detection and Tracking Method for Color-Independent Visual-MIMO. Sensors 2016, 16, 1027. [Google Scholar] [CrossRef] [PubMed]
  27. Kim, J.-E. The Performance Analysis of Color Independent Visual-MIMO based on Color Space. Master’s Thesis, Kookmin University, Seoul, Korea, 2015. [Google Scholar]
  28. Joblove, G.H.; Greenberg, D. Color spaces for computer graphics. ACM SIGGRAPH Comput. Graph. 1978, 12, 20–25. [Google Scholar] [CrossRef]
  29. Berns, R.S. Principles of Color Technology; John Wiley&Sons: New York, NY, USA, 2000. [Google Scholar]
  30. Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Conceptual diagram of the deep-learning neural network in the proposed method.
Figure 1. Conceptual diagram of the deep-learning neural network in the proposed method.
Sensors 22 07176 g001
Figure 2. Structure of the channel identification module.
Figure 2. Structure of the channel identification module.
Sensors 22 07176 g002
Figure 3. Structure of the symbol decision module.
Figure 3. Structure of the symbol decision module.
Sensors 22 07176 g003
Figure 4. Adjacent symbol area used for symbol decision.
Figure 4. Adjacent symbol area used for symbol decision.
Sensors 22 07176 g004
Figure 5. Adjacent symbol area at the LED array boundary.
Figure 5. Adjacent symbol area at the LED array boundary.
Sensors 22 07176 g005
Figure 6. Symbol feature map generated using the adjacent symbol area.
Figure 6. Symbol feature map generated using the adjacent symbol area.
Sensors 22 07176 g006
Figure 7. Distortion of the symbol transmission/reception process as observed in the real world.
Figure 7. Distortion of the symbol transmission/reception process as observed in the real world.
Sensors 22 07176 g007
Figure 8. Real-world channel distortion factors in visual MIMO systems.
Figure 8. Real-world channel distortion factors in visual MIMO systems.
Sensors 22 07176 g008
Figure 9. Differences in symbol images according to camera parameter adjustment. (a) When exposure is appropriate; (b) when exposure is low; (c) when exposure is high; (d) when exposure is appropriate; (e) when exposure is low; (f) when exposure is high.
Figure 9. Differences in symbol images according to camera parameter adjustment. (a) When exposure is appropriate; (b) when exposure is low; (c) when exposure is high; (d) when exposure is appropriate; (e) when exposure is low; (f) when exposure is high.
Sensors 22 07176 g009
Figure 10. Symbol transmission/reception simulation results from channel modeling.
Figure 10. Symbol transmission/reception simulation results from channel modeling.
Sensors 22 07176 g010
Figure 11. Symbol transmission/reception simulation results from channel modeling. ( n s 0.2 ).
Figure 11. Symbol transmission/reception simulation results from channel modeling. ( n s 0.2 ).
Sensors 22 07176 g011
Figure 12. SER performance of the proposed adaptive symbol decision method.
Figure 12. SER performance of the proposed adaptive symbol decision method.
Sensors 22 07176 g012
Table 1. Color space information based on camera exposure changes.
Table 1. Color space information based on camera exposure changes.
Figure 9HSV ModelRGB ModelCIE 1931 Model
H (°)S (%)V (%)RGBxyY
Figure 9a2461001002602550.16910.00990.0120
Figure 9b24010023.500600.16700.00900.0004
Figure 9c1809.41002312552550.31510.33640.9656
Figure 9d34487.139.610113370.63530.23020.0243
Figure 9e35290.612.532370.69720.25370.0019
Figure 9f001002552552550.33330.33331.0000
Table 2. Channel modeling parameter settings.
Table 2. Channel modeling parameter settings.
N 0 ,   σ 2 n x n y n s n v n r n g n b
σ 0.010.010.20.20.010.010.01
Table 3. GCM information for simulations.
Table 3. GCM information for simulations.
ParameterValue
Color spaceCIE 1931
RGB modelCIE RGB
Reference whiteE
LED array size16 (4 × 4)
Number of constellation points4
Intensity (Y value)0.165
Three positions of RGB LEDs in CIE 1931 spaceR: (0.735, 0.265)
G: (0.274, 0.717)
B: (0.167, 0.009)
Table 4. Target color and symbol information.
Table 4. Target color and symbol information.
Target ColorSymbolxyRGB
xy
0.400.33s10.500.33 1689682
s20.40 0.4311811272
s30.30 0.33 97115117
s40.40 0.23 16895138
Table 5. Training data information.
Table 5. Training data information.
Channel DistortionChannel Environment ID
011920
Std. of x, y noise distribution0.00.020.380.40
Std. of S, V noise distribution0.00.020.380.40
Std. of AWGN0.00.020.380.40
Number of frames200200200200
Sample framesSensors 22 07176 i001
Sensors 22 07176 i002
Sensors 22 07176 i003
Sensors 22 07176 i004
Sensors 22 07176 i005
Sensors 22 07176 i006
Sensors 22 07176 i007
Sensors 22 07176 i008
Sensors 22 07176 i009
Sensors 22 07176 i010
Table 6. Verification data information.
Table 6. Verification data information.
Channel DistortionChannel Environment ID
011920
Std. of x, y noise distribution0.010.030.390.41
Std. of S, V noise distribution0.010.030.390.41
Std. of AWGN0.010.030.390.41
Number of frames200200200200
Sample framesSensors 22 07176 i011
Sensors 22 07176 i012
Sensors 22 07176 i013
Sensors 22 07176 i014
Sensors 22 07176 i015
Sensors 22 07176 i016
Sensors 22 07176 i017
Sensors 22 07176 i018
Sensors 22 07176 i019
Sensors 22 07176 i020
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, J.-E.; Kwon, T.-H.; Kim, K.-D. Deep-Learning-Based Adaptive Symbol Decision for Visual MIMO System with Variable Channel Modeling. Sensors 2022, 22, 7176. https://doi.org/10.3390/s22197176

AMA Style

Kim J-E, Kwon T-H, Kim K-D. Deep-Learning-Based Adaptive Symbol Decision for Visual MIMO System with Variable Channel Modeling. Sensors. 2022; 22(19):7176. https://doi.org/10.3390/s22197176

Chicago/Turabian Style

Kim, Jai-Eun, Tae-Ho Kwon, and Ki-Doo Kim. 2022. "Deep-Learning-Based Adaptive Symbol Decision for Visual MIMO System with Variable Channel Modeling" Sensors 22, no. 19: 7176. https://doi.org/10.3390/s22197176

APA Style

Kim, J. -E., Kwon, T. -H., & Kim, K. -D. (2022). Deep-Learning-Based Adaptive Symbol Decision for Visual MIMO System with Variable Channel Modeling. Sensors, 22(19), 7176. https://doi.org/10.3390/s22197176

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop