Next Article in Journal
Clinical Characterization of Oculomotricity in Children with and without Specific Learning Disorders
Previous Article in Journal
Cannabis: Neuropsychiatry and Its Effects on Brain and Behavior
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Accurate and Robust Method for Spike Sorting Based on Convolutional Neural Networks

1
School of Information Science and Engineering, Yanshan University, Qinhuangdao 066004, China
2
Hebei Key Laboratory of Information Transmission and Signal Processing, Yanshan University, Qinhuangdao 066004, China
3
State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing 100875, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2020, 10(11), 835; https://doi.org/10.3390/brainsci10110835
Submission received: 9 October 2020 / Revised: 29 October 2020 / Accepted: 9 November 2020 / Published: 11 November 2020
(This article belongs to the Section Neural Engineering, Neuroergonomics and Neurorobotics)

Abstract

:
In the fields of neuroscience and biomedical signal processing, spike sorting is a crucial step to extract the information of single neurons from extracellular recordings. In this paper, we propose a novel deep learning approach based on one-dimensional convolutional neural networks (1D-CNNs) to implement accurate and robust spike sorting. The results of the simulated data demonstrated that the clustering accuracy in most datasets was greater than 99%, despite the multiple levels of noise and various degrees of overlapped spikes. Moreover, the proposed method performed significantly better than the state-of-the-art method named “WMsorting” and a deep-learning-based multilayer perceptron (MLP) model. In addition, the experimental data recorded from the primary visual cortex of a macaque monkey were used to evaluate the proposed method in a practical application. It was shown that the method could successfully isolate most spikes of different neurons (ranging from two to five) by training the 1D-CNN model with a small number of manually labeled spikes. Considering the above, the deep learning method proposed in this paper is of great advantage for spike sorting with high accuracy and strong robustness. It lays the foundation for application in more challenging works, such as distinguishing overlapped spikes and the simultaneous sorting of multichannel recordings.

1. Introduction

Analyzing the electrophysiological activities of neurons is a basis for exploring brain functions. Neurons in the brain communicate with each other by propagating action potentials, also referred to as “spikes” [1,2,3]. Hence, unraveling the interactions between different neurons is a key step to advancing our understanding of brain science and neural engineering [4]. Prior to interpreting the information carried by the interactions, an essential operation is to separate the spikes of individual neurons picked up by an extracellular electrode [5]. That is, it is necessary to determine which spike corresponds to which neuron, because each electrode records the extracellular field in its vicinity and can detect the spikes emitted by several neurons adjacent to the electrode site [6,7,8,9]. This identification is universally termed spike sorting [10,11,12], which is conducted on the basis of the similarity of spike waveforms. The reason lies in the fact that each neuron fires spikes of a particular shape, depending on the morphology of its dendritic tree, its distance and orientation relative to the recording site, etc. [11,13].
In fact, there are usually four main steps in conventional approaches for spike sorting: (1) Bandpass-filtering (e.g., 300–3000 Hz) the recorded raw data to eliminate the interferences of high-frequency noise and low-frequency field potential; (2) detecting the spikes by determining an amplitude threshold [14] or utilizing other improved methods [15], such as wavelet transforms [16] and fuzzy decision [2]; (3) extracting discriminative features from the detected spikes, frequently using approaches such as principal component analysis (PCA) [17,18,19] and wavelet transform coefficients [14,20,21]; (4) grouping the points in feature space to obtain clusters associated with individual neurons. Many classical and advanced methods have been adopted for this purpose, such as superparamagnetic clustering (SPC) [14], k-means clustering [22], and a mixture of Gaussians [23,24].
In this paper, we focus on a clustering algorithm, i.e., how to correctly isolate spikes originating from different neurons. Compared with semi-automatic methods [25,26], a fully automated process [27,28] requires no human input, consequently reducing time consumption and subjective bias. However, it sacrifices accuracy because it may ignore the distribution of clusters when a subset of a cluster exceeds the implicit boundary [29]. Thus, it is indispensable to manually resort falsely classified spikes, which is time-consuming and labor-intensive. Moreover, with the development of modern microelectronics and electrophysiology, high-density microelectrode arrays containing hundreds of channels have been widely applied in neuronal population recordings [3,30,31,32]. As a result, it is possible to simultaneously record the firing activities of hundreds or even thousands of neurons [33], involving tremendous amounts of data. Thus, current methods require substantial manual intervention, especially in long-term experiments, which makes them slow and unreliable. Inevitably, false classifications affect the results in subsequent analysis. Therefore, it is urgent to develop an effective method to reduce the workload of manual classification while ensuring accuracy and robustness [34]. Here, in light of recent advances in deep learning techniques [1,35], we took the manually classified spikes in an initial interval of the experimental data as the “ground-truth” labels, which were used to train a convolutional neural network (CNN) [36]. Then, this CNN could be used to automatically cluster the remaining data, thereby drastically reducing the time needed for spike sorting with high accuracy.
As an important type of deep learning technique, the CNN has been successfully used in computer vision since the early 21st century [36]. It performs well not only in object detection and brain image analysis [37,38], but also in the medical research field [39,40,41]. In recent years, deep learning techniques have also been used in spike detection and spike sorting. Yang et al. introduced a deep learning algorithm, named the principal component analysis network (PCANet), to reduce the computational complexity of conventional spike sorting algorithms [42]. However, the method did not show better sorting performance on overlapped spikes. The SpikeDeeptector method employed a CNN to select meaningful channels for brain–computer interfaces [15]. It considered a batch of waveforms to construct a single feature vector, which were fed to a deep learning method. Unlike SpikeDeeptector, we used a CNN to distinguish the categories of spikes in this paper. Both simulated data and experimental data were employed to evaluate the performance of our proposed model. By using the “Wave_Clus” database [43], we compared our proposed one-dimensional (1D)-CNN model with a recently developed method named “WMsorting” [44] and a deep-learning-based multilayer perceptron (MLP) model [45]. “WMsorting” proposed a spike feature extraction method using wavelet packet decomposition and mutual information. It is a semi-supervised solution for the automation of a spike sorting framework. The MLP approach constituted a deep-learning-based spike sorting method on extracellular recordings from a single electrode, improving on PCA + K-means clustering. We also compared our 1D-CNN model to some traditional methods, such as a PCA-based feature extraction with fuzzy C-means (FCM) clustering, a correlation coefficient-based (CORR) method with fuzzy clustering, first and second derivative extrema (FSDE) of spike data combined with k-means clustering, and a fusion feature strategy along with a support vector machine (SVM). Furthermore, we used the experimental data recorded from the primary visual cortex of a macaque monkey to test our 1D-CNN model and “WMsorting.”

2. Materials and Methods

2.1. Architecture of the 1D-CNN Model

The standard architecture involved in a CNN includes the convolutional layer, rectified linear activation function, pooling layer, and fully connected layer [36]. Figure 1 illustrates the proposed CNN architecture working on a simulated database. The model contained four convolutional layers and two pooling layers, followed by a fully connected neural network and a softmax classifier. The CNN model used for the experimental data had a similar architecture. The only difference was the length of the input layer, whereby the former had a value of 64 and the latter had a value of 12. The details of the model are described as below.
Convolutional layer: The main building block of a CNN is the convolutional layer, which completes most of the computationally intensive lifting and extracts features from the input signals. In this paper, four convolutional layers were arranged into nine feature maps [36]. The original intention for this choice was because, when the number of convolutional layers was greater than four, the accuracy nearly stopped growing. For all convolutional layers, the stride was 1, the kernel size was 3, and the padding was the same. If the stride is too small, the extracted features are more comprehensive and do not miss too much information. Due to that, the stride was set to 1. Compared with using a larger convolution kernel, using smaller convolution kernels can obtain more feature information with a smaller amount of calculation. Thus, the kernel size of the convolutional layer was set to 3. We filled the edges of the input matrix with zero padding, allowing us to filter the edges of the input matrix. One of the advantages of zero padding was that it allowed us to control the size of the feature map.
Rectified linear activation function and regularization techniques: The rectified linear unit (Relu) layer is used for addressing the problem of optimization by mapping nonlinearity into the data [36]. The reason we chose Relu was that it could be used to alleviate the problem of gradient disappearance. It was designed as an activation function for layers 1, 2, 4, 6, 7, and 8. The softmax function was employed in layer 9 (last layer of the network). Using the softmax classifier, the prediction of a cluster to which the input data belong can be realized. The batch normalization, which kept the inputs closer to a normal distribution, was applied to the output of the convolutional layer. The dropout [46], another regularization technique, was applied to reduce overfitting by randomly setting the values of some input neurons to zero before the fully connected layers.
Pooling layer: The max pooling layer is implemented to reduce the computational complexity of conventional sorting algorithms. Max pooling compares every pair of numbers and outputs the bigger value. In this paper, two max pooling layers were used with a kernel size of 3. The stride was set to 1, and the filter convolved around different layers of the feature map by sliding one unit each time.
Fully connected layer: The output dimensions of the final connected layer depend on the number of classes. For an n-class (n kinds of spikes) problem, the output dimensions were set to n. There were three kinds of spikes in the simulated data, that is, the number of output dimensions was three. In the experimental data, we separately set-up the output dimensions for each channel, depending on the number of clusters determined by the spike sorting algorithm and human intervention.
As shown in Figure 1, the input layer (layer 0) with a resolution of 64 × 1 was convolved with 32 filters of kernel size 3 to form the first layer (layer 1). The second convolutional layer with a kernel size of 3 (64 filters) was applied to produce the second layer. A max pooling layer was employed with a kernel size of 2 (layer 3). A convolutional operation was administered on layer 3 (32 × 64) to form layer 4. The feature maps from layer 4 were once again pooled with a kernel size of 2 to construct the last max pooling operation (layer 5). Then, another round of convolution was employed with a kernel size of 3. The next three layers were the fully connected layers. The neurons of feature maps in layer 6 were fully connected to 300 neurons in layer 7. Layer 7 and layer 8 were respectively fully connected to 100 and n outputs in layers 8 and 9.

2.2. Simulated Database and Experimental Datasets

The simulated database created in “Wave_Clus” [14] has been widely used in the evaluation of several spike sorting algorithms [42,44,47]. In total, 594 different average spike waveforms compiled from real recordings were adopted as templates to construct simulated signals. Randomly selected spikes were superimposed at random times and amplitudes to mimic the background noise. Then, it was feasible to realize different signal-to-noise ratios (SNRs) by altering the ratio between the amplitudes of the noise and signals. More details related to the generation of the simulated database can be found in [43].
There were four big datasets used in our experiments, i.e., C_Easy1, C_Easy2, C_Difficult1, and C_Difficult2. C_Easy1 contained eight noise levels ranging from 0.05 to 0.4 with a step of 0.05. The other three datasets contained four levels ranging from 0.05 to 0.2. “Easy” and “difficult” were set to characterize the overlapping degrees between spikes. Spikes of 64 sampling points were extracted from the continuous data by using the ground-truth spike times denoted in each dataset. Obviously, these spikes could be treated as labeled samples, making them suitable for testing our supervised deep learning method. The sampling rate of the simulated database was 24 kHz. Detailed information of the database is provided in Figure 2.
The experimental datasets were obtained from public databases (Collaborative Research in Computational Neuroscience, CRCNS) [48,49]. The data were recorded from the primary visual cortex (V1) of a macaque monkey with 64-site multi-electrode arrays (eight penetrations, eight sites per penetration). The number of electrode contacts ranged from 65 to 128. After the spikes were isolated via superparamagnetic clustering, manual resorting was conducted to avoid possible errors. The sampling rate of the experimental datasets was 24.4 kHz. The dataset provided spikes of single neurons, including their waveforms and firing times. Here, we removed the channels with very few spikes in at least one cluster (the number of spikes was less than 0.5% of the total), because there were not enough labeled samples to train the 1D-CNN. Figure 3 lists the remaining channels with detailed cluster information, which were used to test the performance of our deep learning method.

2.3. Training and Testing

Six experiments were performed for each dataset. The proportion of data used for the training and testing procedures is illustrated in Figure 4. For each dataset, the training set separately held 5%, 10%, 20%, 30%, 40%, and 50% of the total data, and the corresponding testing set held the remaining 95%, 90%, 80%, 70%, 60%, and 50%, respectively. We separately used 50% of the total data in C_E1_015 and C_D1_015 as the training set and the remaining data as the validation set to adjust and evaluate our model. When our model was determined, other datasets were used to evaluate the performance, and they were divided into a training set and testing set.
To evaluate the performance of the model, accuracy, which was calculated as the percentage of correctly classified samples over all data, was utilized to compute the score of the entire classification, and it was used in the analysis of experimental data. In addition, considering the prevalent imbalance of data distribution in the experimental data, the macro-average F-measure (Macro_F) was also employed to describe the classification effects. Macro_F is more sensitive to the classification quality, and it is defined as follows [50]:
M a c r o _ F = 1 N i = 1 N 2 × P i × R i P i + R i
where P i = T P i / T P i + F P i   is the recall, R i = T P i / T P i + F N i is the precision, and N is the number of clusters. TP denotes true positives, FP denotes false positives, and FN denotes false negatives.

3. Results

3.1. Simulated Database

In this section, the classification accuracy of the proposed 1D-CNN model is compared with “WMsorting”, the deep-learning-based MLP model, and four traditional methods using the “Wave_Clus” database. The four traditional methods utilized PCA-based and correlation-based (CORR) feature extraction with FCM as the clustering method and two other public methods (fusion + SVM and FSDE + K-means). Details of these traditional methods can be found in [44]. The results for the data with different noise levels are shown in Table 1 and Table 2.
Although “WMsorting” showed relatively high accuracy on the C_Easy1, C_Easy2, and C_Difficult2 datasets, it was not robust. The accuracy of “WMsorting” on C_Difficult1_020 was 85.38% and 86.45% when the feature dimensions were 3 and 10, respectively, which were much lower than the results of other datasets. The feature dimension is an important parameter in sorting evaluation, and it is determined with reference to the best Micro_F (micro-average F-measure) [51] score of the PCA-based method. On the other hand, the accuracy of our proposed deep learning method on C_Difficult1_020 was above 98% except for experiment 1, whose accuracy was 95.16%. That is, even when we used only 5% of the spikes to train the 1D-CNN, we still obtained an accuracy improvement of 8.71%. In fact, when the number of labeled spikes increased, i.e., in the other five experiments, our results could enhance the clustering accuracy by more than 10% compared to “WMsorting.” In the other datasets, although the enhancement was not so significant, our deep learning method behaved better than “WMsorting.” Additionally, the improved accuracy is shown in the last column of Table 1, indicating that the best accuracy of our proposed model was relatively higher than the best accuracy of “WMsorting” on all datasets.
Compared to traditional methods, our proposed deep learning method had more obvious advantages. Traditional methods worsened with increasing noise level for all datasets, with an error rate as high as 46%, while our error rate was around 2%. For lower noise levels, our accuracy was still better than that of traditional methods.
The deep-learning-based MLP model used 10% of the data nearest to the cluster centers as the training data; thus, we used experiment 2 (10% of the total data as training set) in the last column of Table 2. As can be seen, the deep-learning-based MLP model did not perform well with a higher noise level, especially on the C_Difficult2_020 or C_Difficult2_015 datasets. The deep-learning-based MLP model’s accuracy on C_Difficult2_020 was 51.55%, which was 42.8% lower than ours.
As the number of spikes used in the training set gradually increased (from 5% to 50% of the total), the accuracy exhibited a slight tendency to increase (0.607% on average). It was also shown that 5% (170 spikes) of the data were enough to train the proposed 1D-CNN model in most cases. On the other hand, with the increase in noise level (from 5% to 20%), the accuracy had no significant reduction. Both factors confirmed that our proposed deep learning method is robust.
Furthermore, we took experiment 6 for two datasets (C_Easy1_noise015 and C_Difficult1_noise015) as an example to illustrate the results in more detail. First, we observed the accuracy of individual classes by constructing confusion matrices, as shown in Figure 5a,b. Each row in the confusion matrix represents a true label, while each column represents a predicted label. The bold percentages denote the proportion of correct true labels for each cluster. Clearly, there was only a small difference between the accuracy of individual classes and the overall accuracy, which is consistent with the balanced spike distribution of each cluster in the “Wave_Clus” database. That is, the proposed 1D-CNN model performed excellently for different clusters, indicating that it is an accurate and robust method for spike sorting.
Moreover, the accuracy for the C_Difficult1_noise015 dataset was clearly lower than that for C_Easy1_noise015. In order to explain the reason behind this, we plotted the correctly clustered spikes in Figure 6 and the falsely classified spikes in Figure 6. Obviously, the spike waveform shapes of the three clusters in C_Easy1_noise015 were quite different and relatively easy to distinguish. On the other hand, the spike waveform shapes in C_Difficult1_noise015 were similar to each other and difficult to distinguish.
Figure 7 shows all the falsely classified spikes and a small subset of the correctly classified spikes in the two datasets. As illustrated in Figure 7a,b, only one spike in “Cluster 2” and three spikes in “Cluster 3” in C_Easy1_noise015 were falsely classified due to their very similar shapes. The spikes in “Cluster 1” in C_Easy1_noise015 were all correctly classified. On the other hand, there were more falsely classified spikes in C_Difficult1_015. Clearly, they were severely distorted by the noise and had similar shapes, which hindered their sorting into the corresponding clusters. However, compared to the accuracy of 91.18% obtained using the “WMsorting” method, our proposed deep learning approach shows a significant improvement with an accuracy of 99.42%.
In addition to clustering accuracy, the loss function is an evaluation index for the performance of deep learning methods. Cross-entropy is a frequently used function that measures the degree of difference between the actual outputs and expected outputs of a CNN model. In other words, a smaller cross-entropy loss function denotes the two outputs being closer. On the other hand, the loss function helps determine whether a model is overfitting or underfitting the data by comparing the training loss and testing loss. The training and testing cross-entropy losses on both datasets (experiment 6 for C_Easy1_noise015 and C_Difficult1_noise015) are plotted in Figure 8. In our proposed model, the training epoch was set to 50. Obviously, although there were occasional rises, the training loss and testing loss decreased along with the increase in the number of epochs. The testing loss of C_Easy1_015 decreased to 0.01708 at the 50th epoch, and the testing loss of C_Difficult1_015 was 0.05861 at the 50th epoch. These values were small enough to demonstrate the robustness of our proposed model. Furthermore, the small difference between the training loss and testing loss showed that our model neither overfitted nor underfitted the data.

3.2. Experimental Datasets

Similarly, six experiments with different proportions of training data were performed on the extracellular recording spikes to further demonstrate the effectiveness of our method. In Figure 9, the classification accuracies are presented according to the number of spike clusters (ranging from two to five). It can be seen that most of the results achieved accuracies higher than 96.5% regardless of the number of labels used to train the 1D-CNN. In fact, the average accuracy of all experiments was 96.53%, which is an acceptable level. However, it should also be noted that the accuracies of some channels were significantly lower than those of others. For example, in the 94th channel, the classification accuracy of experiment 1 (5% of the data used as the training set) was 89.40%. The reason for this is that the number of spikes in the training set was 218, which was not enough to effectively train the 1D-CNN model. When the number of training spikes increased, the classification accuracy improved, e.g., it was 91.80% in experiment 2 (10% of the data used as the training set) and 94.60% in experiment 3 (20% of the data used as the training set).
The above accuracies of each dataset were calculated as a function of the total spikes. As is known, there is always an imbalanced data distribution in experimental datasets; thus, Macro_F is an appropriate criterion to evaluate the classification quality. The results of this evaluation are given in Table 3. Taking the 98th channel as an example, the accuracy in experiment 1 was 95.44%, but the corresponding Macro_F score was 74.66%. Clearly, there was a great difference. The reason is that the minimum number of spikes in the three clusters was 103 and the corresponding number of spikes involved in the training set in experiment 1 was only 21. Obviously, this was not enough to adequately train the 1D-CNN to achieve a desirable output. It should also be noted that the Macro_F score increased to >90% when the number of spikes in the training set was >30% of the total number of spikes (i.e., experiments 4, 5, and 6). In fact, most Macro_F scores were above 95% in experiment 6. In the 84th and 116th channels, scores of 99% were achieved. Thus, as long as there are enough spikes for training the proposed 1D-CNN, it is very likely to obtain a close approximation of the ideal spike sorting results.
The experimental data were also analyzed using “WMsorting”, and the Macro_F scores are shown in Table 3. We separately carried out two experiments using “WMsorting” with three and ten feature dimensions. We used the best obtained accuracy in each channel from our 1D_CNN model and “WMsorting” to calculate the improvement. As can be seen, our 1D_CNN model provided a better result.

4. Discussion

In this paper, we designed a 1D-CNN model to improve the accuracy and robustness of spike sorting. The novel deep learning method exhibited a high sorting accuracy in the simulated database, even when the data suffered from severe noise. In the experimental datasets, our proposed model still performed well in classifying the spikes recorded by the extracellular electrodes in the primary visual cortex of a macaque monkey. Thus, it is reasonable to conclude that our proposed model showed a good performance in terms of accuracy and robustness.
There were two factors influencing the performance of the method, i.e., sampling points and labels. First, the number of sampling points in each spike had a significant impact on the accuracy. As shown in Table 1, the accuracies of C_Easy1_015 and C_Difficult1_015 in experiment 6 are 99.77% and 99.42%, respectively. However, when we downsampled the spike waveforms from 64 points to 16 points, the corresponding accuracies were reduced to 99.14% and 87.91%. Clearly, the reduction seen for the C_Difficult1_015 dataset was much greater than that seen for C_Easy1_015. This is because the spike waveforms of different clusters in C_Difficult1_015 were more similar to each other. The lower accuracies of the experimental datasets with 12 sampling points also verified this effect. That is, with fewer spike sampling points, it was more difficult for 1D-CNN to cluster the data correctly. A possible reason is that the 1D-CNN could not “learn” enough features from the limited spike sampling points. Secondly, as with other deep learning approaches [15,52,53], the number of labels in the training set was of great importance to the clustering accuracy. From the previous analysis of simulated data, it was found that the accuracy could reach over 99.5% when there were more than 60 training spikes of each cluster available in the “easy” datasets. Moreover, the 1D-CNN needed at least 120 spikes of each cluster, with over 98.4% accuracy guaranteed in the “difficult” datasets. As far as the real data were concerned, considering fewer sampling points and severe noise, there should be more than 200 training spikes of each cluster to obtain an accuracy of 95%. Thus, there is no doubt that the accuracy increases with the number of training spikes. When considering both factors, a sufficient number of spikes with enough sampling points in the training set can reliably guarantee a high accuracy and strong robustness. On the other hand, it is time-consuming and laborious to obtain a large number of training spikes. Therefore, it will be necessary to find a tradeoff between performance and cost in real applications.
Although “WMsorting” is a semi-supervised algorithm [54], whereas the method presented here is fully supervised, the degree of manual interventions was almost the same. There were about 170 spikes in the training set for experiment 1 using our proposed method and 180 spikes using “WMsorting.” Hence, the workload when using our method with respect to labeling is equivalent to that when using a semi-supervised algorithm. Meanwhile, the accuracy of our proposed deep learning method was improved with respect to “WMsorting,” especially for the “difficult” datasets.
Compared to traditional methods and the deep-learning-based MLP model [45], our proposed 1D-CNN model had more obvious advantages on robustness. With the increase in noise level for all datasets, traditional methods worsened with an error rate as high as 46.28%, and the deep-learning-based MLP model worsened with an error rate as high as 48.45%. Our error rate was around 2% for all datasets, which was more robust.
We used the Anaconda software (4.4.10, Anaconda Company, Austin, USA) and implemented the proposed CNN architecture using Keras in Python 3. When performed on a dataset with 1738 training spikes and 1739 testing spikes using an Intel(R) Core(TM) i7-7820X central processing unit (CPU) equipped with a Nvidia GeForce RTX 2070 graphics processing unit (GPU) and 16 GB random-access memory (RAM), the model took approximately 17 s for training and 2 s for testing. This means that our proposed deep learning method could classify a spike in one millisecond. If the performance of the hardware were further improved, the computation time would be shorter. Thus, it is feasible to apply our 1D-CNN method in online spike sorting.

5. Conclusions

We proposed a novel deep learning method based on a 1D-CNN, which can be used in online or offline automatic spike sorting. In future works, we will improve this CNN model to effectively cluster overlapping spikes and expand it to simultaneously classify multichannel recordings. As is well known, the shapes of overlapping spikes can appear with various and complicated changes, whereby the superposition time and the unit waveform are always different. This complicates spike sorting. In addition, high-density multielectrode arrays have been used to record the firing activities of large ensembles of neurons. Thus, a spike waveform may be recorded by multiple electrodes at the same time. How to distinguish spikes simultaneously recorded by different electrodes then becomes another important problem to be solved in spike sorting.

Author Contributions

Conceptualization, Z.L.; methodology, Z.L.; validation, Y.W. and N.Z.; formal analysis, X.L.; data curation, N.Z.; writing—review and editing, Z.L.; writing—original draft preparation, Y.W.; supervision, X.L; funding acquisition, Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China (61971374, 61871465, 61603327) and the Natural Science Foundation of Hebei China (F2017203010).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Arel, I.; Rose, D.C.; Karnowski, T.P. Deep machine learning-a new frontier in artificial intelligence research. IEEE Comput. Intell. Mag. 2010, 5, 13–18. [Google Scholar] [CrossRef]
  2. Azami, H.; Escudero, J.; Darzi, A.; Sanei, S. Extracellular spike detection from multiple electrode array using novel intelligent filter and ensemble fuzzy decision making. J. Neurosci. Methods 2015, 239, 129–138. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Berenyi, A.; Somogyvari, Z.; Nagy, A.J.; Roux, L.; Long, J.D.; Fujisawa, S.; Stark, E.; Leonardo, A.; Harris, T.D.; Buzsaki, G. Large-scale, high-density (up to 512 channels) recording of local circuits in behaving animals. J. Neurophysiol. 2014, 111, 1132–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Obeid, I.; Wolf, P.D. Evaluation of spike-detection algorithms for a brain-machine interface application. IEEE Trans. Biomed. Eng. 2004, 51, 905–911. [Google Scholar] [CrossRef] [PubMed]
  5. Schmitzer-Torbert, N.; Jackson, J.; Henze, D.; Harris, K.D.; Redish, A.D. Quantitative measures of cluster quality for use in extracellular recordings. Neuroscience 2005, 131, 1–11. [Google Scholar] [CrossRef]
  6. Rey, H.G.; Pedreira, C.; Quian Quiroga, R. Past, present and future of spike sorting techniques. Brain Res. Bull. 2015, 119, 106–117. [Google Scholar] [CrossRef] [Green Version]
  7. Lefebvre, B.; Yger, P.; Marre, O. Recent progress in multi-electrode spike sorting methods. J. Physiol. Paris 2016, 110, 327–335. [Google Scholar] [CrossRef] [Green Version]
  8. Franke, F.; Jaeckel, D.; Dragas, J.; Müller, J.; Radivojevic, M.; Bakkum, D.; Hierlemann, A. High-density microelectrode array recordings and real-time spike sorting for closed-loop experiments: An emerging technology to study neural plasticity. Front. Neural Circuits 2012, 6, 105. [Google Scholar] [CrossRef] [Green Version]
  9. Buzsáki, G. Large-scale recording of neuronal ensembles. Nat. Neurosci. 2004, 7, 446–451. [Google Scholar] [CrossRef]
  10. Takekawa, T.; Isomura, Y.; Fukai, T. Accurate spike sorting for multi-unit recordings. Eur. J. Neurosci. 2010, 31, 263–272. [Google Scholar] [CrossRef]
  11. Quiroga, R.Q. Spike sorting. Curr. Biol. 2012, 22, 45–46. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Sukiban, J.; Voges, N.; Dembek, T.A.; Pauli, R.; Visservandewalle, V.; Denker, M.; Weber, I.; Timmermann, L.; Grun, S. Evaluation of Spike Sorting Algorithms: Application to Human Subthalamic Nucleus Recordings and Simulations. Neuroscience 2019, 414, 168–185. [Google Scholar] [CrossRef] [PubMed]
  13. Gold, C.; Henze, D.A.; Koch, C.; Buzsaki, G. On the origin of the extracellular action potential waveform: A modeling study. J. Neurophysiol. 2006, 95, 3113–3128. [Google Scholar] [CrossRef] [PubMed]
  14. Quiroga, R.Q.; Nadasdy, Z.; Ben-Shaul, Y. Unsupervised spike detection and sorting with wavelets and superparamagnetic clustering. Neural Comput. 2004, 16, 1661–1687. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Saif-Ur-Rehman, M.; Lienkamper, R.; Parpaley, Y.; Wellmer, J.; Liu, C.; Lee, B.; Kellis, S.; Andersen, R.; Iossifidis, I.; Glasmachers, T.; et al. SpikeDeeptector: A deep-learning based method for detection of neural spiking activity. JNEng 2019, 16, 056003. [Google Scholar] [CrossRef] [Green Version]
  16. Nenadic, Z.; Burdick, J.W. Spike detection using the continuous wavelet transform. IEEE Trans. Biomed. Eng. 2005, 52, 74–87. [Google Scholar] [CrossRef] [Green Version]
  17. Harris, K.D.; Henze, D.A.; Csicsvari, J.; Hirase, H.; Buzsaki, G. Accuracy of tetrode spike separation as determined by simultaneous intracellular and extracellular measurements. J. Neurophysiol. 2000, 84, 401–414. [Google Scholar] [CrossRef]
  18. Shoham, S.; Fellows, M.R.; Normann, R.A. Robust, automatic spike sorting using mixtures of multivariate t-distributions. J. Neurosci. Methods 2003, 127, 111–122. [Google Scholar] [CrossRef]
  19. Einevoll, G.T.; Franke, F.; Hagen, E.; Pouzat, C.; Harris, K.D. Towards reliable spike-train recordings from thousands of neurons with multielectrodes. Curr. Opin. Neurobiol. 2012, 22, 11–17. [Google Scholar] [CrossRef] [Green Version]
  20. Letelier, J.C.; Weber, P.P. Spike sorting based on discrete wavelet transform coefficients. J. Neurosci. Methods 2000, 101, 93–106. [Google Scholar] [CrossRef]
  21. Hulata, E.; Segev, R.; Ben-Jacob, E. A method for spike sorting and detection based on wavelet packets and Shannon’s mutual information. J. Neurosci. Methods 2002, 117, 1–12. [Google Scholar] [CrossRef]
  22. Chah, E.; Hok, V.; Della-Chiesa, A.; Miller, J.J.; O’Mara, S.M.; Reilly, R.B. Automated spike sorting algorithm based on Laplacian eigenmaps and k-means clustering. JNEng 2011, 8, 016006. [Google Scholar]
  23. Rossant, C.; Kadir, S.N.; Goodman, D.F.M.; Schulman, J.; Hunter, M.L.D.; Saleem, A.B.; Grosmark, A.; Belluscio, M.; Denfield, G.H.; Ecker, A.S.; et al. Spike sorting for large, dense electrode arrays. Nat. Neurosci. 2016, 19, 634–641. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Kadir, S.N.; Goodman, D.F.; Harris, K.D. High-dimensional cluster analysis with the masked EM algorithm. Neural Comput. 2014, 26, 2379–2394. [Google Scholar] [CrossRef] [Green Version]
  25. Gibson, S.; Judy, J.W.; Markovic, D. Spike Sorting: The First Step in Decoding the Brain: The first step in decoding the brain. ISPM 2012, 29, 124–143. [Google Scholar] [CrossRef]
  26. Matthews, B.A.; Clements, M.A. Spike sorting by joint probabilistic modeling of neural spike trains and waveforms. Comput. Intell. Neurosci. 2014, 2014, 643059. [Google Scholar] [CrossRef]
  27. Tariq, T.; Satti, M.H.; Kamboh, H.M.; Saeed, M.; Kamboh, A.M. Computationally efficient fully-automatic online neural spike detection and sorting in presence of multi-unit activity for implantable circuits. Comput. Methods Programs Biomed. 2019, 179, 104986. [Google Scholar] [CrossRef]
  28. Chung, J.E.; Magland, J.F.; Barnett, A.H.; Tolosa, V.M.; Greengard, L.F. A Fully Automated Approach to Spike Sorting. Neuron 2017, 95, 1381–1394. [Google Scholar] [CrossRef] [Green Version]
  29. Horton, P.M.; Nicol, A.U.; Kendrick, K.M.; Feng, J.F. Spike sorting based upon machine learning algorithms (SOMA). J. Neurosci. Methods 2007, 160, 52–68. [Google Scholar] [CrossRef]
  30. Frey, U.; Egert, U.; Heer, F.; Hafizovic, S.; Hierlemann, A. Microelectronic system for high-resolution mapping of extracellular electric fields applied to brain slices. Biosens. Bioelectron. 2009, 24, 2191–2198. [Google Scholar] [CrossRef]
  31. Jun, J.J.; Steinmetz, N.A.; Siegle, J.H.; Denman, D.J.; Bauza, M.; Barbarits, B.; Lee, A.K.; Anastassiou, C.A.; Andrei, A.; Aydin, C.; et al. Fully integrated silicon probes for high-density recording of neural activity. Nature 2017, 551, 232–236. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Spira, M.E.; Hai, A. Multi-electrode array technologies for neuroscience and cardiology. Nature Nanotechnol. 2013, 8, 83–94. [Google Scholar] [CrossRef] [PubMed]
  33. Harris, K.D.; Quiroga, R.Q.; Freeman, J.; Smith, S.L. Improving data quality in neuronal population recordings. Nat. Neurosci. 2016, 19, 1165–1174. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Huber, P.J. Robust Statistics; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2004; Volume 523. [Google Scholar]
  35. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [Green Version]
  36. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  37. Nadeem, M.W.; Ghamdi, M.A.A.; Hussain, M.; Khan, M.A.; Khan, K.M.; Almotiri, S.H.; Butt, S.A. Brain Tumor Analysis Empowered with Deep Learning: A Review, Taxonomy, and Future Challenges. Brain Sci. 2020, 10, 118. [Google Scholar] [CrossRef] [Green Version]
  38. Ali, M.B.; Gu, I.Y.; Berger, M.S.; Pallud, J.; Southwell, D.; Widhalm, G.; Roux, A.; Vecchio, T.G.; Jakola, A.S. Domain Mapping and Deep Learning from Multiple MRI Clinical Datasets for Prediction of Molecular Subtypes in Low Grade Gliomas. Brain Sci. 2020, 10, 463. [Google Scholar] [CrossRef]
  39. Ravì, D.; Wong, C.; Deligianni, F.; Berthelot, M.; Andreu-Perez, J.; Lo, B.; Yang, G.-Z. Deep learning for health informatics. IEEE J. Biomed. Health Inform. 2017, 21, 4–21. [Google Scholar] [CrossRef] [Green Version]
  40. Greenspan, H.; Van Ginneken, B.; Summers, R.M. Guest editorial deep learning in medical imaging: Overview and future promise of an exciting new technique. IEEE Trans. Med. Imaging 2016, 35, 1153–1159. [Google Scholar] [CrossRef]
  41. Fujita, H.; Cimr, D. Computer aided detection for fibrillations and flutters using deep convolutional neural network. Inf. Sci. 2019, 486, 231–239. [Google Scholar] [CrossRef]
  42. Yang, K.; Wu, H.; Zeng, Y. A Simple Deep Learning Method for Neuronal Spike Sorting. In Proceedings of the 2017 International Conference on Cloud Technology and Communication Engineering (CTCE2017), Guilin, China, 18–20 August 2017; p. 012062. [Google Scholar]
  43. Quiroga, R.Q. Wave_Clus: Unsupervised Spike Detection and Sorting. 2017. Available online: https://vis.caltech.edu/~rodri/Wave_clus/Wave_clus_home.htm (accessed on 10 November 2020).
  44. Huang, L.; Ling, B.W.; Cai, R.; Zeng, Y.; He, J.; Chen, Y. WMsorting: Wavelet Packets Decomposition and Mutual Information Based Spike Sorting Method. IEEE Trans. NanoBiosci. 2019, 18, 283–295. [Google Scholar] [CrossRef] [PubMed]
  45. Park, I.Y.; Eom, J.; Jang, H.; Kim, S.; Hwang, D. Deep Learning-Based Template Matching Spike Classification for Extracellular Recordings. Appl. Sci. 2019, 10, 301. [Google Scholar] [CrossRef] [Green Version]
  46. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  47. Wu, T.; Zhao, W.; Keefer, E.; Yang, Z. Deep compressive autoencoder for action potential compression in large-scale neural recording. JNEng 2018, 15, 6. [Google Scholar] [CrossRef] [Green Version]
  48. Chu, C.C.; Chien, P.F.; Hung, C.P. Multi-electrode recordings of ongoing activity and responses to parametric stimuli in macaque V1. CRCNS 2014, 61, 41–54. [Google Scholar] [CrossRef]
  49. Chu, C.C.; Chien, P.F.; Hung, C.P. Tuning dissimilarity explains short distance decline of spontaneous spike correlation in macaque V1. Vision Res. 2014, 96, 113–132. [Google Scholar] [CrossRef] [Green Version]
  50. Hou, Y.; Zhou, L.; Jia, S.; Lun, X. A novel approach of decoding EEG four-class motor imagery tasks via scout ESI and CNN. JNEng 2020, 17, 016048. [Google Scholar] [CrossRef]
  51. Özgür, A.; Özgür, L.; Güngör, T. Text categorization with class-based and corpus-based keyword selection. In International Symposium on Computer and Information Sciences; Springer: Berlin/Heidelberg, Germany, 2005; pp. 606–615. [Google Scholar]
  52. Rácz, M.; Liber, C.; Németh, E.; Fiáth, R.; Rokai, J.; Harmati, I.; Ulbert, I.; Márton, G. Spike detection and sorting with deep learning. J. Neural Eng. 2020, 17, 016038. [Google Scholar] [CrossRef]
  53. Saif-Ur-Rehman, M.; Ali, O.; Lienkaemper, R.; Dyck, S.; Metzler, M.; Parpaley, Y.; Wellmer, J.; Liu, C.; Lee, B.; Kellis, S. SpikeDeep-Classifier: A deep-learning based fully automatic offline spike sorting algorithm. arXiv 2019, arXiv:1912.10749. [Google Scholar] [CrossRef]
  54. Huang, G.; Song, S.; Gupta, J.N.D.; Wu, C. Semi-Supervised and Unsupervised Extreme Learning Machines. IEEE Trans. Cybern. 2017, 44, 2405–2417. [Google Scholar] [CrossRef]
Figure 1. Architecture of the proposed convolutional neural network (CNN) model used in the simulated database. The blue region corresponds to convolutional processes, and the red region represents the pooling. The stride of all convolutional layers was 1.
Figure 1. Architecture of the proposed convolutional neural network (CNN) model used in the simulated database. The blue region corresponds to convolutional processes, and the red region represents the pooling. The stride of all convolutional layers was 1.
Brainsci 10 00835 g001
Figure 2. Details of the simulated database. Different colors represent different clusters. The number in parentheses represents the number of overlapped spikes.
Figure 2. Details of the simulated database. Different colors represent different clusters. The number in parentheses represents the number of overlapped spikes.
Brainsci 10 00835 g002
Figure 3. Details of the experimental datasets. Different colors represent different clusters. The number in parentheses represents the number of overlapped spikes.
Figure 3. Details of the experimental datasets. Different colors represent different clusters. The number in parentheses represents the number of overlapped spikes.
Brainsci 10 00835 g003
Figure 4. Proportion of data in training set and testing set for the six experiments. E = experiment.
Figure 4. Proportion of data in training set and testing set for the six experiments. E = experiment.
Brainsci 10 00835 g004
Figure 5. Confusion matrix: (a) C_Easy1_noise015; (b) C_Difficult1_noise015.
Figure 5. Confusion matrix: (a) C_Easy1_noise015; (b) C_Difficult1_noise015.
Brainsci 10 00835 g005
Figure 6. Waveforms of spikes in each cluster: (ac) C_Easy1_noise015; (df) C_Difficult1_noise015.
Figure 6. Waveforms of spikes in each cluster: (ac) C_Easy1_noise015; (df) C_Difficult1_noise015.
Brainsci 10 00835 g006
Figure 7. Falsely classified spikes and randomly selected correctly classified spikes: (a,b) C_Easy1_noise015; (ce) C_Difficult1_noise015.
Figure 7. Falsely classified spikes and randomly selected correctly classified spikes: (a,b) C_Easy1_noise015; (ce) C_Difficult1_noise015.
Brainsci 10 00835 g007
Figure 8. The cross-entropy loss function on the training set and testing set: (a) C_Easy1_015; (b) C_Difficult1_015.
Figure 8. The cross-entropy loss function on the training set and testing set: (a) C_Easy1_015; (b) C_Difficult1_015.
Brainsci 10 00835 g008
Figure 9. Classification accuracies of the recording channels with different numbers of spike clusters: (a) Three clusters; (b) two clusters; (c) four clusters; (d) five clusters.
Figure 9. Classification accuracies of the recording channels with different numbers of spike clusters: (a) Three clusters; (b) two clusters; (c) four clusters; (d) five clusters.
Brainsci 10 00835 g009
Table 1. Comparison of classification accuracy on simulated database (using “WMsorting”). 1D, one-dimensional.
Table 1. Comparison of classification accuracy on simulated database (using “WMsorting”). 1D, one-dimensional.
DatasetNL 1WMsorting1D-CNNImprovement
F 2 = 3F = 10E1 3E2E3E4E5E6
C_Easy100599.6099.5299.6499.5399.4099.6399.6799.770.17
01099.6699.5799.5299.7299.8299.8899.8699.940.28
01599.7199.6399.6199.4999.8299.8499.8199.770.13
02099.5799.4899.3399.3399.6499.8499.7699.880.31
02599.4599.4299.5599.5699.8199.7499.8599.700.40
03099.5799.4899.4999.6599.7199.1899.6299.710.14
03599.4399.3899.4999.4399.6599.5699.2099.600.22
04099.7699.7099.3599.5199.5699.7999.7599.820.06
C_Easy200599.5099.4199.7299.6499.7199.6699.6699.880.38
01099.5299.4099.5899.6299.7599.7699.9199.830.39
01599.4499.3899.3298.8999.2399.3399.6199.770.33
02099.2999.3599.1999.6999.3399.5599.6299.660.31
C_Difficult100594.4795.0098.0799.2199.1199.1198.9799.054.21
01092.8993.7698.9999.0099.0999.4699.6699.836.07
01590.1891.1897.7698.4398.6399.1898.7599.428.24
02085.3886.4595.1698.5499.0198.8398.7398.8312.38
C_Difficult200599.2399.4499.2298.8899.4899.6699.8599.820.41
01098.9399.5199.4899.7899.7899.8499.9999.940.48
01598.0599.5199.5799.5899.5399.6799.7199.710.20
02095.9999.6699.5899.7599.6199.7599.6799.830.17
1 Noise level; 2 feature dimensions; 3 experiment.
Table 2. Comparison of classification accuracy on simulated database (using traditional methods and deep-learning-based multilayer perceptron (MLP) model). PCA, principal component analysis; FCM, fuzzy C-means; FSDE, first and second derivative extrema; CORR, correlation-based; SVM, support vector machine.
Table 2. Comparison of classification accuracy on simulated database (using traditional methods and deep-learning-based multilayer perceptron (MLP) model). PCA, principal component analysis; FCM, fuzzy C-means; FSDE, first and second derivative extrema; CORR, correlation-based; SVM, support vector machine.
DatasetNL 1PCA + FCMFSDE + K-meansCORR + FCMFusion + SVMMLP1D-CNN
F 2 = 3F = 10F = 3F = 3F = 10F = 10E2 3
C_Easy100599.3799.3594.6297.5097.3898.6699.2699.53
01099.7299.7295.5494.0496.4598.9899.4399.72
01599.2599.2894.4590.5494.9498.2299.2599.49
02099.4099.4095.0888.7792.4397.3599.1999.33
02599.2499.24 84.4186.8495.45 99.56
03098.7398.59 81.5080.8388.66 99.65
03597.7695.16 77.0273.8083.22 99.43
04096.4968.54 75.5864.6278.12 99.51
C_Easy200598.4898.6894.8193.2096.0492.2398.6899.64
01097.1698.2494.8386.0282.1992.9398.4999.62
01592.5294.4994.9683.0582.8289.8097.1998.89
02085.2088.6092.7179.8178.2286.2495.2099.69
C_Difficult100595.8672.5494.5083.4886.0897.5898.7899.21
01089.5666.1194.7865.6971.5594.8198.9399.00
01576.4161.3393.8157.4958.8487.8597.5598.43
02063.0354.0590.6053.7253.8178.5996.6298.54
C_Difficult200598.6998.8194.3891.5094.5087.4098.4998.88
01098.6498.7694.4890.9696.3388.0794.6699.78
01594.3997.3387.1888.1796.0274.6582.2099.58
02084.6383.3781.7184.7795.4867.2551.5599.75
1 Noise level; 2 feature dimensions; 3 experiment.
Table 3. Macro_F scores on experimental datasets. 1D, one-dimensional.
Table 3. Macro_F scores on experimental datasets. 1D, one-dimensional.
ChannelsNumber of ClustersWMsorting1D_CNNImprovement
F 1 = 3F = 10E1 2E2E3E4E5E6
125288.9486.2187.3390.9694.4997.0094.4997.508.56
66394.5695.3695.1696.3396.8497.3497.6798.172.81
69396.4396.4997.3397.3497.6696.3497.1598.171.68
77384.0587.8489.4591.7891.6796.3395.9797.339.49
79391.5992.2891.4995.4994.9996.3596.3396.163.88
84392.2694.0296.5198.1798.0098.0097.6799.004.98
91394.3795.2994.3396.0095.3295.3295.3296.301.01
92392.3194.3893.3395.5194.6395.6595.5195.651.27
94387.2988.6188.5088.1193.0192.1493.5094.365.75
98366.1872.7374.6680.0679.3390.2590.0090.1417.41
99383.2987.1681.6990.5891.7492.0990.3593.296.13
100389.9290.6488.3193.3896.0094.9896.1797.667.02
101384.2885.6287.3490.7894.0194.0093.8095.8510.23
108390.2990.6193.3094.3494.6595.6794.8795.504.89
109390.2992.5294.1894.0095.8394.0195.0097.494.97
114392.4991.0691.8292.7393.4395.6696.9996.323.83
115389.9088.6191.3392.7894.4494.6295.6396.146.24
122386.1190.6090.3593.3094.4496.8495.9796.015.41
124379.2385.0786.7288.2488.4893.7792.6896.8311.76
67485.3487.0693.5494.4793.4895.9995.8496.369.30
68478.0680.3581.9984.1088.0392.0090.1790.8010.45
70485.9688.7289.0192.0292.9994.9993.7295.246.52
83484.1986.1092.4394.2594.5293.5294.3596.7610.66
95488.0689.7992.8690.6193.1792.9493.5496.076.28
107485.1086.7191.2193.8395.7696.4995.6595.018.30
116484.1286.1995.4494.9795.6898.6298.7599.0012.81
82546.2253.5867.7272.4778.4180.2778.4383.6330.05
126586.1789.6597.1997.4997.3898.3998.5998.408.75
1 Feature dimensions; 2 experiment.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Z.; Wang, Y.; Zhang, N.; Li, X. An Accurate and Robust Method for Spike Sorting Based on Convolutional Neural Networks. Brain Sci. 2020, 10, 835. https://doi.org/10.3390/brainsci10110835

AMA Style

Li Z, Wang Y, Zhang N, Li X. An Accurate and Robust Method for Spike Sorting Based on Convolutional Neural Networks. Brain Sciences. 2020; 10(11):835. https://doi.org/10.3390/brainsci10110835

Chicago/Turabian Style

Li, Zhaohui, Yongtian Wang, Nan Zhang, and Xiaoli Li. 2020. "An Accurate and Robust Method for Spike Sorting Based on Convolutional Neural Networks" Brain Sciences 10, no. 11: 835. https://doi.org/10.3390/brainsci10110835

APA Style

Li, Z., Wang, Y., Zhang, N., & Li, X. (2020). An Accurate and Robust Method for Spike Sorting Based on Convolutional Neural Networks. Brain Sciences, 10(11), 835. https://doi.org/10.3390/brainsci10110835

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop