Next Article in Journal
Predicting the Output of Solar Photovoltaic Panels in the Absence of Weather Data Using Only the Power Output of the Neighbouring Sites
Previous Article in Journal
Accuracy of Intracranial Pressure Monitoring—Single Centre Observational Study and Literature Review
Previous Article in Special Issue
Victim Detection and Localization in Emergencies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Sound Event Localization and Detection Using Imbalanced Real and Synthetic Data via Multi-Generator

Department of Computer Engineering, Chosun University, Gwangju 61452, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(7), 3398; https://doi.org/10.3390/s23073398
Submission received: 20 February 2023 / Revised: 17 March 2023 / Accepted: 21 March 2023 / Published: 23 March 2023

Abstract

:
This study proposes a sound event localization and detection (SELD) method using imbalanced real and synthetic data via a multi-generator. The proposed method is based on a residual convolutional neural network (RCNN) and a transformer encoder for real spatial sound scenes. SELD aims to classify the sound event, detect the onset and offset of the classified event, and estimate the direction of the sound event. In Detection and Classification of Acoustic Scenes and Events (DCASE) 2022 Task 3, SELD is performed with a few real spatial sound scene data and a relatively large number of synthetic data. When a model is trained using imbalanced data, it can proceed by focusing only on a larger number of data. Thus, a multi-generator that samples real and synthetic data at a specific rate in one batch is proposed to prevent this problem. We applied the data augmentation technique SpecAugment and used time-frequency masking to the dataset. Furthermore, we propose a neural network architecture to apply the RCNN and transformer encoder. Several models were trained with various structures and hyperparameters, and several ensemble models were obtained by “cherry-picking” specific models. Based on the experiment, the single model of the proposed method and the model applied with the ensemble exhibited improved performance compared with the baseline model.

1. Introduction

SELD is commonly performed to preprocess audio signals. SELD can be divided into subtasks called sound event detection (SED) and sound source localization (SSL).
SED aims to classify a specific sound event in a sound [1] and detect the onset and offset of each sound event [2,3]. In real life, situations such as background noise, overlapping sound sources, and the occurrence of sound sources belonging to the same event class can occur. Overlapping sources refer to situations where multiple event classes occur simultaneously in a given time instance, which can be addressed by applying polyphonic SED that can detect multiple overlapping event classes at the same time. Additionally, there are scene-dependent and scene-independent approaches in SED. Scene-dependent approaches provide information on the acoustic scene for both training and testing, while scene-independent approaches do not provide such information. Such SED tasks can be utilized in various fields in which sound is detected, such as surveillance [4] or audio captioning [5].
SSL is the task of estimating the location of a sound source in three-dimensional coordinates using the recorded multichannel audio signal, with the microphone being the reference point for the recording. Like the human auditory system, which utilizes physical variables such as time, intensity, and spectral shape to perform SSL, deep learning-based SSL also estimates these variables, such as time, intensity, and spectrogram, to localize sound sources. SSL is not only a task of estimating direction, but there are also tasks that consider distance as well. In this study, SSL simplifies the direction-of-arrival (DOA) estimation by only considering azimuth and elevation without estimating the distance to the microphone array. Such SSL is used in various fields such as source separation [6], speech recognition [7], and speech enhancement [8]. In SELD, two subtasks, SED and SSL, are performed simultaneously.
There is a challenge called DCASE that performs SELD. The DCASE challenge was first held in 2013 and has continued until 2022. The DCASE Task 3 challenge on SELD was held in 2019, and various deep-learning-based methods have been proposed. Yearly, participants in the Task 3 challenge attempt to achieve the SELD in various environments. DCASE 2019 aimed to perform SELD using audio data containing only static sound sources. DCASE 2020 was similar to that of 2019; however, a dynamic sound source was added [9]. In 2021, directional interference was added to the 2020 audio data [10].
As the challenge progressed, numerous deep learning-based methods for improving SELD performance were proposed. The proposed method differentiates input features depending on the type of audio data format used, such as multichannel microphone array (MIC) and first-order ambisonic (FOA). Both formats commonly use mel-energy, but different approaches have been proposed for each format: intensity vectors(IV) for encoding time-difference characteristics in FOA format, and the use of Spatial Cue-Augmented Log-Spectrogram (SALSA) [11] and SALSA-Lite [12] for MIC format. SALSA employs a linear-frequency log-power spectrogram for each channel and an eigenvector-based phase vector (EPV). On the other hand, SALSA-Lite uses a modified input feature that replaces EPV with simple frequency-normalized inter-channel phase difference (NIPD). SALSA-Lite is more computationally efficient than SALSA. Furthermore, activity-coupled Cartesian direction-of-arrival (ACCDOA) [13] and Multi-ACCDOA [14] have been proposed as output formats for the SELD models. A neural-network-based SELD generally comprises a two-branch [9,15] and two-stage [16] method. The two-branch method estimates SED and DoA using two branches composed of fully connected layers in a single network. This approach linearly combines the weights and losses of the model for multi-objectives. However, the model’s performance can be significantly affected by the weight and might be difficult to tune. The two-stage method uses two separate networks to estimate SED and DoA. This approach avoids the multi-objective problem because it builds a network for each target. However, it increases the complexity of the system and the overall size of the network. A method proposed to address the issues of the two methods is ACCDOA. ACCDOA uses a single network and a single branch by assigning the sound event activity to the length of the corresponding Cartesian DoA vector. However, since ACCDOA is a class-wise output format, overlapping events cannot be detected in the same class. This problem was solved using multi-ACCDOA, which extended track-wise. The multi-ACCDOA format can detect overlapping sounds in the same class, resolving the problem of ACCDOA.
DCASE 2019–2021 were conducted using the synthesized data. However, DCASE 2022 Task 3 differs from previous competitions in that it includes a relatively small amount of real spatial acoustic scene data and a relatively large amount of synthetic data generated using specific indoor impulse responses [17]. There is a difference with the dataset. Furthermore, since the final evaluation uses only real data, the model’s performance is highly dependent on how well it generalizes when trained using large amounts of synthetic data. Therefore, in order to achieve optimal performance, participants need to solve the challenge of training a model with a disproportionate amount of data appropriately.
The two methods discussed subsequently are suggested for effective training, using imbalanced real and synthetic data. First, an RCNN and transformer encoder [18]-based neural network architecture are proposed. Furthermore, a multi-generator method is introduced that samples actual and synthetic data at a predetermined rate in a single batch. For the performance of the proposed SELD model, the error rate and F1-score are used as evaluation metrics for SED using the DCASE 2022 dataset, and the class-dependent localization error and recall are used as evaluation metrics for SSL [19]. The performance of the proposed SELD model is then compared with that of the DCASE 2022 baseline model.
Various proposed techniques were applied in the DCASE 2022 Task3 baseline. The model uses 7-channel data as input. In the FOA format, 4-channel mel-spectrogram and 3-channel IV are used, while in the MIC format, 4-channel mel-spectrogram and 3-channel SALSA-Lite are used. In addition, the baseline model consists of three layers of CNN with a kernel size of 3 × 3 , two layers of Bi-GRU, and two layers of the FC layer.
The remainder of this paper is organized as follows. Section 2 describes the structure and method of the proposed model. Section 3 compares and evaluates the performances of the models. Finally, the conclusions are provided in Section 4.

2. Proposed Sound Event Detection and Localization Method

2.1. Architecture

The structure of the proposed model is shown in Figure 1. The proposed neural network model is based on the RCNN structure. RCNN is an advanced form of residual network (ResNet) that uses residual blocks, which have the same input and output dimensions and add the input values to the output values, to minimize information loss [20]. In our proposed method, the stem block is formed using convolutional blocks to utilize the RCNN structure of the model. Here, the kernel sizes of the convolution block are 5 × 5 and 1 × 1. The residual block comprises a convolutional neural network (CNN) layer with residual connections, batch normalization, and a max pooling layer. The structure of the residual block is shown in Figure 2. The CNN with residual blocks utilized a kernel size of 3 × 3 , and a total of three layers of residual blocks were used. Then, the transformer encoder and fully connected layers are formed. The hyper tangent is used as an activation function. The transformer encoder is a sequence model that preserves the positional information of input data and can simultaneously learn all positional relationships in the input sequence. Unlike recurrent neural network (RNN), it does not require sequential processing, which makes it faster, and can achieve better performance than RNN when the input sequence is long [18]. The attention model used in the transformer encoder of the proposed model employs multi-head attention. For the hyperparameters, we set the encoder dimension to 128, the number of layers to 1, the number of heads to 4, and the dimension of the feedforward layer to 256, 512, 1024, and 2048. The hyper tangent is used as an activation function. Finally, a single model with a single branch is trained and Multi-ACCDOA is applied as the final output format since SELD is performed for overlapping sources. The format is defined as follows:
P R 3 × N × C × T ,
where N, C, and T denote the track, class, and frame index, respectively.
We used the FOA format of four channels as the audio data format. Four mel-spectrograms extracted from each channel and the IV of three channels were used as input features of the model. The extracted mel-spectrograms and IVs were stacked to form a total of seven input channels. The IV is used to estimate the DOA in FOA B-format recordings [21]. The FOA B-format comprises four channels and can be expressed as W, X, Y, and Z. IV can be calculated using the four-channel spectrogram that is the short-time Fourier transform output [22].
I t , k = R ( W * , h [ t , k ] ) = [ I X [ t , k ] , I Y [ t , k ] , I Z [ t , k ] ] T ,
where h [ t , k ] = [ X [ t , k ] , Y [ t , k ] , Z [ t , k ] ] T . R ( · ) represents the real part of the complex number, and ( · ) * represents the complex conjugate number. The IVs are calculated for the X, Y, and Z channels based on the w channel in the mel-spectrogram of the four channels.

2.2. Multi-Generator

The DCASE dataset consisted of 1200 synthetic data and 121 real data. According to the DCASE regulations, additional synthetic data were obtained, and a dataset consisting of 6000 synthetic data and 121 real data was complied with. Among the synthetic data, 5700 were used for training, and the real dataset was composed of 67 examples designated by DCASE. The ratio of real and synthetic data in the given dataset is imbalanced. Owing to the limited number of real data, a multi-generator was presented to utilize the data efficiently for training. The generator plays a role in loading data for model training and evaluation. When an imbalanced dataset is used for training, only a relatively large amount of data can be focused on. Thus, we utilized the multi-generator to prevent this scenario. The multi-generator distinguishes generators according to the type of data. Further, it allows for sampling of real and synthetic data according to the predetermined ratio. Accordingly, it does not focus only on one data type. Here, the batch size was set to 128. We used two generators, one for real data and one for synthetic data, to sample a batch consisting of 8 real data and 120 synthetic data.
The DCASE dataset consisted of 1200 synthetic data and 121 real data. We complied with the competition regulations and acquired additional synthetic data, resulting in the construction of a dataset consisting of 6000 synthetic data and 121 real data. Among these, 6000 synthetic data and 67 real data were used for training. For validation data, we used the real data separated by DCASE. The ratio of real and synthetic data in the given dataset is imbalanced, which leads to an imbalance in the ratio of the data used for training. Owing to the limited number of real data, a multi-generator was presented to utilize the data efficiently for training. The generator performs a role same as that of a dataloader, loading the dataset. A single generator was used to train the model in the existing training method. However, when an imbalanced dataset is used for training, only a relatively large amount of data can be focused on. Thus, we utilized the multi-generator to prevent this scenario. Multi-generator regularly samples synthetic and real data at a predetermined rate. Accordingly, it does not focus only on one data type. Here, the batch size was set to 128. We used two generators, one for real data and one for synthetic data, to sample a batch consisting of 8 real data and 120 synthetic data.
The dataset used in this study consists of 6000 synthetic data and 121 real data. Among the synthetic data, 5700 were used for training, and the real dataset was composed of 67 examples designated by DCASE. The ratio of real and synthetic data in the given dataset is imbalanced. Owing to the limited number of real data, a multi-generator was presented to utilize the data efficiently for training. The generator plays a role in loading data for model training and evaluation. When an imbalanced dataset is used for training, only a relatively large amount of data can be focused on. Thus, we utilized the multi-generator to prevent this scenario. The multi-generator distinguishes generators according to the type of data. In addition, it allows for sampling of real and synthetic data according to the predetermined ratio. Accordingly, it does not focus only on one data type. Here, the batch size was set to 128. We used two generators, one for real data and one for synthetic data, to sample a batch consisting of 8 real data and 120 synthetic data.

2.3. Training Method

The hyperparameters to train the model are as follows: Nesterov momentum Adam [23] was employed as the optimizer. The dropout and learning rates were set to 0.2 and 10 3 , respectively. The learning rate scheduler was also employed for training. Particularly, we trained the neural network model with the same learning rate until 30 epochs. Subsequently, cosine annealing scheduling was utilized [24]. This starts with a relatively large learning rate and is dropped rapidly to the predefined minimum value. Here, the maximum and minimum learning rates were 10 3 and 10 4 , respectively.
Furthermore, the SpecAugment data augmentation method, widely used in speech processing, was employed [25]. Two types of masking were applied in SpecAugment: frequency and time masking. In frequency masking, consecutive frequencies of a spectrogram are randomly masked. Time masking is similar to frequency masking; however, consecutive time frames of a spectrogram are randomly masked. The spectrogram with two masking applied is shown in Figure 3. SpecAugment was randomly applied to each batch of input data. Time masking could be applied to the entire sequence, with a width of 8 frames, masking roughly 0.02 s of time information per batch. Frequency masking is also applied to the entire sequence, with 8 mel-bins masked out of a total of 64 mel-bins.

3. Performance Evaluations

Model performance was measured only on a limited number of real data in the development dataset. In addition, it was evaluated according to the SELD evaluation metrics used in the DCASE challenge; error rate, F1-score, localization error, and localization recall [19]. Here, error rate and F1-score were utilized as evaluation metrics for SED, and localization error ( L E C D ) and localization recall ( L R C D ) were used as DOA evaluation metrics, respectively. The error rate was determined using the number of errors in terms of insertions (I), deletions (D), substitutions (S), and reference events (N), and it is represented as
E R = D + I + S N .
The F1-score was calculated over all test data based on the total number of false positives (FP), false negatives (FN), and true positives (TP). The F1-score is expressed as
P = T P T P + F P , R = T P T P + F N , F = 2 P R P + R ,
where P and R are the mean precision and recall, respectively.
Class-dependent localization metrics are calculated based on class-aware localization. The class-aware localizations L E C and L R C are computed only between predictions and references of class c. L E C and L R C are defined as
L E c = | | A c D c | | 1 | | A c | | 1 ,
L R c = l | | A c ( l ) | | 1 l N c ( l ) ,
where A c , D c , and N c indicate the association matrix, distance matrix, and references, respectively. c 1 , , C represents the class index, and l = 1 , , L represents temporal frame outputs or some other preferred averaged segmental representation. Using the calculated L E c and L R c , the overall class-dependent L E C D and L R C D are determined as the class means as follows
L E C D = 1 C · L c l L E c ( l ) ,
L R C D = 1 C c L R c .
The SELD score is not included in the ranking calculation, but it is used as a performance metric to assess overall performance during model training. The SELD score is calculated by averaging four metrics: error rate, F1-score, localization error, and localization recall. The SELD score is calculated as follows:
S E L D s c o r e = E r r o r r a t e + ( 1 F 1 ) + L E + L R 4 .
The resulting value ranges from 0 to 1, with better model performance resulting in values closer to 0.
The performance between models follows the DCASE ranking system. Sort in ascending order based on the cumulative rank of the four metrics. For example, if model A is ranked as error rate: 1, F1-score: 1, localization error: 3, localization recall: 1, then the cumulative rank is 1 + 1 + 3 + 1 = 5. If the individual ranks of model B are error rate: 2, F1-score: 3, localization error: 1, localization recall: 2, the cumulative rank is 8. As a result, the overall ranking is A, B.
Performance was compared using a validation set composed only of real data, and the training was assumed to converge if there was no improvement in performance after 100 epochs. Table 1 shows the model’s performance for the training method as an evaluation metric. The proposed three training methods differ in the types of data used. Method 1 used only synthesized data and Methods 2–3 used the same synthesized data and real data but were trained with different approaches. Methods 1–3 were evaluated with real data to calculate performance metrics. Method 2 was fine-tuned with real data after training with synthetic data. Method 3 trained the model using multi-generators, as the proposed method in this paper. Such models were constructed with the same neural network architecture. Comparing method 1 and method 2, using real data for training showed better performance in all metrics. This indicates that there are differences between synthetic and real data. Therefore, using a small amount of real data appropriately in training is an important factor in improving performance. When comparing Method 2 and Method 3, the fine-tuning model received a better score in error rate ( 0.63 ), while the model with a multi-generator received better scores in the other metrics (F1: 0.31, L E : 23.17, L R : 62.0). As DCASE uses four evaluation metrics, the multi-generator, which showed overall improvement in three metrics, was deemed appropriate to use.
We trained various models by modifying hyperparameters based on models that applied RCNN and transformer encoder or bi-directional long-short-term memory (Bi-LSTM) to the RNN layer. We trained various types of models by changing the hyperparameters, such as the dimension of the transformer encoder. Specifically, we set the dimension to 256, 512, 1024, and 2048 and trained the models accordingly. Additionally, the L1 loss function or mean squared error (MSE) loss function was applied as a loss function. We cherry-picked the top 7 models with the best performance among the various models trained and presented the results in Table 2. The model using Bi-LSTM achieved high in terms of error rate ( 0.65 ) and localization error ( 22.24 ). In terms of error rate ( 0.65 ), F1-score ( 0.33 ), and localization recall ( 62.0 ), it was confirmed that the model with a transformer encoder showed the excellent performance. Out of all seven models, six models have transformer encoders constructed. In other words, the transformer encoder model outscored overall in all metrics. Accordingly, Model 7 with a 512-dimensional transformer encoder and L1 loss function outperforms compared to the other models by cumulative ranking.
The ensemble technique [26] was also applied to the model shown in Table 2. Particularly, the ensemble was applied to seven models in numerous forms. After that, the four ensemble models listed in Table 3 are lastly determined by cumulative ranking. Model I is the result of an ensemble of seven models, and Models II to IV are the results of an ensemble of six models in different combinations. Model II was ensemble of six models listed in Table 2, which were selected as 1, 3, 4, 5, 6, and 7. Model III was an ensemble of six models, which were selected as 1, 2, 4, 5, 6, and 7. Model IV was an ensemble of six models, which were selected as 1, 2, 3, 4, 6, and 7. All four ensemble models were conducted similarly in terms of error rate ( 0.59 ). Model I scored the highest in F1-score ( 0.35 ) and localization error ( 20.7 ), and Model IV showed the best in localization recall ( 59.0 ). In the cumulative ranking, the ensemble model, Model I, performed better than the other ensemble models.
The final performance results for the validation and evaluation dataset are depicted in Table 4. The proposed method models (Single Model 7, Ensemble Model 3) were based on the highest score in the cumulative ranking. In the case of the single model, the Model 7 received the highest cumulative ranked shown in Table 2. In the case of the ensemble model, Model III had the highest cumulative rank shown in the DCASE challenge result. Note that the evaluation dataset was not provided to participants, and model outcomes were measured by DCASE. Therefore, there is no performance evaluation for the proposed method on the eval set that was not submitted. On the validation data, the ensemble model performed as well on error rate ( 0.59 ) and F1-score ( 0.35 ), whereas the localization error ( 26.41 ) and localization recall ( 62.0 ) were better handled by the single model. On the evaluation data, the baseline model scored higher in localization recall ( 51.4 ). The proposed methods showed overall better performance on the Dev set compared to the baseline model. In addition, it was confirmed that the proposed method improved all three performance metrics except for Localization Recall on the Eval set. Since SELD evaluates based on four performance metrics, the proposed method that showed performance improvement in three metrics can be considered significant. Moreover, the proposed method achieved higher ranks than the baseline model in the cumulative ranking [27]. Consequently, it is implied that the proposed method appears to be better than the baseline and benefits SELD tasks.

4. Conclusions

In this paper, the sound event localization and detection method using the RCNN and transformer encoder was proposed for real spatial sound scenes. We made efforts to construct stem blocks and minimize the vanishing gradient by utilizing residual blocks. Additionally, we enhanced the model structure by replacing the conventional RNN with transformer encoders. To address the issue of the imbalanced dataset, we applied a multi-generator approach. This approach allows for proper sampling of real and synthetic data according to the defined ratio. Moreover, we applied SpecAugment as a data augmentation technique. We evaluated the proposed approach using performance metrics used in SELD and observed improved performance compared to the baseline model. In future research, we aim to advance the proposed approach by applying additional data preprocessing and deep learning techniques to further improve the overall performance.

Author Contributions

Conceptualization, C.C.; data curation, Y.S.; funding acquisition, C.C.; investigation, Y.S.; methodology, Y.S.; project administration, C.C.; software, Y.S.; supervision, C.C.; validation, Y.S. and C.C.; visualization, Y.S.; writing—original draft preparation, Y.S.; writing—review and editing, C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the research fund from Chosun University, 2021.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. McLoughlin, I.; Zhang, H.; Xie, Z.; Song, Y.; Xiao, W. Robust sound event classification using deep neural networks. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 540–552. [Google Scholar] [CrossRef] [Green Version]
  2. Benetos, E.; Dixon, S. Polyphonic music transcription using note onset and offset detection. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 37–40. [Google Scholar]
  3. Mesaros, A.; Heittola, T.; Virtanen, T.; Plumbley, M.D. Sound event detection: A tutorial. IEEE Signal Process. Mag. 2021, 38, 67–83. [Google Scholar] [CrossRef]
  4. Foggia, P.; Petkov, N.; Saggese, A.; Strisciuglio, N.; Vento, M. Audio surveillance of roads: A system for detecting anomalous sounds. IEEE Trans. Intell. Transp. Syst. 2016, 17, 279–288. [Google Scholar] [CrossRef]
  5. Drossos, K.; Adavanne, S.; Virtanen, T. Automated audio captioning with recurrent neural networks. In Proceedings of the 2017 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 15–18 October 2017; pp. 374–378. [Google Scholar]
  6. Chazan, S.E.; Hammer, H.; Hazan, G.; Goldberger, J.; Gannot, S. Multi-microphone speaker separation based on deep DOA estimation. In Proceedings of the European Signal Processing Conference (EUSIPCO), A Coruna, Spain, 2–6 September 2019; pp. 1–5. [Google Scholar]
  7. Lee, H.Y.; Cho, J.W.; Kim, M.; Park, H.M. DNN-based feature enhancement using DOA-constrained ICA for robust speech recognition. IEEE Signal Process. Lett. 2016, 23, 1091–1095. [Google Scholar] [CrossRef]
  8. Xenaki, A.; Bünsow Boldt, J.; Græsbøll Christensen, M. Sound source localization and speech enhancement with sparse Bayesian learning beamforming. J. Acoust. Soc. Am. 2018, 143, 3912–3921. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Politis, A.; Adavanne, S.; Virtanen, T. A Dataset of Reverberant Spatial Sound Scenes with Moving Sources for Sound Event Localization and Detection. In Proceedings of the Detection and Classification of Acoustic Scenes and Events Workshop (DCASE), Virtual, 2–4 November 2020; pp. 165–169. [Google Scholar]
  10. Politis, A.; Adavanne, S.; Krause, D.; Deleforge, A.; Srivastava, P.; Virtanen, T. A Dataset of Dynamic Reverberant Sound Scenes with Directional Interferers for Sound Event Localization and Detection. In Proceedings of the Detection and Classification of Acoustic Scenes and Events Workshop (DCASE), Virtual, 15–19 November 2021; pp. 125–129. [Google Scholar]
  11. Nguyen, T.N.T.; Watcharasupat, K.N.; Nguyen, N.K.; Jones, D.L.; Gan, W.S. Salsa: Spatial cue-augmented log-spectrogram features for polyphonic sound event localization and detection. IEEE/ACM Trans. Audio Speech Lang. Process. 2022, 30, 1749–1762. [Google Scholar] [CrossRef]
  12. Nguyen, T.N.T.; Jones, D.L.; Watcharasupat, K.N.; Phan, H.; Gan, W.S. SALSA-Lite: A fast and effective feature for polyphonic sound event localization and detection with microphone arrays. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 716–720. [Google Scholar]
  13. Shimada, K.; Koyama, Y.; Takahashi, N.; Takahashi, S.; Mitsufuji, Y. ACCDOA: Activity-coupled cartesian direction of arrival representation for sound event localization and detection. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 915–919. [Google Scholar]
  14. Shimada, K.; Koyama, Y.; Takahashi, S.; Takahashi, N.; Tsunoo, E.; Mitsufuji, Y. Multi-ACCDOA: Localizing and Detecting Overlapping Sounds from the Same Class with Auxiliary Duplicating Permutation Invariant Training. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 316–320. [Google Scholar]
  15. Adavanne, S.; Politis, A.; Nikunen, J.; Virtanen, T. Sound event localization and detection of overlapping sources using convolutional recurrent neural networks. IEEE J. Sel. Top. Signal Process. 2018, 13, 34–48. [Google Scholar] [CrossRef] [Green Version]
  16. Cao, Y.; Kong, Q.; Iqbal, T.; An, F.; Wang, W.; Plumbley, M.D. Polyphonic Sound Event Detection and Localization Using a Two-Stage strategy. In Proceedings of the Detection and Classification of Acoustic Scenes and Events Workshop (DCASE), New York, NY, USA, 25–26 October 2019; pp. 30–34. [Google Scholar]
  17. Politis, A.; Shimada, K.; Sudarsanam, P.; Adavanne, S.; Krause, D.; Koyama, Y.; Takahashi, N.; Takahashi, S.; Mitsufuji, Y.; Virtanen, T. STARSS22: A dataset of spatial recordings of real scenes with spatiotemporal annotations of sound events. arXiv 2022, arXiv:2206.01948. [Google Scholar]
  18. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  19. Politis, A.; Mesaros, A.; Adavanne, S.; Heittola, T.; Virtanen, T. Overview and evaluation of sound event localization and detection in DCASE 2019. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 29, 684–698. [Google Scholar] [CrossRef]
  20. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  21. Ahonen, J.; Pulkki, V.; Lokki, T. Teleconference application and B-format microphone array for directional audio coding. In Proceedings of the Audio Engineering Society Conference, Saariselka, Finland, 15–17 March 2007. [Google Scholar]
  22. Yasuda, M.; Koizumi, Y.; Saito, S.; Uematsu, H.; Imoto, K. Sound event localization based on sound intensity vector refined by DNN-based denoising and source separation. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 651–655. [Google Scholar]
  23. Dozat, T. Incorporating Nesterov Momentum into Adam. In Proceedings of the International Conference on Learning Representations (ICLR), San Juan, PR, USA, 2–4 May 2016; pp. 1–4. [Google Scholar]
  24. Loshchilov, I.; Hutter, F. SGDR: Stochastic gradient descent with warm restarts. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017; pp. 1–16. [Google Scholar]
  25. Park, D.S.; Chan, W.; Zhang, Y.; Chiu, C.C.; Zoph, B.; Cubuk, E.D.; Le, Q.V. Specaugment: A simple data augmentation method for automatic speech recognition. In Proceedings of the Interspeech, Graz, Austria, 15–19 September 2019; pp. 2613–2617. [Google Scholar]
  26. Dietterich, T.G. Ensemble methods in machine learning. In Proceedings of the International Workshop on Multiple Classifier Systems, Cagliari, Italy, 21–23 June 2000; pp. 1–15. [Google Scholar]
  27. Dcase.Community. Sound Event Localization and Detection Evaluated in Real Spatial Sound Scenes—DCASE. 2022. Available online: https://dcase.community/challenge2022/task-sound-event-localization-and-detection-evaluated-in-real-spatial-sound-scenes-results (accessed on 31 August 2022).
Figure 1. Architecture of the proposed RCNN and transformer-based model.
Figure 1. Architecture of the proposed RCNN and transformer-based model.
Sensors 23 03398 g001
Figure 2. Block diagram of a residual convolutional block comprising a CNN layer with residual connections, batch normalization, and max pooling layer.
Figure 2. Block diagram of a residual convolutional block comprising a CNN layer with residual connections, batch normalization, and max pooling layer.
Sensors 23 03398 g002
Figure 3. Example of time and frequency masking.
Figure 3. Example of time and frequency masking.
Sensors 23 03398 g003
Table 1. Comparison of the trained model performances depending on the training methods. The values that have the best performance are bolded.
Table 1. Comparison of the trained model performances depending on the training methods. The values that have the best performance are bolded.
MethodValidation
(Only Real Data)
Error RateF1-ScoreLELRSELD Score
Method 1
(training with synthetic data)
0.810.1939.9453.00.578
Method 2
(training with synthetic data,
fine tuning with real data)
0.630.2739.4955.00.507
Method 3
(using multi-generator)
0.700.3123.1762.00.475
Table 2. Performance comparison of the trained single models. The values that have the best performance are bolded.
Table 2. Performance comparison of the trained single models. The values that have the best performance are bolded.
MethodValidation
(Only Real Data)
Error RateF1-ScoreLELRSELD Score
Model 1
(Transformer encoder, 512 dim, MSE)
0.670.3324.6361.00.467
Model 2
(Transformer encoder, 2048 dim, L1)
0.690.3024.0358.00.486
Model 3
(Transformer encoder, 512 dim, L1)
0.650.3124.8159.00.472
Model 4
(Model 1 additional training)
0.660.3025.7262.00.471
Model 5
(Bi-LSTM, MSE)
0.650.3122.2453.00.483
Model 6
(Transformer encoder, 256 dim, MSE)
0.650.3130.0159.00.479
Model 7
(Model 3 additional training)
0.650.3126.4162.00.467
Table 3. Performance comparison of the ensemble models. The values that have the best performance are bolded.
Table 3. Performance comparison of the ensemble models. The values that have the best performance are bolded.
Ensemble ModelValidation
(Only Real Data)
Error RateF1-ScoreLELRSELD Score
Model I0.590.3520.757.00.446
Model II0.590.3424.858.00.452
Model III0.590.3533.857.00.464
Model IV0.590.3423.059.00.447
Table 4. Comparison of the performance of the proposed method with that of the baseline model. The values that have the best performance are bolded.
Table 4. Comparison of the performance of the proposed method with that of the baseline model. The values that have the best performance are bolded.
Challenge ResultDev setEval Set
Error RateF1-ScoreLELRSELD ScoreError RateF1-ScoreLELRSELD Score
Baseline of DCASE 20220.710.2129.346.00.5510.610.2422.951.40.496
Prposed method
(Single Model 7)
0.650.3126.4162.00.467-----
Prposed method
(Ensemble, Model III)
0.590.3533.857.00.4640.590.3119.850.70.471
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shin, Y.; Chun, C. Sound Event Localization and Detection Using Imbalanced Real and Synthetic Data via Multi-Generator. Sensors 2023, 23, 3398. https://doi.org/10.3390/s23073398

AMA Style

Shin Y, Chun C. Sound Event Localization and Detection Using Imbalanced Real and Synthetic Data via Multi-Generator. Sensors. 2023; 23(7):3398. https://doi.org/10.3390/s23073398

Chicago/Turabian Style

Shin, Yeongseo, and Chanjun Chun. 2023. "Sound Event Localization and Detection Using Imbalanced Real and Synthetic Data via Multi-Generator" Sensors 23, no. 7: 3398. https://doi.org/10.3390/s23073398

APA Style

Shin, Y., & Chun, C. (2023). Sound Event Localization and Detection Using Imbalanced Real and Synthetic Data via Multi-Generator. Sensors, 23(7), 3398. https://doi.org/10.3390/s23073398

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop