Next Article in Journal
GDReBase: A Knowledge Base for Relations between Human Gut Microbes and Diseases Based on Deep Learning
Next Article in Special Issue
Exploiting the Rolling Shutter Read-Out Time for ENF-Based Camera Identification
Previous Article in Journal
City Ditty: An Immersive Soundscape Sketchpad for Professionals of the Built Environment
Previous Article in Special Issue
Blockchain Integration in the Era of Industrial Metaverse
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Prediction of the Long-Term Environmental Acoustic Pattern of a City Location Using Short-Term Sound Pressure Level Measurements

Research Group in Advanced Telecommunications (GRITA), Universidad Católica de Murcia (UCAM), 30107 Guadalupe, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(3), 1613; https://doi.org/10.3390/app13031613
Submission received: 30 December 2022 / Revised: 20 January 2023 / Accepted: 24 January 2023 / Published: 27 January 2023

Abstract

:
To manage noise pollution, cities use monitoring systems over wireless acoustic sensor networks. These networks are mainly composed of fixed-location sound pressure level sensors deployed in outdoor sites of the city for long-term monitoring. However, due to high economic and human resource costs, it is not feasible to deploy fixed metering stations on every street in a city. Therefore, these continuous measurements are usually complemented with short-term measurements at different selected locations, which are carried out by acoustic sensors mounted on vehicles or at street level. In this research, the application of artificial neural networks is proposed for estimation of the long-term environmental acoustic pattern of a location based on the information collected during a short time period. An evaluation has been carried out through a comparison of eight artificial neural network architectures using real data from the acoustic sensor network of Barcelona, Spain, showing higher accuracy in prediction when the complexity of the model increases. Moreover, time slots with better performance can be detected, helping city managers to deploy temporal stations optimally.

1. Introduction

Sound waves, or noise emissions, are one of the pollutants that urban citizens are most concerned about [1]. To identify, measure, and determine exposure to environmental noise, city rulers are developing data strategies to capture, transform, and analyze information using Internet of Things (IoT) and big data technologies.
European Directive 2002/49/EC aims to establish a common approach for the assessment and management of environmental noise in order to standardize procedures and metrics. The goal is to avoid, prevent, and reduce harmful effects, including annoyance, for citizens as a result of exposure to different noise sources [2]. The directive specifically promotes agglomerations of people such as cities or clusters of cities to create strategic noise mapping (SNM) and then share the findings with the public. Additionally, the outcome of these noise maps has led to the formation of action plans for noise reduction in areas identified as having high noise exposure (noise exposure protection zones).
More recently, numerous large cities have begun to deploy Wireless Acoustic Sensor Networks (WASN), which are based on IoT technologies [3], in order to gather noise data that can be analyzed and utilized to update SNM and action plans. These WASNs are usually made up of two different types of stations: fixed-location sensors for long-term monitoring, and temporal location sensors for short-term monitoring. The latter can take the form of temporarily deployed sensors, instrumented vehicles with an acoustic sensor together with a geopositioning system to locate the measurement, or regular sound measure devices known as sonometers [4].
While fixed stations remain at one place all their lifetime, allowing the continuous monitoring of noise levels to identify trends and seasonality, temporal stations are placed in a particular site during an established period of time (minutes, hours, or days) to measure the acoustic soundfield by gathering short-term data.
To recognize environmental acoustic patterns or behaviors, using the average or median of noise indicators for the overall assessment period, generally at least one year, is recommended by the Directive [2]. Therefore, short-term data are not usually considered due to the lack of capability to capture seasonality components such as holidays or weekends. After the analysis of previously enunciated long-term statistics, two principal types of environmental acoustic patterns are usually recognized: special regime areas and quiet areas. Special regime areas include locations where the noise indicator exceeds a high threshold, while quiet areas include locations where the noise indicator is below a low threshold. Although other patterns with complex behavior can exist, advanced statistical techniques are required to recognize them.
In a number of out prior works [5,6], we applied unsupervised learning techniques to group the nodes of a WASN in clusters with the same behavior and recognize complex patterns on this basis. These complex patterns can provide insights to city managers for establishing personalized strategies and defining new acoustic areas. In the current research, the application of a supervised machine learning technique, Artificial Neural Network (ANN), is proposed to predict the long-term acoustic behavior group to which a location belongs by means of short-term measurements. In this way, temporal stations can be used by city managers to identify the environmental acoustic pattern of a site, enhancing the value of the WASN and improving the SNM.
During last few years, machine learning algorithms have been considered in a number of studies involving environmental acoustic data captured by WASNs.
Many of the studies found in the literature use supervised machine learning approaches to analyze audio signals. In New York City, a comprehensive dataset [7] of labeled audio recordings was generated utilizing a WASN [8] for the design and assessment of machine learning techniques. This dataset was used to perform methods for both identification [9] and categorization [10] of acoustic scenes and events. Recently, a deep learning structure was created using this dataset [11] to retrieve urban sound events such as car horns and human speech from multi-label audio recordings. In a European project called DYNAMAP [12], multiple machine learning techniques were evaluated for detecting [13,14,15,16] abnormal noise sources such as birds, bikes, vehicles with heavy loads passing over rough surfaces, horn vehicle noise, music in a car or in the street, ambulance sirens, airplanes, thunder storms, etc., in order to eliminate events unrelated to road traffic noise and create a noise map. In addition to the previously mentioned methods, other techniques utilizing supervised machine learning have been utilized for classifying sound sources. Maijala et al. [17] introduced a pattern classification algorithm that used Mel-frequency cepstral coefficients as features to determine the primary noise source in the acoustic environment. In this research, two types of supervised classifiers, namely, artificial neural networks with two hidden layers of 10, 30, 50, or 100 neurons and a Gaussian mixture model, were compared. Ye et al. [18] introduced an aggregation scheme combining local features and short-term sound recording features with long-term descriptive statistics to create a deep convolutional neural network for classifying urban sound events.
Regarding machine learning techniques for sound pressure level and acoustic pattern prediction, a number of studies have been published within the last few years. Das et al. [19,20] proposed an ANN architecture with only one hidden layer to predict annoyance levels of traffic noise. The architecture complexity of the trained ANNs (see Section 2.4 for details about this definition) were N e t _ 3 _ 1 , with six variables in the input layer for the first study and N e t _ 1 _ 1 , N e t _ 2 _ 1 , N e t _ 3 _ 1 , N e t _ 4 _ 1 , N e t _ 5 _ 1 with five variables in the input layer for the second study. By utilizing short-term data and concentrating on traffic noise, unsupervised machine learning techniques such as dimensionality reduction and clustering were employed to optimize the location and quantity of monitoring sites [21]. A separate publication [22] presented a methodology for more efficiently estimating day-period and night-period sound pressure levels on urban roads in Milan, Italy in comparison to the legislative road classification by using equivalent sound pressure levels of a 1-h period from a 24 h measurement campaign. Subsequently, in order to link each street in the area of examination to one of the two noise profiles found through clustering, several non-acoustic parameters were examined [22]. In another recent study, the intermittency ratio indicator was paired with the equivalent sound pressure level of a 1-h time frame in order to improve the categorization of different types of streets within the two identified clusters [23].
Regarding the identification of the long-term environmental acoustic pattern of a city, Torija et al. [24] investigated the necessary stabilization time, short-term variability, and impulsiveness of the sound pressure level to accurately characterize the temporal composition of urban soundscapes. The authors used data from sound level meters to analyze sound pressure levels in urban environments, and found that a stabilization time of at least 30 min was required to obtain reliable measurements of sound pressure level. The same study suggested that measurements should be taken over a longer period of time to achieve a more accurate characterization of urban soundscapes, and that the short-term variability and impulsiveness of sound pressure levels should be considered as well. In a later study, Gajardo et al. [25] analyzed data collected from sound level meters in various urban environments and concluded that hourly averages of sound levels may not be representative of the true levels of noise exposure. Therefore, using longer measurement periods such as 24 h, to obtain more accurate representation of noise levels in urban environments was recommended. On the other hand, regarding the prediction of the equivalent sound level using short-term measurements, Brambilla et al. [26] focused on the stabilization time for road traffic noise measurements and concluded that a time of at least 10 min is necessary for reliable estimation of the equivalent sound pressure level of 1-h period; in addition, factors such as traffic volume, traffic composition, and road type can affect the required stabilization time.
An environmental acoustic pattern refers to the distribution and variation of sound levels in a specific environment. These patterns can be affected by a variety of factors, such as land use, weather conditions, and human activity. In urban environments, the environmental acoustic pattern is typically characterized by high levels of noise pollution from sources such as traffic, construction, and industrial activities. However, these noise sources create a complex and dynamic acoustic environment which is highly dependent on the time of day and location. In this research, the environmental acoustic pattern of a location refers to the classification of a location using the equivalent sound pressure level during the day, evening, and night periods over a year, as recommended by the Directive [2] and defined in Section 2.2.
The contribution of the current research is to use an unsupervised learning algorithm to estimate the corresponding environmental acoustic pattern of a location among the recognized long-term behaviors based on one-year acoustic data. This is carried out by using one-hour equivalent acoustic data to design and test algorithms based on ANNs, which are trained using shorter periods of time with a large amount of available data and require parallel processing to optimize the data pipelines.
This rest of this paper is structured as follows. The datasets, algorithms, and methodology used for training and testing the models are presented in Section 2. Then, in Section 3, the results obtained from the analysis are displayed and discussed. Finally, Section 4 summarizes the main conclusions of this work.

2. Materials and Methods

In this section, the materials and methods applied during this research are presented. The data source containing the sound pressure level values of the sites and the collection methodology are described in Section 2.1. The environmental acoustic patterns recognized in a previous work [5] are summarized in Section 2.2. Next, the curated short-term datasets used in this research and their transformations are detailed in Section 2.3. Section 2.4 presents the machine learning models that have been trained and evaluated in this work. Finally, the metrics used in the evaluation of the models are defined in Section 2.5.
The data preparation, transformation, analysis, modeling, and visualization were executed utilizing the Statistical Programming Language R [27], which involved the integration of a local environment using R version 4.2.1 with a free cloud-based environment provided by Posit Cloud using R version 4.2.2. The scripts applied in this research are available at the Github repository https://github.com/AntonioPL/BCN_Noise (accessed on 6 January 2023). In order to ensure the reproducibility of the research, the seed was fixed using the R function set.seed() in every task that incorporated a random step.

2.1. Data Source

The historical data used in this research contain sound pressure level values from 70 fixed acoustic nodes deployed in Barcelona, Spain, to build a WASN, as described in publications by Camps et al. [28] and Farres et al. [29]. The map in Figure 1 shows the widespread distribution of the nodes in the whole city.
These fixed acoustic nodes are equipped with remote Cesva TA120 [30] sonometers, which capture sound pressure levels continuously 24 × 7 (24 h and 7 days a week) and every minute send the A-frequency weighting equivalent sound pressure level of a 1-min period, denoted as L Aeq 1 m , as defined in Equation (1) following ISO1996-2 [31]:
L eq 1 m = 10 · log 1 60 t 0 t 0 + 60 p 2 ( t ) p 0 2 d t dBA ,
where [ t 0 , t 0 + 60 ] is a 1-min interval beginning at time t 0 , p ( t ) is the sound pressure level at time t in Pascal pressure units ( P a ), and p 0 = 20 μ pa is the sound pressure reference value.
These data are captured every minute and stored in the central data storage [28], where transformation are performed before the data are absorbed into the smart city platform of BCN called Plataforma de Sensors i Actuadors de Barcelona [32].
More than 97 million of L Aeq 1 m records captured by BCN city council in the full years from 2018 to 2020 for the 70 nodes were exported from the smart city platform in 73 Excel files with wide data format for use in this research. These files contain sound pressure level values of every minute for every node in the described period. It is worth noting that there were a number of null values due to sensor errors and maintenance periods; these were removed during the curation phase. Basic statistics and available records from the nodes can be found in [5].

2.2. Environmental Acoustic Patterns

To evaluate the environmental acoustic behavior of a site, European Directive 2002/49/EC [2] recommends the use of the L Aeq T indicators corresponding to day, evening, and night periods over 24 h for a specific station on every day during one year, denoted as L d 1 y , L e 1 y and L n 1 y , respectively, and the overall assessment period noise indicator DEN (day, evening, night), represented by L den . To take into account the temporal variability of the sound pressure level values during the different periods of the day, the yearly standard deviation of L den 1 d , denoted s d 1 y ( L den 1 d ), has been proposed to describe the variability or volatility of the sound pressure level of the nodes during a year [5,6]. In these previous works of ours, four environmental acoustic patterns were recognized using these four noise indicators, calculated from the described dataset as inputs of several unsupervised learning techniques. Therefore, the nodes of BCN’s WASN were classified into one of these patterns, and are shown in Figure 1 in different colors together with their locations.
Table 1 shows the average values of the four previously defined noise indicators for the different pattern categories that allow the description of their behaviours. Analytically, there are three pattern categories in which day and evening sound pressure level values are similar and in which both are higher than night sound pressure level values by a statistical significance amount. The first pattern category, shown by the 23 nodes with the black color tag in Figure 1, has higher sound pressure values ( L d 1 y , L e 1 y and L n 1 y ) than the second pattern category, shown by the 27 nodes with the magenta color tag in Figure 1. Both categories have higher sound pressure values than the fourth category, which includes the 11 nodes shown with the brown color tag in Figure 1. Therefore, these pattern categories represent nodes with high, medium, and low values of sound pressure with the described behavior. Moreover, a negative correlation between sound pressure levels indicators and variability ( s d 1 y ( L den 1 d )) can be observed, i.e., the fourth category is the one with the highest variability, followed by the second and first categories, in this order. The remaining third category, indicated by the blue color tag, contains nine nodes. This category shows a different behavior than the others, with evening sound pressure level value being higher than the other periods, which all have similar values. Moreover, this third category presents the highest variability of all categories.
These environmental acoustic patterns can be contextualized by features such as the type of roads, use of the area, and noise sources. Interpretation of these characteristics allows for a deeper understanding and appreciation of the behavior patterns of different city areas. This can be valuable for residents, tourists, businesses, and city managers.
Behavior Category 1 groups the locations related to main routes of road traffic, in particular, major thoroughfares and intersections with very high road traffic intensity. Noise levels are relatively consistent during the day, with peak noise levels occurring during rush hour when traffic is heaviest. At night, noise levels decrease somewhat due to reduced traffic volume, though they are still significantly higher than in quiet residential areas. The sound pressure level is high and fluctuation is low, as shown in Table 1.
Behavior Category 2 groups those locations related to the regular areas in a city, which typically include medium-density residential, commercial, and office building areas along with public spaces such as parks, squares, and sidewalks. The noise pollution is moderate to high, with a wide range of noise sources. During the day the noise level is relatively consistent, with peak noise levels occurring during peak hours of activity such as rush hour and lunchtime. At night, noise levels decrease somewhat due to reduced activity, though they remain higher than in quiet residential areas. In this category, the sound pressure level is moderate to high and the fluctuation is moderate, as shown in Table 1.
Behavior Category 3 groups the locations related to shopping, entertainment, and nightlife activity. The noise pollution is high throughout the day, evening, and night due to the high level of human activity and mix of commercial and entertainment venues. During the day, noise levels are high and relatively consistent, with peak noise levels occurring during peak hours of shopping and entertainment activity. In the evening, noise levels continue to be high, and are more fluctuating, with an increase in human activity as people go out for entertainment and nightlife. At night, noise levels are high and fluctuating, with an increase in human activity in nightlife venues such as bars, clubs, and restaurants. In this category, both the sound pressure level and fluctuation are high, as is shown in Table 1.
Finally, Behavior category 4 is related to quiet residential areas in a city, where the noise pollution is low during the day, evening, and night periods. These areas are characterized by lower levels of human activity and fewer noise sources, providing a relatively peaceful and quiet environment for residents. Noise levels are low during the day, with occasional spikes from passing vehicles and aircraft or distant construction and maintenance work. In the evening noise levels decrease even further due to lower traffic and other human activities. Noise levels at night decrease significantly, as expected. However, occasional high noise level events, e.g., from passing vehicles or aircraft, can explain the elevated fluctuation of the sound pressure level shown in Table 1.
In the current research, short-period measurement data are used to estimate the corresponding behavior recognized using long-term data, i.e., the environmental acoustic pattern category is the output variable of the proposed supervised learning algorithm.

2.3. Curated Modelling Datasets

To train and evaluate the machine learning models, 24 short-term period curated datasets were prepared and denoted using numbers from 0 to 23, sequentially corresponding with every one-hour time slot of the day. Each instance contains 60 sound pressure level values for a particular node on a specific date at the fixed hour, i.e., dataset number X contains all the sound pressure level values captured from X:00 until X:59 in hh:mm format for every node at any date from January 2018 until December 2020. Table 2 shows the distribution of these 24 datasets, detailing the amount of available, valid, and null instances and the average instances per node for every dataset. As a summary, there are 1,621,145 valid instances, which is 93.66% of the total available instances, with an average of 23,159 instances per node.
The following tasks were applied to these curated datasets to train and test the models implemented with the machine learning technique described in Section 2.4. First, instances with null values were removed. Then, every dataset was randomly split into two subsets, called the training and and test sets. The training subset contained 80% of the curated dataset instances, and was the input for model training, while the test subset contained 20% of the curated dataset instances, and was used to evaluate the models.

2.4. Artificial Neural Networks

To estimate the target variable or pattern category described in Section 2.2 using the curated datasets described in Section 2.3, supervised learning algorithms were considered. In particular, several feed-forward multilayer Artificial Neural Networks were built.
A feed-forward multilayer ANN is a mathematical model composed of elements called neurons [33] grouped by layers and relationships between the elements of a layer with the elements of the previous layer by activation functions, as displayed in Figure 2.
There are three different types of layers, as can be seen in Figure 2. The input layer, represented by the green circles, is fed the input dataset, meaning that the size of this layer must be equal to the size of every instance in the dataset (60 in this work). The output layer, represented by the blue circles, is populated by the output variable to be estimated, meaning that the size must match the number of categories described in Section 2.2 (four in this work). Moreover, there are one or more intermediate layers, usually known as hidden layers, which are represented by the orange circles in Figure 2. The hidden layers can have different sizes. To fit the parameters of these layers to the data, a backpropagation algorithm [35] was used, with normalized exponential (softmax) as the activation function of the output layer and rectified linear unit (ReLU) as the activation function of the hidden layers.
In this article, the notation N e t _ X 1 _ _ X n _ Y represents a feed-forward ANN architecture with n hidden layers with size X i for hidden layer i and size Y for the output layer. In particular, eight architectures with a different number of hidden layers and different amounts of neurons in the layers were trained on the 24 datasets described in Section 2.1, resulting in 192 models used in the comparison. The eight detailed architectures are the following: N e t _ 16 _ 4 , N e t _ 32 _ 4 , N e t _ 64 _ 4 , N e t _ 16 _ 16 _ 4 , N e t _ 32 _ 32 _ 4 , N e t _ 64 _ 32 _ 4 , N e t _ 16 _ 16 _ 16 _ 4 , and N e t _ 64 _ 32 _ 16 _ 4 . Note that all the models have four neurons in the output layer.

2.5. Performance Metrics

In this study, the classification performance of the trained models was measured globally and for each category using three different metrics: Accuracy, F1-Score, and Balanced Accuracy.
Accuracy, the percentage of elements correctly labeled by the model, was calculated using Equation (2) to evaluate the global performance of the models:
A c c u r a c y = i = 1 C T P i N ,
where N is the quantity of elements, C is the number of categories, and T P i is the quantity of elements belonging to real category i correctly labeled by the model as category i for every category i. By definition, the Accuracy is a real number between 0 and 1. A high Accuracy indicates good global performance of the model, with the best result reaching 1 when all the elements are correctly labeled by the model.
On the other hand, F1-Score and Balanced Accuracy were calculated for every category i to evaluate the performance of the model over every category. Thr F1-Score is the harmonic mean of the trade-off metrics, precision and recall, as defined in Equation (3), for every category i:
F 1 S c o r e i = 2 P r e c i s i o n i R e c a l l i P r e c i s i o n i + R e c a l l i ,
where
P r e c i s i o n i = T P i T P i + F P i , a n d
R e c a l l i = T P i T P i + F N i .
F N i is the quantity of elements belonging to real category i incorrectly labeled by the model as a category different from i, while F P i is the quantity of elements not belonging to real category i incorrectly labeled by the model as category i. The maximum possible F1-score value is 1, which indicates perfect precision and recall, while the minimum possible value is 0, which is the case if either precision or recall is zero.
Second, Balanced Accuracy is the arithmetical mean of the trade-off metrics, sensitivity and specificity, as shown in Equation (6):
B a l a n c e d   A c c u r a c y i = S e n s i t i v i t y i + S p e c i f i c i t y i 2 ,
where
S e n s i t i v i t y i = T P i T P i + F N i ,
S p e c i f i c i t y i = T N i T N i + F P i ,
and F N i is the quantity of elements belonging to real category i incorrectly labeled by the model as belonging to a category other than i. The highest possible Balanced Accuracy value is 1, indicating perfect Sensitivity and Specificity, and the lowest possible value is 0, which is the case if both Sensitivity and Specificity are zero.
In summary, when evaluating a particular category, the closer the Balanced Accuracy and F1-Score are to 1, the better the model can correctly classify observations.

3. Results and Discussion

This section presents and discusses the results obtained in the comparison between trained models for the different ANN architectures. This evaluation was carried out in three approaches: global performance, time slot, and environmental acoustic pattern.
First, the performance of the different models on the test datasets was calculated using Accuracy as a global metric. Table 3 shows the Accuracy of the 192 trained models for the eight ANNs defined in Section 2.4 on the test subsets of the 24 datasets representing each hourly time slot, where time slot X corresponds to the interval from X:00 hour to X:59 hour, as defined in Section 2.3.
The global performance of the models depends on the time slot and the model, as expected; N e t _ 64 _ 32 _ 4 , from 21:00 to 21:59, shows the highest Accuracy at 0.6943, resulting in the best combination of architecture and hourly time slot. This is a particular insight, very valuable for city managers; however, this asseveration is difficult to generalize for other cities. Analyzing these results, a discussion is provided in the following paragraphs in order to obtain more general conclusions.
Due to the existence of four categories, adopting a random model supposes an Accuracy of 0.25 in each of them. As shown in Table 3, in general, all models across every hour exceed the random model except one. As the pattern categories are not equally distributed (see Table 1), a baseline model could be the selection of the most representative category with an Accuracy of 0.39 (=27/70). Even though the Accuracy of the models ranges from 0.150 to 0.694, 160 of the 192 models (83%) have an Accuracy higher than the baseline model. In addition, a one-sided parametric hypotheses testing was carried out with the hypotheses represented in Equation (9):
H 0 : μ 0.39 H 1 : μ > 0.39
Considering the central limit theorem, the test statistic follows a Student’s t-distribution with 191 degrees of freedom, and the estimator of the test is 18.929, equivalent to a p v a l u e < 2.2 10 16 . Therefore, the null hypothesis is rejected, leading to the conclusion that the improvement when using an ANN to estimate the long-term environmental acoustic pattern of a spot based on short-term data is statistically significant.
Regarding the optimum hourly time slot to capture data that best represent the long-term pattern, Table 3 shows that on average every hourly time slot improves the baseline model; the better hourly time slots to predict environmental acoustic behaviors are 14, 17, 20, 4, and 21, in which the averaged Accuracy is higher than 0.55. Finally, the worst time slots to capture data are 7, 10, 8, 12, 22, and 11, in which the Accuracy is lower than 0.49.
However, outliers decrease the representativeness of the mean value. Therefore, a median Accuracy analysis was performed to minimize the impact of outliers in the above results. The Accuracy distribution for each hourly time slot ordered by the median of the Accuracy is shown in Figure 3. These hourly time slots are colored yellow for daytime periods (from 7:00 a.m. to 7:00 p.m.), orange for evening periods (from 7:00 p.m. to 11:00 p.m.), and gray for night periods (from 11:00 p.m. to 07:00 a.m.), as defined in Directive 2002/49/EC [2].
Figure 3 shows that 21, 18, and 17 are the best hourly time slots, in this order. On the other hand, the worst time slots are 22, 8, and 12. The top seven most accurate hourly time slots are in the interval 14 to 21. From this interval, only 19 is not in the ranking, falling to the twelfth position. Therefore, the period from 14:00 to 22:00 is recommended to capture data and estimate the location acoustic pattern.
Next, the impact of the complexity of the ANN architecture on the performance in the classification was analyzed. Figure 4 shows a comparison of the distribution of the Accuracy performance metric. The fill color represents the number of hidden layers, with light blue, blue, and dark blue standing for 1, 2, and 3, respectively. Although the models with a higher quantity of hidden layers have the highest average Accuracy, the amount of neurons in the layers does not significantly affect Accuracy.
Finally, the performance of the models in relation to each pattern category, as described in Section 2.2, was evaluated using Balanced Accuracy and F1-score.
Figure 5 shows that Pattern Category 1 has the best performance on average (0.63 F1-Score and 0.75 Balanced Accuracy), followed by Categories 2 (0.53 F1-Score and 0.59 Balanced Accuracy) and 4 (0.32 F1-Score and 0.60 Balanced Accuracy). Category 3 is the most difficult to predict (0.07 F1-Score and 0.50 Balanced Accuracy). This observation is inverse correlated with the s d 2019 ( L den 1 d ) of each pattern category, meaning that its higher the volatility makes this category harder to predict. It is important to note that one-hour time slots are used as a short-term measurement period; thus, improvements in the predictions for Pattern Category 3 can be achieved by combining data from two or more hourly time slots.
To obtain further insights into the prediction ability, an hourly time slot F1-Score performance comparison was carried out for every category.
Figure 6 shows similar trends in Categories 1 and 2 regarding median F1-Score for each hourly time slot. Most of these have low variability, meaning that in general any period could be used to predict these environmental pattern categories. On the contrary, Category 3 has low performance at all time slots, improving in the nightly period, but not enough to be confident in the prediction. Therefore, other strategies, for example, increasing the size of the time period or including several hourly time slots as input data, should be considered in future works. Finally, Category 4 presents a wide range of performance values, highlighting the nightly period 21:00–02:00 as the best period. Moreover, the variability of the F1-Score distribution for Category 4 is the highest of all categories.

4. Conclusions

In this paper, we carried out an evaluation of the suitability of predicting the long-term environmental acoustic pattern of a position based on information collected in a short-term interval using artificial neural networks. For this, we used a dataset with sound pressure level values from the city of Barcelona, Spain, captured with a wireless acoustic sensor network. Using several performance metrics, we performed a comparison between 192 models designed with eight different architectures and trained using hourly sound pressure level datasets.
In general, the results show that artificial neural networks can classify short-term acoustic data into one of several recognized long-term environmental acoustic patterns. From a global perspective, models with higher quantity of hidden layers have better performance, even though this performance is not affected by the amount of neurons, and the performance increases if the data are gathered in an hourly time slot included in the interval from 14:00 to 22:00. Regarding particular environmental acoustic patterns, those with lower sound pressure level variability are easier to estimate using hourly sound pressure level measurements.
The provided insights are crucial to define the data collection methodology in order to assure the most accurate pattern category prediction and avoid bias created by stable routines with temporal stations. Moreover, it is recommended to capture data at the same time slot in different locations, as this improves recognition of the specific environmental acoustic behavior of a place.

Author Contributions

Conceptualization, A.P. and J.M.N.; data curation, A.P.; investigation, A.P. and J.M.N.; methodology, A.P.; software, A.P.; modeling A.P.; supervision, J.M.N.; validation, J.M.N. and A.P.; writing—original draft, A.P. and J.M.N.; writing—review and editing, A.P. and J.M.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by MCIN/AEI/10.13039/501100011033 under grant PID2020-112827GB-I00.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data source described in this paper has been shared by the Departament d’Avaluació i Gestió Ambiental in Ajuntament de Barcelona, which owns and has rights over all data.

Acknowledgments

The authors acknowledge Júlia Camps Farrés and Alejandro Aparicio Estrems for sharing the data.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANNArtificial Neural Networks
BCNBarcelona City (Spain)
IoTInternet of Things
GPSGlobal Positioning System
ReLURectified Linear Unit
SNNStrategic Noise Map
WASNWireless Acoustic Sensor Network

References

  1. Zipf, L.; Primack, R.B.; Rothendler, M. Citizen scientists and university students monitor noise pollution in cities and protected areas with smartphones. PLoS ONE 2020, 15, e0236785. [Google Scholar] [CrossRef] [PubMed]
  2. European Commission. Directive 2002/49/EC of the European Parliament and of the Council of 25 June 2002 Relating to the Assessment and Management of Environmental Noise; European Commission: Brussels, Belgium, 2002.
  3. Zanella, A.; Bui, N.; Castellani, A.; Vangelista, L.; Zorzi, M. Internet of Things for Smart Cities. IEEE Internet Things J. 2014, 1, 22–32. [Google Scholar] [CrossRef]
  4. Garrido, J.C.; Mosquera, B.M.; Echarte, J.; Sanz, R. Management Noise Network of Madrid City Council. In InterNoise19, Proceedings of the Inter-Noise and Noise-Con Congress Conference, Madrid, Spain, 16–19 June 2019; Institute of Noise Control Engineering: Madrid, Spain, 2019; pp. 996–1997. [Google Scholar]
  5. Pita, A.; Rodriguez, F.J.; Navarro, J.M. Cluster Analysis of Urban Acoustic Environments on Barcelona Sensor Network Data. Int. J. Environ. Res. Public Health 2021, 18, 8271. [Google Scholar] [CrossRef]
  6. Pita, A.; Rodriguez, F.J.; Navarro, J.M. Analysis and Evaluation of Clustering Techniques Applied to Wireless Acoustics Sensor Network Data. Appl. Sci. 2022, 12, 8550. [Google Scholar] [CrossRef]
  7. Cartwright, M.; Mendez, A.E.M.; Cramer, J.; Lostanlen, V.; Dove, G.; Wu, H.H.; Bello, J. Sonyc urban sound tagging (sonyc-ust): A multilabel data-set from an urban acoustic sensor network. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2019, DCASE19, New York, NY, USA, 25–26 October 2019; pp. 35–39. [Google Scholar]
  8. Bello, J.P.; Silva, C.; Nov, O.; Dubois, R.L.; Arora, A.; Salamon, J.; Doraiswamy, H. SONYC: A system for monitoring, analyzing, and mitigating urban noise pollution. Commun. ACM 2019, 62, 68–77. [Google Scholar] [CrossRef] [Green Version]
  9. Wang, Y.; Salamon, J.; Bryan, N.J.; Bello, J.P. Few-shot sound event detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2020), Barcelona, Spain, 4–8 May 2020; pp. 81–85. [Google Scholar]
  10. Salamon, J.; Bello, J. Deep convolutional neural networks and data augmentation for environmental sound classification. IEEE Signal Process. Lett. 2017, 24, 279–283. [Google Scholar] [CrossRef]
  11. Fan, J.; Nichols, E.; Tompkins, D.; Méndez, A.E.M.; Elizalde, B.; Pasquier, P. Multi-Label Sound Event Retrieval Using A Deep Learning-Based Siamese Structure with A Pairwise Presence Matrix. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2020), Barcelona, Spain, 4–8 May 2020; pp. 3482–3486. [Google Scholar]
  12. Bellucci, P.; Peruzzi, L.; Zambon, G. LIFE DYNAMAP project: The case study of Rome. Appl. Acoust. 2017, 117, 193–206. [Google Scholar] [CrossRef]
  13. Alsina-Pagès, R.M.; Alías, F.; Socoró, J.C.; Orga, F. Detection of anomalous noise events on low-capacity acoustic nodes for dynamic road traffic noise mapping within an hybrid WASN. Sensors 2018, 18, 1272. [Google Scholar] [CrossRef] [Green Version]
  14. Socoró, J.C.; Albiol, X.; Sevillano, X.; Alias, F. Analysis and automatic detection of anomalous noise events in real recordings of road traffic noise for the LIFE DYNAMAP project. In Proceedings of the Inter-Noise and Noise-Con Congress Conference (InterNoise16), Hamburg, Germany, 21–24 August 2016; pp. 1879–2861. [Google Scholar]
  15. Socoró, J.C.; Alías, F.; Alsina-Pagès, R.M. An Anomalous Noise Events Detector for Dynamic Road Traffic Noise Mapping in Real-Life Urban and Suburban Environments. Sensors 2017, 17, 2323. [Google Scholar] [CrossRef]
  16. Alías, F.; Socoró, J.C. Description of Anomalous Noise Events for Reliable Dynamic Traffic Noise Mapping in Real-Life Urban and Suburban Soundscapes. Appl. Sci. 2017, 7, 146. [Google Scholar] [CrossRef] [Green Version]
  17. Maijala, P.; Shuyang, Z.; Heittola, T.; Virtanen, T. Environmental noise monitoring using source classification in sensors. Appl. Acoust. 2018, 129, 258–267. [Google Scholar] [CrossRef]
  18. Ye, J.; Kobayashi, T.; Murakawa, M. Urban sound event classification based on local and global features aggregation. Appl. Acoust. 2017, 117, 246–256. [Google Scholar] [CrossRef]
  19. Das, C.P.; Swain, B.K.; Goswami, S.; Das, M. Prediction of traffic noise induced annoyance: A two-staged SEM-Artificial Neural Network approach. Transp. Res. Part Transp. Environ. 2021, 100, 103055. [Google Scholar] [CrossRef]
  20. Das, C.; Rath, S.; Swain, B.; Goswami, S.; Das, M. Artificial Neural Network Modeling of Traffic Noise Induced Annoyance Amongst Exposed Population. Indian J. Environ. Prot. 2022, 42, 1042–1050. [Google Scholar]
  21. Zambon, G.; Benocci, R.; Brambilla, G. Cluster categorization of urban roads to optimize their noise monitoring. Environ. Monit. Assess. 2016, 188, 26. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Zambon, G.; Benocci, R.; Bisceglie, A.; Roman, H.E.; Bellucci, P. The LIFE DYNAMAP project: Towards a procedure for dynamic noise mapping in urban areas. Appl. Acoust. 2017, 124, 52–60. [Google Scholar] [CrossRef]
  23. Brambilla, G.; Benocci, R.; Confalonieri, C.; Roman, H.E.; Zambon, G. Classification of urban road traffic noise based on sound energy and eventfulness indicators. Appl. Sci. 2020, 10, 2451. [Google Scholar] [CrossRef] [Green Version]
  24. Torija, A.J.; Ruiz, D.P.; Ramos-Ridao, A. Required stabilization time, short-term variability and impulsiveness of the sound pressure level to characterize the temporal composition of urban soundscapes. Appl. Acoust. 2011, 72, 89–99. [Google Scholar] [CrossRef]
  25. Gajardo, C.P.; Barrigón Morillas, J.M. Stabilisation patterns of hourly urban sound levels. Environ. Monit. Assess. 2015, 187, 4072. [Google Scholar] [CrossRef]
  26. Brambilla, G.; Benocci, R.; Potenza, A.; Zambon, G. Stabilization Time of Running Equivalent Level LAeq for Urban Road Traffic Noise. Appl. Sci. 2023, 13, 207. [Google Scholar] [CrossRef]
  27. Available online: https://www.r-project.org/ (accessed on 6 January 2023).
  28. Camps, J. Barcelona noise monitoring network. In Proceedings of the EuroNoise, Maastricht, The Netherlands, 31 May–3 June 2015; pp. 218–220. [Google Scholar]
  29. Farrés, J.C.; Novas, J.C. Issues and challenges to improve the Barcelona Noise Monitoring Network. In Proceedings of the 11th European Congress and Exposition on Noise Control Engineering, Heraklion, Greece, 27–31 May 2018; pp. 27–31. [Google Scholar]
  30. CESVA TA120 Noise Measuring Sensor for Smart Solutions. Available online: https://www.cesva.com/en/products/sensors-terminals/TA120/ (accessed on 6 January 2023).
  31. ISO 1996-2:2017; Acoustics—Description, Measurement and Assessment of Environmental Noise—Part 2: Determination of Environmental Noise Levels. International Organization for Standardization: Geneva, Switzerland, 2017.
  32. Plataforma BCNSentilo. Available online: http://connecta.bcn.cat/connecta-catalog-web/component/map (accessed on 6 January 2023).
  33. McCulloch, W.S.; Pitts, W. A Logical Calculus of the Ideas Immanent in Nervous Activity. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  34. Dastres, R.; Soori, M. Artificial Neural Network Systems. Int. J. Imaging Robot. 2021, 21, 13–25. [Google Scholar]
  35. Werbos, P. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. Ph.D. Dissertation, Harvard University, Cambridge, MA, USA, 1974. [Google Scholar]
Figure 1. Map showing location and pattern category (indicated by different colors) of the nodes of the WASN of BCN. Category 1 in black, Category 2 in magenta, Category 3 in cyan and Category 4 in brown.
Figure 1. Map showing location and pattern category (indicated by different colors) of the nodes of the WASN of BCN. Category 1 in black, Category 2 in magenta, Category 3 in cyan and Category 4 in brown.
Applsci 13 01613 g001
Figure 2. Artificial neural network architecture (adapted from Dastres and Soori [34]).
Figure 2. Artificial neural network architecture (adapted from Dastres and Soori [34]).
Applsci 13 01613 g002
Figure 3. Statistical distributions of the Accuracy variable for each hourly time slot, represented though box and whisker plot and sorted by the median value of the Accuracy in decreasing order.
Figure 3. Statistical distributions of the Accuracy variable for each hourly time slot, represented though box and whisker plot and sorted by the median value of the Accuracy in decreasing order.
Applsci 13 01613 g003
Figure 4. Statistical distributions of the variable Accuracy for every ANN Model ordered by mean value (red circle) represented though box and whisker plots. The colors group the models by their quantity of hidden layers.
Figure 4. Statistical distributions of the variable Accuracy for every ANN Model ordered by mean value (red circle) represented though box and whisker plots. The colors group the models by their quantity of hidden layers.
Applsci 13 01613 g004
Figure 5. Statistical distributions of F1-Score (a) and Balanced Accuracy (b) performance metrics for every environmental acoustic pattern, represented as box and whisker plots.
Figure 5. Statistical distributions of F1-Score (a) and Balanced Accuracy (b) performance metrics for every environmental acoustic pattern, represented as box and whisker plots.
Applsci 13 01613 g005
Figure 6. Statistical distributions of the F1-Score variable for every hourly time slot, broken down into the four environmental acoustic behaviors (subfigures (ad) for category patterns 1, 2, 3, and 4, respectively) represented though box and whisker plot.
Figure 6. Statistical distributions of the F1-Score variable for every hourly time slot, broken down into the four environmental acoustic behaviors (subfigures (ad) for category patterns 1, 2, 3, and 4, respectively) represented though box and whisker plot.
Applsci 13 01613 g006
Table 1. Environmental acoustic pattern classification of BCN WASN (adapted from Pita, Navarro, and Rodriguez [5]).
Table 1. Environmental acoustic pattern classification of BCN WASN (adapted from Pita, Navarro, and Rodriguez [5]).
Pattern Category L d 2019 (dB) L e 2019 (dB) L n 2019 (dB) sd 2019 ( L den 1 d )NodesColor
170.7470.7966.391.5023black
266.4066.0462.282.0627magenta
366.0568.2566.713.789cyan
461.1160.5756.242.6111brown
Table 2. Hourly sound pressure level datasets.
Table 2. Hourly sound pressure level datasets.
Dataset (Hourly Time Slot)InstancesValid InstancesInstances with Nulls% Rows with NullsAverage Instances per Node
072,21368,20840055.55%974.40
172,21368,10741065.69%972.96
272,21367,74144726.19%967.73
372,21367,89243215.98%969.89
472,21367,87043436.01%969.57
572,21368,02441895.80%971.77
672,21368,03241815.79%971.89
772,08967,88542045.83%969.79
871,96767,65243156.00%966.46
971,99867,36246366.44%962.31
1071,99867,25747416.58%960.81
1172,02967,03349966.94%957.61
1272,05966,96750927.07%956.67
1372,06066,93151297.12%956.16
1471,99867,04749516.88%957.81
1572,02967,48345466.31%964.04
1672,02967,47645536.32%963.94
1772,12167,47746446.44%963.96
1872,12167,53345886.36%964.76
1972,12067,52645946.37%964.66
2072,15167,21849336.84%960.26
2172,21367,11251017.06%958.74
2272,21367,20650076.93%960.09
2372,21368,10641075.69%972.94
Table 3. Accuracy of the trained models over the 24-h time slot datasets. Different colors represent an accuracy heat-map to visually identify patterns.
Table 3. Accuracy of the trained models over the 24-h time slot datasets. Different colors represent an accuracy heat-map to visually identify patterns.
ModelHourly Time Slot
01234567891011
Net_16_40.5430.5740.5380.4470.6090.5460.5440.5520.4940.5660.5380.564
Net_32_40.5970.5490.5330.5460.5210.3930.4540.4760.3920.4590.5150.508
Net_64_40.3920.5250.4720.3930.5640.5560.5440.3900.4600.3930.5320.516
Net_16_16_40.6040.5570.4750.4920.5840.5380.5620.5020.5170.5710.3610.259
Net_32_32_40.5510.5380.5310.5560.5190.5430.4420.5140.5040.4920.4350.535
Net_64_32_40.6020.5320.5650.5300.5760.5500.5230.5530.5490.5830.4800.390
Net_16_16_16_40.5150.5220.5680.5590.5150.5980.5270.3960.3940.5240.5840.571
Net_64_32_16_40.5600.5610.5550.4980.5280.4800.4490.4880.4410.5150.3960.355
Average0.5450.5450.5300.5030.5520.5250.5060.4840.4690.5130.4800.462
ModelHourly Time Slot
121314151617181920212223
Net_16_40.6150.6350.6640.5440.5570.6410.6220.5290.6390.6830.4850.568
Net_32_40.5310.5170.4900.5780.5990.5710.6340.3920.4230.3890.6660.386
Net_64_40.4360.5510.5300.5790.4850.5400.4680.5880.5850.5840.1500.611
Net_16_16_40.4540.5370.6260.6350.6100.6300.6340.6570.4900.6150.5480.605
Net_32_32_40.3840.3390.5170.3890.3850.4520.5610.5040.5950.4970.4940.427
Net_64_32_40.6160.3850.6520.3870.6310.6180.4800.6120.6430.6950.5570.396
Net_16_16_16_40.3320.5310.5750.5860.5720.6140.6290.5560.6650.6090.3900.396
Net_64_32_16_40.3800.4970.5640.3820.3430.5330.3700.4730.4200.3340.4080.587
Average0.4680.4990.5770.5100.5230.5750.5500.5390.5580.5510.4620.497
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Navarro, J.M.; Pita, A. Machine Learning Prediction of the Long-Term Environmental Acoustic Pattern of a City Location Using Short-Term Sound Pressure Level Measurements. Appl. Sci. 2023, 13, 1613. https://doi.org/10.3390/app13031613

AMA Style

Navarro JM, Pita A. Machine Learning Prediction of the Long-Term Environmental Acoustic Pattern of a City Location Using Short-Term Sound Pressure Level Measurements. Applied Sciences. 2023; 13(3):1613. https://doi.org/10.3390/app13031613

Chicago/Turabian Style

Navarro, Juan M., and Antonio Pita. 2023. "Machine Learning Prediction of the Long-Term Environmental Acoustic Pattern of a City Location Using Short-Term Sound Pressure Level Measurements" Applied Sciences 13, no. 3: 1613. https://doi.org/10.3390/app13031613

APA Style

Navarro, J. M., & Pita, A. (2023). Machine Learning Prediction of the Long-Term Environmental Acoustic Pattern of a City Location Using Short-Term Sound Pressure Level Measurements. Applied Sciences, 13(3), 1613. https://doi.org/10.3390/app13031613

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop