Next Article in Journal
Application and Development of QKD-Based Quantum Secure Communication
Next Article in Special Issue
LFDNN: A Novel Hybrid Recommendation Model Based on DeepFM and LightGBM
Previous Article in Journal
On the Asymptotic Capacity of Information-Theoretic Privacy-Preserving Epidemiological Data Collection
Previous Article in Special Issue
Manipulating Voice Attributes by Adversarial Learning of Structured Disentangled Representations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Source Acquisition Device Identification from Recorded Audio Based on Spatiotemporal Representation Learning with Multi-Attention Mechanisms

1
Hubei Key Laboratory for High-Efficiency Utilization of Solar Energy and Operation Control of Energy Storage System, Hubei University of Technology, Wuhan 430068, China
2
National Engineering Research Center for Multimedia Software, School of Computer Science, Wuhan University, Wuhan 430072, China
3
Department of Digital Media Technology, Central China Normal University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(4), 626; https://doi.org/10.3390/e25040626
Submission received: 13 February 2023 / Revised: 4 April 2023 / Accepted: 5 April 2023 / Published: 6 April 2023
(This article belongs to the Special Issue Information-Theoretic Approaches in Speech Processing and Recognition)

Abstract

:
Source acquisition device identification from recorded audio aims to identify the source recording device by analyzing the intrinsic characteristics of audio, which is a challenging problem in audio forensics. In this paper, we propose a spatiotemporal representation learning framework with multi-attention mechanisms to tackle this problem. In the deep feature extraction stage of recording devices, a two-branch network based on residual dense temporal convolution networks (RD-TCNs) and convolutional neural networks (CNNs) is constructed. The spatial probability distribution features of audio signals are employed as inputs to the branch of the CNN for spatial representation learning, and the temporal spectral features of audio signals are fed into the branch of the RD-TCN network for temporal representation learning. This achieves simultaneous learning of long-term and short-term features to obtain an accurate representation of device-related information. In the spatiotemporal feature fusion stage, three attention mechanisms—temporal, spatial, and branch attention mechanisms—are designed to capture spatiotemporal weights and achieve effective deep feature fusion. The proposed framework achieves state-of-the-art performance on the benchmark CCNU_Mobile dataset, reaching an accuracy of 97.6% for the identification of 45 recording devices, with a significant reduction in training time compared to other models.

1. Introduction

Source acquisition device identification from recorded audio comprises research on the identification of recording devices by analyzing the intrinsic characteristics of audio [1] and is an important topic in the field of digital audio passive forensics [2,3,4]. In secure forensic applications, recording device identification can be used to authenticate the source of audio files to confirm security, and device-related information in the audio is equivalent to an embedded passive watermark, which is a unique fingerprint belonging to the attributed device [5,6]. In addition, it can also assist in determining whether the audio has been tampered with by detecting whether the device information in the audio is consistent, thereby confirming the integrity and reliability of the data [4,7]. Similarly, it can be used to determine the authenticity of a voice in order to identify whether it is AI-generated or not.
Research on recording device source identification focuses on the identification of the source recording device of the audio data, i.e., the attribution category of the target audio is derived by matching with the database through identification. This identification process relies on the signature information of the recording device embedded in the audio data, which is a kind of recording-device-related discriminatory information that is generated in the process of recording the target sound using the recording device. The target audio is often regarded as a part of the human voice perception system, but ambient information is also recorded during the recording process. Device-related information that is not easily perceptible to the human ear is also embedded during the audio generation process. This device-related information is derived from differences in the transfer functions of different devices during recording, which are due to differences in the selection of recording-related components and the design of electronic circuits within each device, causing convolutional distortion in the input speech during recording, leaving device-signature information in the recording [8]. Existing studies show that this information is mainly distributed in the low- and medium-frequency bands and is not easily detectable by the human ear.
The development of research on audio recording device identification in the existing literature has gone through several stages, and the existing methods can be divided into three main stages. The first class of methods is based on research on feature engineering of audio signals, such as the Mel frequency cepstrum coefficient (MFCC) [8,9] and the band energy difference (BED) [5] method to extract spectral-like features or other distinguishing features using acoustically relevant a priori knowledge, as well as voice activity detection (VAD) [10] and spectral subtraction [11] methods to reduce irrelevant signals in features. The second class of methods is based on machine learning models, such as inputting features into supervised machine learning models such as support vector machine (SVM) [8] and the Gaussian mixture model (GMM) [12,13,14], which distinguish feature data by mapping them to a high-dimensional space; during model training, the information of feature data in each dimension is transformed into probability density information corresponding to each category. The third class of methods is representation learning based on neural network models, applying deep learning algorithms to the recording device recognition process and constructing network models that fit with the feature data, such as using deep neural networks (DNNs) to refine feature data to improve information density [15] and using CNNs for deep feature extraction from the feature matrix [16,17].
With the widespread streaming of media on the Web, research on streaming data has become a hot topic, among which spatiotemporal representation learning has been rapidly developed due to its sensitivity to the spatiotemporal properties of streaming data [18,19]. Spatiotemporal representation learning can automatically learn hierarchical feature representations from spatiotemporal-type data based on the powerful function approximation capability of deep learning; it is particularly effective for streaming data such as video or audio, and some researchers have used it in audio–visual tasks via a two-branch attention mechanism [20] to model the relation between the global feature of one modality and the local features of another. Similarly, we used a dual-attention mechanism for information complementation in audio understanding for features from different abstraction levels. Inspired by spatiotemporal representation learning, we propose a spatiotemporal information fusion network based on RD-TCN and CNN for recording device identification and incorporate multi-attention mechanisms in this paper. The temporal representation part is constructed with a TCN based on residual dense blocks, and the spatial representation part is constructed with a CNN for deep spatial feature extraction. We use a spatial attention mechanism, temporal attention mechanism, and branch attention mechanism according to the characteristics of feature information to automatically learn the weight assignment to multi-dimensional feature parameters and achieve efficient feature fusion, respectively. The main contributions of this paper are summarized as follows:
  • In this paper, we propose a two-branch network to implement spatiotemporal representation learning for recording device identification. The extraction of deep temporal features is performed by a residual dense TCN, and the extraction of deep spatial features is performed by a CNN. The whole representation learning process is optimized by a structured loss function. The implementation codes of this research are available at https://github.com/CCNUZFW/STRLMA (accessed on 20 March 2023).
  • In order to collaborate with the spatiotemporal representation network to obtain a better representation of recording devices, we design three attention mechanisms: a spatial attention mechanism, temporal attention mechanism, and branch attention mechanism. The spatial and temporal attention mechanisms assign weights to the input features Gaussian super vector (GSV) and MFCC, respectively, to enhance the representativeness of both features. The branch attention mechanism is applied to the fusion of two-way branches to promote the learning of key information in the fusion process.
  • Compared with six baseline methods, the proposed framework achieves state-of-the-art performance on the benchmark CCNU_Mobile dataset, reaching an accuracy of 97.6% for the identification of 45 recording devices, with a significant reduction in training time compared to baseline models.
The rest of this paper is organized as follows. Section 2 is a review of related work in the existing literature. In Section 3, we provide a problem definition of this research and summarize the notations that appear in this paper. In Section 4, we introduce the main framework of this paper, describing in detail the spatiotemporal representation learning model for extracting sound source information. Section 5 focuses on the experimental validation phase and develops the analysis of the experimental results. Section 6 presents the conclusion and future research outlook.

2. Related Work

In this section, we review related studies on audio recording device identification and classify them into the following two categories based on the technical characteristics of each method.

2.1. Recording Device Identification Based on Traditional Feature Engineering and Machine Learning

In the early stage, some researchers drew on the research in fields such as speaker recognition and used features from these areas as features for source device identification, which has achieved good results. Hanilci et al. [8] proposed a device source identification feature based on MFCC and used an SVM classifier and a vector quantization (VQ) classifier to classify MFCC features in the back end. Their results show that the SVM classifier is better than the VQ classifier. Although this method achieves good results, it also has some drawbacks. For example, most of the recording segments are obtained by recording the human voice, which contains too much speaker information. Therefore, Aggarwal et al. [11] proposed, for the first time, extracting MFCC features from noise spectrum signals of non-speech signals. This approach is better than extracting MFCC features from speech segments based on the analysis of experimental results. The features applied in these methods are not specifically designed for recording device identification.
In addition to MFCC using the speaker recognition domain, GSV [21] has been introduced into the device source identification domain. GSV features are mean vectors in GMM [22], and in device source identification involving GSV features, it is common to use MFCC features to train GMM models, transform the MFCC feature recognition problem into a GMM probability distribution problem, and extract the GSV features. Kotropoulos et al. [12] obtained a GMM model by training MFCC features, extracted GSV features from it to characterize different categories of device source information, and used a neural network with a radial-basis function kernel in the back end for classification, which achieved a good recognition accuracy. Although the original GSV features can achieve certain results in the field of device source identification, the GSV features are probability space density functions containing speaker and device sources, which are not necessarily applicable for direct characterization of device source information. Therefore, Jiang et al. [13] proposed a kernel-based GSV feature in order to make GSV features more applicable to device source identification tasks. This feature projects the traditional GSV into another device information space. Comparative experiments are conducted at the back end using SVM and a sparse representation-based classifier (SRC). The experiments prove that the kernel-based GSV feature achieves better results. These models do not scale well for newly registered audio samples or recording devices.
Although traditional acoustic features have achieved good recognition results in device source identification, most of these traditional features are constructed based on the human ear auditory system and cannot fully characterize the variability among recording devices. Jin et al. [23] proposed a device source identification method based on device source noise estimation. This method first used the silent segment speech signal to extract the device source noise signal as a device source carrier; then, the spectral features were extracted on this carrier, and an SVM classifier was used on the back end to conduct experiments on a dataset containing 24 different models of cell phones. The proposed features performed best, with the highest recognition accuracy of 94.53%. Although the features based on the above method achieve high identification accuracy, the computational effort increases exponentially due to the complexity of the mapping function in the feature extraction process. Therefore, in order to simplify the computational process and save time cost, Luo et al. [5] proposed a simpler spectral feature extraction method. This method first divides the device source speech signal into frames and then performs Fourier transform directly and characterizes the device source differences by calculating the BED between two adjacent frames. This method is relatively intuitive and also effectively reduces the computational effort while still achieving good results in experiments on a dataset of 141 device sources. These shallow recognition models have limited ability to handle a large number of speech samples.

2.2. Recording Device Identification Based on Deep Learning

In recent years, deep learning has been widely used and achieved improved performances in many fields [24,25], mainly due to the powerful representation learning capability of deep learning [26,27]. Li et al. [15,28] proposed two kinds of deep representation features based on supervised learning: one of the deep learning features based on supervised learning uses MFCC features and labels to train a DNN, then extracts bottleneck features from the middle layer of the DNN. The other deep learning feature uses MFCC features and labels to train the deep autoencoding network, then takes the bottleneck features of the middle layer as the device source features afterwards. The experimental results show that the deep representation learning features outperformed the general features. Combined with previous developments in the field of device source identification, spectral features and 2D image form features can also work well in the field of device source identification.
Influenced by research in image processing and recognition, several researchers have proposed methods to transform audio signals into two-dimensional images to characterize device source features. Qin et al. [16] trained CNN models by transforming speech signals into a spectral map as input. Lin et al. [29] combined deep learning methods with traditional spectral features, using an attention mechanism to assign adaptive network feature weights to different bands of the spectrum in the same device source speech. In response to the methods emerging in the field of traditional device source identification, Qi et al. [30] first used both denoising processing and spectral subtraction to obtain a noisy signal, then extracted the Fourier histogram coefficients of the noisy signal as the input features of a deep network model. Deep learning models can not only train large datasets, but also have powerful generalization and migration properties. Baldini et al. [17] used shallow CNN models in the back end to surpass traditional classification methods and achieved better results. However, the fitting of shallow networks does not fully reflect the fitting effect of deep learning. In a previous study [1], we proposed a multifeature fusion recording device source identification method based on an attention mechanism, using deep representation learning to extract key information features for device source identification. These methods use deep representation learning methods to obtain better results compared to shallow models. Furthermore, these methods focus more on the representation learning of spatial features and less on the representation learning of temporal features. Therefore, in this paper, we propose a recording device identification method based on spatiotemporal representation learning and a multi-attention mechanism, considering the spectral spatial properties of GSV features and the temporal characteristics of MFCC features, and construct a model of the spatiotemporal characteristics of recording-device-related information in audio data.

3. Preliminaries

In this section, we first formally define the task of audio recording device identification. Then, we explain the definitions related to deep temporal features and deep spatial features used in this paper. Next, we describe the proposed multi-attention mechanism. Table 1 shows some important notations. The following section provides a more detailed explanation of their roles.

3.1. Problem Definition

Definition 1 
(Audio recording device identification problem). Audio recording device identification aims at detecting the recording device identity of a test audio signal ( A t ) from an enrollment database ( { A d e | d = 1 , 2 , , D } ) as:
d * = arg max d f A 1 e , A t ; φ , f A 2 e , A t ; φ , , f A D e , A t ; φ ,
where f ( · ) denotes a function used to calculate the similarity; φ stands for the parameters of the back end; A e and A t are the enrollment and test device features, respectively; and D > 1 denotes the number of enrollment devices. If A t can never be outside of the D registered devices, then the recording device identification problem is a closed-set problem; otherwise, it is an open-set problem. A flow chart of the recording device identification process is presented in Figure 1.

3.2. Deep Feature and Multi-Attention Mechanism Definition

Definition 2 
(Deep temporal feature). The deep temporal features are extracted from the MFCC by RD-TCN. For time-series feature data, which are generally frame-processed features, it is necessary to not only analyze the information of feature vectors within frames but also to analyze the information association between frames. Deep temporal features are extracted from the original features using deep learning, which requires the use of networks with sequence modeling capabilities, such as RNN and LSTM; a modified TCN is used in this paper. The extracted deep temporal features reflect the temporal information in the feature data, including the analysis of similarity information and mutation information between adjacent frames, as well as the analysis of long-term fluctuation information.
Definition 3 
(Deep spatial feature). The deep spatial features are extracted from the GSV by CNN. GSV is a probability distribution feature extracted by GMM that maps the original feature data to the high-dimensional feature space during GSV extraction, and the extracted GSV reflects the statistical information of Gaussian distribution in the sample space. The extracted deep spatial features reflect the spatial information in the feature data, including the analysis of the correlation between different Gaussian components in the GSV feature matrix, as well as the analysis of the local associations of different regions in the two-dimensional space of the feature matrix.
Definition 4 
(Multi-Attention mechanism). The multi-attention mechanism proposed in this paper consists of three attention mechanisms, namely a spatial attention mechanism, temporal attention mechanism, and branch attention mechanism, which act in different stages for feature reconstruction and feature fusion, respectively. The core idea is learning the importance distribution in the input features according to the corresponding attention blocks, then enhancing the information related to classification in the features by weight assignment. The reconstruction of the original features by the spatial attention mechanism and the temporal attention mechanism highlights the classification-related information and improves the utilization of effective information in the model. The fusion of the two-way network information by the branch attention mechanism achieves complementary utilization of spatiotemporal feature information and improves the information density.

4. Methods

The spatiotemporal representation learning framework includes feature reconstruction, deep feature extraction, feature fusion, and classification decisions, and the whole task is jointly trained with end-to-end architecture. The overall framework is shown in Figure 2.
  • The feature reconstruction phase is divided into temporal feature reconstruction and spatial feature reconstruction. Temporal feature reconstruction assigns adaptive weights to the features at the temporal scale through the attention mechanism and marks the significant sequences to improve the effects of important feature sequences on the model. Similarly, reconstruction for spatial information involves learning the spatial feature information using the attention mechanism to assign different weights to enhance spatial features.
  • In the deep feature extraction stage, the temporal branch based on RD-TCN is used to extract the deep temporal features, and the spatial branch based on CNN is used to extract the deep spatial features.
  • In the feature fusion phase, a branch attention mechanism is designed for the fusion of deep temporal features and deep spatial features.
  • In the classification decision phase, we apply a multiloss joint computation strategy in order to build an end-to-end network system and optimize the learning process of the two-way branch network and the decision end.
The process of the proposed spatiotemporal representation learning model is shown in Algorithm 1.

4.1. Feature Reconstruction Phase

4.1.1. Feature Reconstruction of MFCC Based on a Temporal Attention Mechanism

The MFCC feature is a frequency cepstrum feature based on short-term Fourier transform, which is one of the most commonly used features in the field of source recording device identification. The extraction process of MFCC includes preprocessing (frame splitting and windowing), fast Fourier transform (FFT), Mel filtering, logarithmic operation, and discrete cosine transform (DCT). The specific extraction process is as follows:
  • First, in order to obtain a stable representation of the audio signal in the frequency domain, the audio signal ( A ) needs to be framed. A Hamming window with frame length ( f l ) and frame shift ( f s ) is used to obtain the short time frames;
  • Then, the frequency spectral information of each frame is obtained by performing a fast Fourier transform on the framed and windowed signal;
  • Then, Mel-scale triangular filters are used to filter the frequency spectra of frames;
  • Then, the logarithmic amplitude spectrum at the output of each filter bank is calculated, and the M-dimension MFCC vectors are obtained by DCT calculation.
Although MFCC is capable of representing time-based changes in audio information, this raw information contains a large amount of redundant information. The temporal attention mechanism is a typical technique used for time-series data classification to attenuate and remove the noisy or irrelevant parts. In the process of device source identification, the temporal attention mechanism possesses strong practical significance. For example, the presence of many human voices in the device source speech interferes with the device source information, and sequence segments with many similar human voices can be attenuated by the temporal attention mechanism, which can be used to overcome the problem of masking and irrelevant signals. Specifically, the attention mechanism can be used to measure the relevance of each time step, and the temporal attention weights provide significance to meaningful values in the sequence, which helps to distinguish similar regions and mark important information. The temporal attention mechanism constructs weights in the same way as the spatial attention mechanism, using convolutional layers to construct learnable parameters, then adjusting the temporal feature maps.
The temporal attention mechanism proposed in this paper first extracts the internal correlation between the contextual relationships and feature vectors in the temporal feature spectrum through a convolutional layer, then assigns weights to the time series using a single-scale maximum pooling layer to control the importance coefficients of each feature vector on the time scale. Finally, weights indicating the degree of importance are assigned to each time node in the temporal feature spectrum by multiplying by the input feature sequence.
In summary, the temporal attention mechanism with X as input and X as output can be formulated as
W t e m = F t e m X , θ t e m = σ ( F C ( M P ( δ ( C o n v 2 ( δ ( C o n v 1 ( X ) ) ) ) ) ) ) , X = W t e m X .
where σ is sigmoid activation; δ is ReLU activation; and F C , M P , and C o n v denote the fully connected layer, maximum pooling layer, and convolutional layer, respectively. In the training phase, attentional temporal pooling and temporal convolutional networks are jointly trained to guide our model for effective information extraction in the temporal dimension. The temporal attention mechanism is designed to emphasize the importance distribution of sequential information, and its network structure and parameters are designed to learn the temporal correlation of input features on the time scale. The specific structure is shown in Figure 3.
Algorithm 1 The proposed spatiotemporal representation learning model.
  Input: MFCC feature X : a sequence of MFCC vectors { x 1 , x 2 , , x T } , GSV feature
       G : a feature matrix of shape ( M , K ) .
  Output: The prediction of the attributed recording device for the input sample
 1 Reconstruct the input temporal feature X into X by temporal attention mechnism:
W t e m = F t e m X , θ t e m = σ ( F C ( M P ( δ ( C o n v 2 ( δ ( C o n v 1 ( X ) ) ) ) ) ) ) , X = W t e m X .
 2 Reconstruct the input spatial feature G into G by spatial attention mechnism:
W s p a = F s p a ( G , θ s p a ) = σ ( F C ( M P ( δ ( C o n v ( G ) ) ) ) ) , G = W s p a G .
 3 Extract deep spatial features G ^ through CNN blocks of spatial network branch:
G ^ = F C N N ( G ) .
 4 Extract deep temporal features X ^ through RD-TCN blocks of the temporal network branch:
X ^ = F R D T C N ( X ) .
 5 Compute spatial feature loss L s p a :
y ^ s p a ( d ) = arg max d = d * S o f t m a x ( W C N N G ^ + b ) , L s p a = d = 1 D y d log ( y ^ s p a ( d ) ) .
 6 Compute temporal feature loss L t e m :
y ^ t e m ( d ) = arg max d = d * S o f t m a x ( W R D T C N X ^ + b ) , L t e m = d = 1 D y d log ( y ^ t e m ( d ) ) .
 7 Concatenate deep temporal features X ^ and deep spatial features G ^ , and assign weights by branch attention mechanism to achieve feature fusion:
Y = [ X ^ , G ^ ] , W b r a = F b r a Y , θ b r a = σ ( F C ( M P ( δ ( C o n v 2 ( δ ( C o n v 1 ( Y ) ) ) ) ) ) ) , Y = W b r a Y .
 8 Compute classification loss L c l a :
y ^ c l a ( d ) = arg max d = d * S o f t m a x ( W c l a Y + b ) , L c l a = d = 1 D y d log ( y ^ c l a ( d ) ) .
 9 Compute the overall loss L t o t a l :
L t o t a l = α L s p a + β L t e m + γ L c l a .
10 Predict the source recording device d * .

4.1.2. Feature Reconstruction of GSV Based on the Spatial Attention Mechanism

GSV features have been shown to be effective in research on recording device source identification [12,13]; the core idea is that the probability distribution of any shape can be approximated by multiple Gaussian distribution functions. GSV is constructed through the adoption of GMM. In adapting the parameters of GMM by the MAP adaptation algorithm using the target device data, the target device GMM parameters are obtained, and the mapping of feature data in the GMM feature space in the parameter learning process is highly correlated with the target category in the mapping direction.
The discrepancy information in the GMM model of each target device source mainly exists in its mean value, and the mean vector thus extracted is the GSV feature. For the number of Gaussian components (K), the dimension of each MFCC vector is M, and the mean value of each target model is a matrix of ( M , K ) . The GSV features are obtained by concatenating the mean vectors of each target model.
The extraction process of GSV features involves the following three steps:
Step 1: If a audio data correspond to a feature ( X ), where X = { x 1 , x 2 , x T } , and assuming its dimensionality is M, the formula used to calculate its likelihood function is:
p x t λ = k = 1 K ω k p k x t ,
where this density function is obtained by weighting K single Gaussian density functions ( p k ( x t ) ), where the mean μ k and the covariance Σ k of each Gaussian component are of sizes ( 1 , M ) and ( M , M ) , respectively.
p k x t = 1 ( 2 π ) M / 2 Σ k 1 / 2 exp 1 2 x t μ k T Σ k 1 x t μ k ,
where the mixture weights ( w k ) satisfy k = 1 K w k = 1 . Assuming that λ denotes the set of model parameters, λ = { w k , μ k , Σ k | k = 1 , 2 , , K } , which is derived by expectation maximization (EM) iterative training.
Step 2: Using the EM algorithm to estimate the iterative parameter ( λ ), first, assign λ an initial value; then, estimate the new parameter ( λ ) in order to satisfy p ( X λ ) p ( X λ ) . In order to ensure that the likelihood of λ under λ is the highest possible, the new parameters are iteratively trained again; the estimation formulae of each parameter are shown in Equations (5)–(7).
w k = 1 T t = 1 T p k x t , λ ,
μ k = t = 1 T p k x t , λ x t t = 1 T p k x t , λ ,
Σ k = t = 1 T p k x t , λ x t μ t ) x t μ t T t = 1 T p k x t , λ .
where w k denotes the mixture weights, μ k denotes the mean, and Σ k is the covariance matrix.
Step 3: Finally, the feature vectors of D devices are adapted by MAP to obtain the device-specific source GMM, and the mean vector of GMM is extracted, which is the GSV features.
In device source identification, the information density of different parts of the spatial features is not equal, and only the parts relevant to the classification task need to be attended to. The spatial attention mechanism finds the most important parts of the feature map for processing and displays the importance in the form of probability maps or probability feature vectors to emphasize important information and suppress useless information. The spatial attention mechanism designed in this paper aims to reconstruct the features by including weights for the device source features so that the features retain sufficient device-related spatial information. The spatial attention mechanism uses convolution and pooling to construct learnable parameters, which are jointly optimized with the whole model to construct efficient device source features.
The spatial attention mechanism proposed in this paper captures the spatial relationships in the input feature maps through a 2D convolutional layer and controls the weight assignment of the importance of spatial information by maximum pooling. Then, spatial-attention-based weight assignment is achieved by multiplying the corresponding elements in the input features. In summary, the spatial attention mechanism, with G as the input and G as the output, can be formulated as:
W s p a = F s p a ( G , θ s p a ) = σ ( F C ( M P ( δ ( C o n v ( G ) ) ) ) ) , G = W s p a G .
where σ is the sigmoid activation; δ is the ReLU activation; and F C , M P , and C o n v denote the fully connected layer, maximum pooling layer, and convolution layer, respectively. The spatial attention mechanism aims to establish a mapping of the importance level of location information, and its network structure and parameters are designed to learn the spatial correlation of the input feature matrix. The specific structure is shown in Figure 4. The structure diagrams of the attention blocks are drawn similarly, but they have different types of inputs and outputs (the temporal dimension in the temporal feature ( X ) is much larger than the dimension of the spatial feature ( G ), T ≫ K > M ). They also differ in terms of their structure and parameters set by the design purpose.

4.2. Deep Feature Extraction Phase

4.2.1. Deep Spatial Feature Extraction Based on CNN

In this paper, a CNN is used to extract spatial feature information from the input data, and after completing the spatial feature extraction, a fully connected layer is used to remove the location information and reduce the sensitivity of parameters in the subsequent fusion process. CNN networks have excellent processing ability for 2D data. The spatial information in the input data is extracted by transforming the input 2D data into a feature map through feature extraction with a convolutional kernel. The CNN local perception and parameter-sharing feature greatly reduces the network parameters, ensures the sparsity of the network, and preserves the local correlation of the samples. In summary, the CNN block, with G as the input feature and G ^ as the output feature, can be formulated as:
G ^ = F C N N ( G ) .
The CNN used for deep spatial feature extraction is composed of a convolutional layer, pooling layer, and fully connected layer. The convolutional layer is obtained by convolving the feature surface with the local region of the layer’s feature surface through a convolutional kernel; this operation is able to extract the deep features from the feature surface. Setting the convolution kernel as ( H , L , C ) , where C is the number of channels and ( H , L ) is the size of a single convolution kernel, the feature is passed through the convolution calculation to the convolution layer to form the data as:
v i , j = f h = 1 H l = 1 L c = 1 C a h , l , c j w h , l , c i + b i ,
where i denotes the i-th channel of the convolution layer; C denotes the C channels associated with the convolution layer in the C channels of the input layer (or pooling layer) ( C C ); w h , l , c i denotes the convolution kernel required for the i-th channel of the convolution layer; a h , l , c j is the j-th input of the input layer (or pooling layer) (depends on the value of the input layer or pooling layer and the step size); f ( · ) is the activation function, usually chosen as a ReLU, sigmoid, or tanh function; and v i , j denotes the specific value of the j-th value of the i-th channel.
The resolution of the eigenfaces is then reduced by pooling operations, while also maintaining the spatial invariance of the eigenfaces.
v j l = d o w n v j l 1 ,
where d o w n ( v j l 1 ) denotes the j-th feature mapping after pooling for the l 1 -th layer.
After convolution, the data from the previous layer are normalized with different weights by discarding the location information through the fully connected layer to obtain an output result with the following formula for the feature information output from the fully connected layer.
g ^ j = f w j l f c v j l + b j ,
where f c ( v j l ) denotes the expansion of the pooling layer into a fully connected form, and g ^ j is the the j-th value of spatial feature information extracted by the CNN.
The spatial feature losses ( L s p a ) are expressed as:
y ^ s p a ( d ) = arg max d = d * S o f t m a x ( W C N N G ^ + b ) , L s p a = d = 1 D y d log ( y ^ s p a ( d ) ) .

4.2.2. Deep Temporal Feature Extraction Based on RD-TCN

Time-series features are commonly modeled using recurrent neural networks (RNNs) [31,32] and their variants, as they have a recurrent regression structure suitable for modeling time series; however, their performance is still affected by their inability to be parallel, and they have disadvantages such as limited ability to analyze information within a sequence and high training time consumption. In general, CNNs are limited by the convolutional kernel size and cannot capture long-term dependent information well and are therefore considered unsuitable for modeling of time-series information. However, some recent works have shown that specific convolutional neural network structures can also achieve good results; for example, TCN was compared with a variety of RNN structures and found to be capable of matching or even surpassing RNN models in a variety of tasks.
TCN has the advantage of parallel processing of data. Unlike RNN-type networks, which require sequential processing of data in a sequential manner, TCN can perform parallel processing of given temporal data, thereby greatly reducing the training time. In addition, TCN can set parameters such as the number of network layers, convolutional kernel size, and the dilation rate according to the task, thereby indirectly setting the receptive field size to accommodate feature data of different complexities.
TCNs use causal convolution in order to enable the extraction of temporal information. Causal convolution has a unidirectional structure, which means that there is a preceding cause before there is a subsequent effect, and it is a strictly time-constrained model. Pure causal convolution still suffers from the problem that the modeling length of the time scale is limited by the size of the convolutional kernel, and a very large number of hidden layers needs to be stacked in order to extract longer time series of dependencies. Thus, dilation convolution is used to complete the temporal convolution operation. Dilation convolution allows the input to be sampled at intervals during convolution. Therefore, the dilation convolution makes the size of the effective receptive field grow exponentially with the number of layers, allowing the convolution kernel to parse the extracted features. This allows the convolutional network to obtain a sufficiently large receptive field with relatively few layers.
Here, we introduce improvements on the basis of a TCN by constructing an RD-TCN using residual dense blocks, which further enhances the information utilization compared with a normal TCN. A structural diagram of RD-TCN block is shown in Figure 5. In summary, the RD-TCN block, with X as the input feature and X ^ as the output feature, can be formulated as:
X ^ = F R D T C N ( X ) .
We propose the use of a residual dense structure [33] instead of a residual structure as in an ordinary TCN in order to make efficient use of all the layered information in the convolutional layers through the residual dense blocks. We construct residual dense blocks in TCN to achieve skip-layer connectivity; the residual dense blocks allow the network to transfer information in a cross-layer manner, fusing lower-layer features with higher-layer features to enhance information utilization. The network structure of the residual dense block is shown in Figure 6. In this paper, residual dense blocks are used as building blocks for the RD-TCN, as they contain ordinary residual concatenation layers and dense feature fusion with local residual learning. The residual dense block supports continuous memory, and after extracting multiple layers of local dense features, it further fuses the global features, then adaptively retains the layered features in a global manner, thereby achieving implicit deep supervision.
As shown in Figure 5, the activation values in the l-th layer are represented by V l R C × T . Each layer has the same number of filters (C), which enables us to combine activation values from different layers using skip connections later. When the convolution kernel size is set to three, i.e., the number of time steps beyond which the convolution is applied, the temporal convolution is calculated as
V t l = δ ( W ( 1 ) V t 2 s l 1 + W ( 2 ) V t s l 1 + W ( 3 ) V t l 1 + b ) ,
where V t l is the result of the dilated convolution at time t of the l-th layer, the input to the first layer is the deep temporal feature X , and s denotes the rate parameter of the dilation convolution. The result obtained after adding the residual dense connections is calculated as
X ^ = F C ( δ ( W [ V 3 , V 2 , V 1 , X ] ) + X ) ,
where W denotes the a set of weights, where the bias term is omitted for simplicity, and [ V 3 , V 2 , V 1 , X ] refers to the concatenation of the feature maps in the axis of the channel.
The temporal feature loss ( L t e m ) are expressed as
y ^ t e m ( d ) = arg max d = d * S o f t m a x ( W R D T C N X ^ + b ) , L t e m = d = 1 D y d log ( y ^ t e m ( d ) ) .

4.3. Feature Fusion Based on the Branch Attention Mechanism

In this model, attention factors are added to the temporal features to adjust the weights of temporal scales to enhance the temporal feature vectors, and attention factors are also added to the spatial features to adjust the information distribution of spatial features. After that, the temporal and spatial information is extracted, and finally, the information is fused using the branch attention mechanism. This allows for an increase in the efficiency of temporal and spatial information extraction and improves the identification accuracy.
This module addresses the problem of fusing different types of features by assigning weights to different types of features through network learning, thereby improving the effects of important features on model training. The attention fusion mechanism designed in this paper includes a weight learning layer and a dot product assignment layer, where the weights to be learned are calculated by the operations of convolution and pooling. In summary, the branch attention mechanism, with X ^ and G ^ as the input and Y as the output, can be formulated as:
Y = [ X ^ , G ^ ] , W b r a = F b r a Y , θ b r a = σ ( F C ( M P ( δ ( C o n v 2 ( δ ( C o n v 1 ( Y ) ) ) ) ) ) ) , Y = W b r a Y .
where [ X ^ , G ^ ] refers to the concatenation of the features X ^ and G ^ in the axis of the channel. The branch attention mechanism fuses the feature outputs of the two-way branches, and its structure and parameters are designed to aid in the feature assignment during the fusion process. The specific structure is shown in Figure 7.
Branch attention can be considered a dynamic branch selection mechanism to determine which branches need to be focused on when used in networks with multibranch structures. The branch attention mechanism first constructs the learnable parameters by convolution, while a set of feature maps can be obtained after convolution, followed by compression of the features into a set of real numbers using a pooling layer, which is also equivalent to a convolution operation with a global sense field. The last layer is the Softmax layer, which adds nonlinearity to the weights. After the Softmax layer, a set of weights is generated for the features, which is used to represent the correlation and importance between the branches of the feature network. The weights are then assigned to each branch, and the features are fused in a dot-product manner.
The classification losses ( L c l a ) are expressed as
y ^ c l a ( d ) = arg max d = d * S o f t m a x ( W c l a Y + b ) , L c l a = d = 1 D y d log ( y ^ c l a ( d ) ) .

4.4. Classification Decision Based on the Joint Loss Function

The overall loss function of the framework proposed in this paper consists of temporal feature loss, spatial feature loss, and classification loss, as shown in the following equation.
L t o t a l = α L s p a + β L t e m + γ L c l a ,
where L t o t a l is the overall loss of the model; L t e m and L s p a are the temporal feature loss and spatial feature loss, respectively; L c l a is the classification loss of the decision layer; and α , β , γ denote the proportion coefficients of the three losses to adjust the proportion of each partial loss in the overall loss. The temporal feature loss and spatial feature loss are used to ensure the accuracy of the learning direction of important information in the feature information extraction stage and to update the weights of the branching network over time, respectively.
The classification loss in the decision layer is used to ensure the consistency of the content between the predicted and true categories and to backpropagate the learning update of the model weights. The learning algorithm of the objective function for the loss and network weights is shown in Algorithm 2.  
Algorithm 2: Algorithm for model objective function learning
Entropy 25 00626 i001

5. Experimental Results and Analysis

In this section, we show the extensive experiments conducted to validate the effectiveness and advantages of our method, and we compare the performance of our method with several baselines on a dataset containing 28,890 audio data samples. To validate the effectiveness of each independent part of the framework, we conducted an ablation study based on the idea of the controlled variable method to observe the core modules and key hyperparameters of our method. To obtain the optimal parameters and structure, we conducted experiments using our method for the parameter and structure settings.

5.1. Dataset

In this paper, we use the CCNU_Mobile dataset [1] to validate the proposed method. The CCNU_Mobile dataset consists of audio data recorded by 45 different device models and recording devices from 9 different brands, including Apple, Huawei, Honor, Nubia, Oppo, Vivo, Xiaomi, Samsung, and ZTE; the device models are shown in Table 2.
The recording corpus of this dataset is from the TIMIT dataset. During the recording process of the CCNU_Mobile dataset, all the training data in the TIMIT dataset were first spliced into a long audio file with a duration of about 110 min, then played in a quiet dedicated recording studio environment, using 45 devices to record. After the recording was completed, the recorded long audio was cut into segments, with 642 audio samples recorded per device; each recording sample is about 10 s in duration, and the audio samples are all single-channel audio files saved in .wav format with a sampling rate of 32,000 Hz and a bit rate of 512 kbps. We randomly selected 514 samples from each device category for the training set and another 128 samples for the test. The training and test sets accounted for 80% and 20% of the total samples, respectively, and the validation set was 20% of the training set.

5.2. Evaluation Metrics

For evaluation purposes, we used classification accuracy ( A c c ) as the performance metric, which is defined as:
A c c = S c r S t × 100 % ,
where S t denotes the total number of samples participating in the test, and S c r denotes the number of samples that were correctly identified. The recognition results of the samples during the test are computed by the final layer of the model, using the Softmax layer, which maps the output of each neuron in the penultimate fully connected layer to the ( 0 , 1 ) interval to obtain the scores of each category in the multiclassification task and calculates the probability of belonging to each category to obtain the recognition results. For a total of n categories ( S o f t m a x d ) represented by numerical values, where D denotes the number of categories and d ( 0 , D ] , the Softmax calculation formula is:
P S o f t m a x i = e v i d D e v d ,
where S o f t m a x i denotes the i-th output, and v i denotes the value of the i-th category. The final Softmax value obtained for each category sums to 1.

5.3. Baselines

In order to evaluate the performance of the methods proposed in this paper, we needed other baselines for comparison. The baseline methods used in the comparison experiments are described below. The details of the baselines are as follows: the Gaussian mixture model–universal background model (GMM-UBM) and MFCC-SVM are two classical methods in the field of recording device source identification, and we used them as reference standards. I-vector and BED features are improved in terms of features to solve the feature information representation problem, and we also used them as features for the baseline methods. In addition, GSV-CNN was added to the baseline method as a representative method of deep learning methods, and a multifeature fusion method using multiple features and deep learning models was added to the baseline method as a novelty.
GMM-UBM [14]: This method uses training GMM and calculates probability scores for each category for classification; using UBM to train GMM reduces the computational effort.
MFCC-SVM [8]: This method uses MFCC, an inverse spectral feature widely used in the audio recognition field, as a feature and SVM as a classification model.
I-vector-SVM [34]: The i-vector method reduces the dimensionality by obtaining the audio feature vector of the high-dimensional target device source, projecting it in the subspace, using factor analysis to eliminate the factors that add redundancy to obtain the low-dimensional feature vector.
BED-SVM [5]: This method uses a spectral feature extraction method that calculates the baseband energy difference, which is more intuitive to characterize the device source differences and effectively reduces the computational effort.
X-vector TDNN-based systems [35,36]: This method is a high-performance method in the field of speaker recognition. We conducted two experiments on X-vector TDNN-based systems: one using SVM as the back-end classification after extracting X-vectors based on TDNN and the other using a Softmax layer as the back-end direct classification of TDNN.
GSV-CNN [17]: This method constructs a representative CNN model for identification, and in this experiment, we used the GSV feature as the input.
Multifeature fusion [1]: This method uses three feature inputs and deep and shallow feature fusion using a CNN/DNN.

5.4. Experimental Settings

(1) Framework settings: To extract the temporal MFCC features, a Hamming window function is chosen; the length of each signal frame is 30 ms, the overlap is 15 ms, and the length of the extracted single MFCC vector is 39. The spatial GSV features are extracted using 64 fitted Gaussian components, and the feature matrix shape is (39, 64). Table 3 and Table 4 show the detailed parameters specific to the three attention mechanisms and the network framework, respectively.
(2) Training settings: The method was trained in the same environment as all baseline methods during the training process, with a total of 200 training epochs, a batch size of 128, an initial l r = 0.001, and the l r representing 1/10 of the previous 30 epochs. The experiment was conducted with an Adam optimizer with β 1 = 0.9 and β 2 = 0.999. The experiment-related feature extraction steps and model implementation steps were performed in Matlab 2020a and TensorFlow 2.1. The relevant experimental hardware configuration is as follows: CPU: Intel® Xeon Gold 5218 × 2; GPU: NVIDIA® TITAN RTX (24 GB video memory); and memory: 32 GB.

5.5. Results and Discussion

5.5.1. Comparison with Baseline Methods

For the accuracy comparison between this method and the baseline methods on the CCNU_Mobile dataset, Table 5 shows the mean values of the 10 results obtained from the experiments for each method. From the experimental results in Figure 8, it can be seen that the deep-learning-based methods generally outperform the traditional methods, and the choice of features can have an impact on recognition under the same model selection; however, the magnitude of the impact is not significant. As shown in Figure 8, our proposed method obtains the highest ACC while demonstrating smaller variance in multiple repetitions of the experiment relative to other deep learning methods, which implies that our proposed method achieves more stable recognition performance. In the classification of 45 categories of recording devices, the proposed method achieves a recognition rate of 97.68%, corresponding to approximately 5626 correctly identified samples among 5760 test audio samples. As shown in Figure 9 and Figure 10, although the performance of our method is not optimal in terms of inference time, it remains in the same order of magnitude as the methods with the shortest inference time, while the training time is significantly reduced.

5.5.2. Ablation Experiments of Attention Mechanisms

In order to verify the effectiveness of the three attention mechanisms, four experiments were conducted, in which three attention mechanisms were present and one of them was removed. It can be seen from the experimental results in Table 6 that when one of the three attention mechanisms is absent, the model without a spatial attention mechanism obtains the worst effect, with an accuracy rate reduced by 0.2% compared with the model without a modal attention mechanism and the model without a temporal attention mechanism. This result shows that the spatial attention mechanism has the greatest influence on the network model investigated in this paper and that the spatial attention mechanism is also the model that requires the most network parameters. When the three attention mechanisms exist simultaneously, the model achieves the best effect, reaching 97.6%, directly proving the effectiveness of the three attention mechanisms proposed in this paper and indirectly proving that the three attention mechanisms achieved their respective intended effects.

5.5.3. Validation Experiments of the RD-TCN Temporal Feature Extraction Network

In order to verify the effectiveness of the RD-TCN network, comparison experiments were conducted.The parameters of the ordinary TCN-based temporal feature extraction and the RD-TCN-based temporal feature extraction are shown in Table 4. The spatial feature extraction network adopts a structure of a CNN with the same network parameters as in Table 4.
In the RD-TCN temporal feature extraction network, the original residual blocks of the TCN network are replaced with residual dense blocks. From the experimental results in Table 7, it can be seen that the recognition accuracy of the ordinary TCN-based temporal feature extraction network reaches 97.4%, while the TCN network with the addition of residual dense blocks reaches 97.6%, which is an improvement of 0.2% compared with the ordinary TCN temporal feature extraction network, proving that the RD-TCN temporal feature extraction network is effective.

5.5.4. Experiments for the Joint Loss Function

To optimize the model training process, we used the strategy of joint loss function, which was optimized by combining three losses. To explore the effectiveness of the joint loss function, we compared it with a model using a single cross-entropy loss function and tested the optimal parameters by controlling the proportional coefficients.
Table 8 shows the experimental results for the loss function settings. The ACC for all experiments is higher than 97.1% when using the joint loss function optimization strategy. The joint loss function optimization strategy significantly outperforms the network model using a single loss. The experimental results show that the network can converge better by co-optimizing multiple losses. Experiments comparing the joint loss function with different proportional coefficients shows that the network model can achieve the best results when the proportional coefficients are (0.25, 0.5, 0.25) and (0.25, 0.25, 0.5), reaching 97.6%, which indicates that the proper allocation of proportional loss coefficients can achieve better results.

6. Conclusions

In this paper, we propose a spatiotemporal representation learning method with a multi-attention mechanism for recording device source identification. In this paper, we used multiple features as input, which is a kind of multiple knowledge representation; this strategy effectively improves the recognition accuracy and contributes to the generalization ability and interpretability of the model [37]. The two features used were extracted from different abstraction levels, among which GSV is extracted based on MFCC, which is an information supplement based on a priori domain knowledge and eliminates category-independent information in the extraction process to avoid the interference of speech content or speaker information, and enriches the feature information in terms of probability density distribution information. In terms of structure, the method includes temporal and spatial feature reconstruction, temporal and spatial feature extraction, temporal and spatial feature fusion, and joint loss calculation stages. In terms of the contributions of this paper, first, feature reconstruction of temporal and spatial features was performed using temporal attention and spatial attention, respectively, significantly highlighting the temporal and spatial information of device source features and subsequently optimizing the fusion of both features using the branch attention mechanism. Secondly, in order to solve the problem of long training time required for existing deep learning models for device identification, in this paper, we used an RD-TCN network for temporal feature extraction, which also improved the recognition speed of the model on the basis of improving the accuracy rate.
In terms of experimental performance, the present method shows a small improvement in recognition accuracy relative to methods proposed in our previous work; however, the training time of this method is significantly reduced, and its performance is more stable. The recognition performance on this dataset is close to saturation, and even small improvements are meaningful, so we will subsequently explore application-oriented and effective recognition methods. In future work, we will further optimize the model to obtain more significant recognition performance improvement. However, the present method still has some shortcomings in the feature extraction stage, which require the early extraction of two features before model training, increasing the complexity of the application. We will investigate how to extract more expressive features and research feature extraction methods that are better-matched to the task of recording device identification. In terms of identification models, we will explore the application of self-supervised learning in this area and attempt to improve the application of transformer-based representational learning in recording device identification. We will also improve our method to be applicable to more complex recognition scenarios to promote additional applications.

Author Contributions

Conceptualization, Z.W. and C.Z.; methodology, Z.W., S.F. and D.Z.; software, S.F. and D.Z.; validation, Z.W., S.F. and D.Z.; formal analysis, Z.W. and C.Z.; investigation, Z.W. and C.Z.; resources, Z.W. and C.Z.; data curation, Z.W. and C.Z.; writing—original draft preparation, Z.W., S.F. and D.Z.; writing—review and editing, Z.W.; visualization, S.F.; supervision, Z.W.; project administration, Z.W.; funding acquisition, Z.W. and C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research work presented in this paper was supported by the National Natural Science Foundation of China (Nos. 62177022 and 61901165), the AI and Faculty Empowerment Pilot Project (No. CCNUAI&FE2022-03-01), the Collaborative Innovation Center for Informatization and Balanced Development of K-12 Education by the MOE and Hubei Province (No. xtzd2021-005), Natural Science Foundation of Hubei Province (No. 2022CFA007), and the National Natural Science Foundation of China (No. 61501199).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data will be made available upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MFCCMel frequency cepstrum coefficient
BEDBand energy difference
VADVoice activity detection
SVMSupport vector machine
GMMGaussian mixture model
DNNDeep neural network
GSVGaussian super vector
VQVector quantization
SRCSparse representation-based classifier
2DTwo-dimensional
FFTFast Fourier transform
DCT Discrete cosine transform
EM Expectation maximization
RNN Recurrent neural network

References

  1. Zeng, C.; Zhu, D.; Wang, Z.; Wang, Z.; Zhao, N.; He, L. An end-to-end deep source recording device identification system for Web media forensics. Int. J. Web Inf. Syst. 2020, 16, 413–425. [Google Scholar] [CrossRef]
  2. Maher, R.C. Audio forensic examination. IEEE Signal Process. Mag. 2009, 26, 84–94. [Google Scholar] [CrossRef]
  3. Wang, Z.; Yang, Y.; Zeng, C.; Kong, S.; Feng, S.; Zhao, N. Shallow and Deep Feature Fusion for Digital Audio Tampering Detection. EURASIP J. Adv. Signal Process. 2022, 2022, 69. [Google Scholar] [CrossRef]
  4. Zeng, C.; Yang, Y.; Wang, Z.; Kong, S.; Feng, S. Audio Tampering Forensics Based on Representation Learning of ENF Phase Sequence. Int. J. Digit. Crime Forensics 2022, 14, 94. [Google Scholar] [CrossRef]
  5. Luo, D.; Korus, P.; Huang, J. Band Energy Difference for Source Attribution in Audio Forensics. IEEE Trans. Inf. Forensics Secur. 2018, 13, 2179–2189. [Google Scholar] [CrossRef]
  6. Cuccovillo, L.; Aichroth, P. Open-set microphone classification via blind channel analysis. In Proceedings of the IEEE 2016 International Conference on Communications and Signal Processing (ICCSP), Shanghai, China, 20–25 March 2016; pp. 2074–2078. [Google Scholar] [CrossRef]
  7. Zhao, H.; Malik, H. Audio Recording Location Identification Using Acoustic Environment Signature. IEEE Trans. Inf. Forensics Secur. 2013, 8, 1746–1759. [Google Scholar] [CrossRef]
  8. Hanilci, C.; Ertas, F.; Ertas, T.; Eskidere, O. Recognition of Brand and Models of Cell-Phones from Recorded Speech Signals. IEEE Trans. Inf. Forensics Secur. 2012, 7, 625–634. [Google Scholar] [CrossRef]
  9. Hadoltikar, V.A.; Ratnaparkhe, V.R.; Kumar, R. Optimization of MFCC parameters for mobile phone recognition from audio recordings. In Proceedings of the IEEE 2019 3rd International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 12–14 June 2019; pp. 777–780. [Google Scholar] [CrossRef]
  10. Hanilci, C.; Kinnunen, T. Source cell-phone recognition from recorded speech using non-speech segments. Digit. Signal Process. 2014, 35, 75–85. [Google Scholar] [CrossRef] [Green Version]
  11. Aggarwal, R.; Singh, S.; Roul, A.K.; Khanna, N. Cellphone identification using noise estimates from recorded audio. In Proceedings of the IEEE 2014 International Conference on Communications and Signal Processing (ICCSP), Melmaruvathur, India, 3–5 April 2014; pp. 1218–1222. [Google Scholar] [CrossRef]
  12. Kotropoulos, C.; Samaras, S. Mobile phone identification using recorded speech signals. In Proceedings of the IEEE 2014 International Conference on Digital Signal Processing (DSP), Hong Kong, China, 20–23 August 2014; pp. 586–591. [Google Scholar] [CrossRef]
  13. Jiang, Y.; Leung, F.H.F. Source Microphone Recognition Aided by a Kernel-Based Projection Method. IEEE Trans. Inf. Forensics Secur. 2019, 14, 2875–2886. [Google Scholar] [CrossRef]
  14. Garcia-Romero, D.; Espy-Wilson, C.Y. Automatic acquisition device identification from speech recordings. In Proceedings of the IEEE 2010 International Conference on Acoustics, Speech and Signal Processing, Dallas, TX, USA, 14–19 March 2010; pp. 1806–1809. [Google Scholar] [CrossRef]
  15. Li, Y.; Zhang, X.; Li, X.; Zhang, Y.; Yang, J.; He, Q. Mobile Phone Clustering from Speech Recordings Using Deep Representation and Spectral Clustering. IEEE Trans. Inf. Forensics Secur. 2018, 13, 965–977. [Google Scholar] [CrossRef]
  16. Qin, T.; Wang, R.; Yan, D.; Lin, L. Source Cell-Phone Identification in the Presence of Additive Noise from CQT Domain. Information 2018, 9, 205. [Google Scholar] [CrossRef] [Green Version]
  17. Baldini, G.; Amerini, I.; Gentile, C. Microphone Identification Using Convolutional Neural Networks. IEEE Sens. Lett. 2019, 3, 6001504. [Google Scholar] [CrossRef]
  18. Wang, S.; Cao, J.; Yu, P. Deep Learning for Spatio-Temporal Data Mining: A Survey. IEEE Trans. Knowl. Data Eng. 2020, 34, 3681–3700. [Google Scholar] [CrossRef]
  19. Lyu, L.; Wang, Z.; Yun, H.; Yang, Z.; Li, Y. Deep Knowledge Tracing Based on Spatial and Temporal Representation Learning for Learning Performance Prediction. Appl. Sci. 2022, 12, 7188. [Google Scholar] [CrossRef]
  20. Wu, Y.; Zhu, L.; Yan, Y.; Yang, Y. Dual Attention Matching for Audio-Visual Event Localization. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6291–6299. [Google Scholar] [CrossRef]
  21. Campbell, W.M.; Sturim, D.E.; Reynolds, D.A. Support vector machines using GMM supervectors for speaker verification. IEEE Signal Process. Lett. 2006, 13, 308–311. [Google Scholar] [CrossRef]
  22. Reynolds, D.A. A mixture modeling approach to text-independent speaker ID. J. Acoust. Soc. Am. 1990, 87, S109. [Google Scholar] [CrossRef]
  23. Jin, C.; Wang, R.; Yan, D.; Tao, B.; Chen, Y.; Pei, A. Source Cell-Phone Identification Using Spectral Features of Device Self-noise. In Proceedings of the Digital Forensics and Watermarking: 15th International Workshop (IWDW), Beijing, China, 17–19 September 2016; pp. 29–45. [Google Scholar] [CrossRef]
  24. Wang, Z.; Zuo, C.; Zeng, C. SAE Based Unified Double JPEG Compression Detection System for Web Image Forensics. Int. J. Web Inf. Syst. 2021, 17, 84–98. [Google Scholar] [CrossRef]
  25. Zeng, C.; Ye, J.; Wang, Z.; Zhao, N.; Wu, M. Cascade Neural Network-Based Joint Sampling and Reconstruction for Image Compressed Sensing. Signal Image Video Process. 2022, 16, 47–54. [Google Scholar] [CrossRef]
  26. Wang, Z.; Wang, Z.; Zeng, C.; Yu, Y.; Wan, X. High-Quality Image Compressed Sensing and Reconstruction with Multi-Scale Dilated Convolutional Neural Network. Circuits Syst. Signal Process. 2023, 42, 1593–1616. [Google Scholar] [CrossRef]
  27. Zeng, C.; Yan, K.; Wang, Z.; Yu, Y.; Xia, S.; Zhao, N. Abs-CAM: A Gradient Optimization Interpretable Approach for Explanation of Convolutional Neural Networks. Signal Image Video Process. 2022, 1–8. [Google Scholar] [CrossRef]
  28. Li, Y.; Zhang, X.; Li, X.; Feng, X.; Yang, J.; Chen, A.; He, Q. Mobile phone clustering from acquired speech recordings using deep Gaussian supervector and spectral clustering. In Proceedings of the IEEE 2017 International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 2137–2141. [Google Scholar] [CrossRef]
  29. Lin, X.; Zhu, J.; Chen, D. Subband Aware CNN for Cell-Phone Recognition. IEEE Signal Process. Lett. 2020, 27, 605–609. [Google Scholar] [CrossRef]
  30. Qi, S.; Huang, Z.; Li, Y.; Shi, S. Audio recording device identification based on deep learning. In Proceedings of the IEEE 2016 International Conference on Signal and Image Processing (ICSIP), Beijing, China, 13–15 August 2016; pp. 426–431. [Google Scholar] [CrossRef] [Green Version]
  31. Graves, A.; Mohamed, A.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE 2013 International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar] [CrossRef] [Green Version]
  32. Zaremba, W.; Sutskever, I.; Vinyals, O. Recurrent neural network regularization. arXiv 2014, arXiv:1409.2329. [Google Scholar] [CrossRef]
  33. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar] [CrossRef] [Green Version]
  34. Rao, W.; Mak, M. Boosting the Performance of I-Vector Based Speaker Verification via Utterance Partitioning. IEEE Trans. Audio Speech Lang. Process. 2013, 21, 1012–1022. [Google Scholar] [CrossRef] [Green Version]
  35. Snyder, D.; Garcia-Romero, D.; Sell, G.; Povey, D.; Khudanpur, S. X-Vectors: Robust DNN Embeddings for Speaker Recognition. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5329–5333. [Google Scholar] [CrossRef]
  36. Peddinti, V.; Povey, D.; Khudanpur, S. A time delay neural network architecture for efficient modeling of long temporal contexts. In Proceedings of the 16th Annual Conference of the International Speech Communication Association (Interspeech 2015), Dresden, Germany, 6–10 September 2015; pp. 3214–3218. [Google Scholar] [CrossRef]
  37. Yang, Y.; Zhuang, Y.; Pan, Y. Multiple knowledge representation for big data artificial intelligence: Framework, applications, and case studies. Front. Inf. Technol. Electron. Eng. 2021, 22, 1551–1558. [Google Scholar] [CrossRef]
Figure 1. Flow chart of recording device identification.
Figure 1. Flow chart of recording device identification.
Entropy 25 00626 g001
Figure 2. Spatiotemporal representation learning framework ( W t e m , W s p a : weights of temporal and spatial features; L t o t a l : overall loss of the model; L s p a : spatial feature loss; L t e m : temporal feature loss; L c l a : classification loss; α , β , γ : proportion coefficients of the three losses).
Figure 2. Spatiotemporal representation learning framework ( W t e m , W s p a : weights of temporal and spatial features; L t o t a l : overall loss of the model; L s p a : spatial feature loss; L t e m : temporal feature loss; L c l a : classification loss; α , β , γ : proportion coefficients of the three losses).
Entropy 25 00626 g002
Figure 3. Temporal attention mechanism structure diagram. (T: number of MFCC vectors; M: length of per MFCC vector; ⊙: element-wise product (Hadamard product of features-by-column and weight column vector ( W t e m ))).
Figure 3. Temporal attention mechanism structure diagram. (T: number of MFCC vectors; M: length of per MFCC vector; ⊙: element-wise product (Hadamard product of features-by-column and weight column vector ( W t e m ))).
Entropy 25 00626 g003
Figure 4. Spatial attention mechanism structure diagram. (M: length of per MFCC vector; K: number of Gaussian mixture components; ⊙: element-wise product (Hadamard product of spatial feature and weight matrix ( W s p a ))).
Figure 4. Spatial attention mechanism structure diagram. (M: length of per MFCC vector; K: number of Gaussian mixture components; ⊙: element-wise product (Hadamard product of spatial feature and weight matrix ( W s p a ))).
Entropy 25 00626 g004
Figure 5. Schematic diagram of a TCN based on residual dense blocks ( V l : activation values in the l-th layer; s: dilation rate; +: concatenation operation; ⊕: addition operation).
Figure 5. Schematic diagram of a TCN based on residual dense blocks ( V l : activation values in the l-th layer; s: dilation rate; +: concatenation operation; ⊕: addition operation).
Entropy 25 00626 g005
Figure 6. Structure of the residual dense block of the temporal convolution. ⊕: addition operation.
Figure 6. Structure of the residual dense block of the temporal convolution. ⊕: addition operation.
Entropy 25 00626 g006
Figure 7. Branch attention mechanism structure diagram. (⊙: element-wise product (Hadamard product of features by column and weight column vector W b r a )).
Figure 7. Branch attention mechanism structure diagram. (⊙: element-wise product (Hadamard product of features by column and weight column vector W b r a )).
Entropy 25 00626 g007
Figure 8. Comparison of the ACC with baseline methods.
Figure 8. Comparison of the ACC with baseline methods.
Entropy 25 00626 g008
Figure 9. Comparison of the training time with baseline methods.
Figure 9. Comparison of the training time with baseline methods.
Entropy 25 00626 g009
Figure 10. Comparison of the inference time with baseline methods.
Figure 10. Comparison of the inference time with baseline methods.
Entropy 25 00626 g010
Table 1. A summary of notations.
Table 1. A summary of notations.
NotationDescription
A Audio signal
f l Frame length
f s Frame shift
X = { x 1 , x 2 , , x T } , x t R M MFCC features extracted after framing of samples
KNumber of Gaussian components
G R M × K GSV features of samples
C o n v Convolution layer
M P Maximum pooling layer
F C Fully connected layer
σ Sigmoid activation function
δ ReLU activation function
w , W Weights
λ = { w , μ , Σ } Model parameters of GMM
L t o t a l , L s p a , L t e m , L c l a Components of the loss function
α , β , γ Proportion coefficients of the three losses
Table 2. Brands and models of recording device in the CCNU_Mobile dataset.
Table 2. Brands and models of recording device in the CCNU_Mobile dataset.
BrandModels
APPLEiPhone6(4), iPhone6s(3), iPhone7p, iPhoneX, iPhoneSE, iPad7, Air1, Air2(2)
XIAOMImi2s, note3, mi5, mi8, mi8se(2), mix2, redmiNote4x, redmi3S
HUAWEINova, Nova2s, Nova3e, P10, P20, TAG-AL00
HONORhonor7x, honor8(3), honorV8, honor9, honor10
VIVOx3f, x7, y11t
ZTEC880a, G719c
SAMSUNGS8, sphd710
OPPOR9s
NUBIAZ11
Table 3. Network parameters of the three attention mechanisms.
Table 3. Network parameters of the three attention mechanisms.
Spatial Attention MechanismTemporal Attention MechanismBranch Attention Mechanism
Conv(16, (3*3), (2*2))Conv (16, (5*5), (3*3))Conv (16, (3*3), (1*2))
Conv(32, (3*3), (2*2))Conv (32, (3*3), (3*3))Conv (32, (5*5), (1*3))
Pooling ((3*3), (2*2))Pooling ((5*1), (5*1))Pooling ((1*5), (1*5))
FlattenPooling ((7*1), (7*1))Pooling ((1*7), (1*7))
Dense (2496)FlattenFlatten
Reshape (39*64)Dense (650)Dense (2)
MultiplyReshape (650*1)Reshape (2*1)
/MultiplyMultiply
Table 4. Network parameters of the deep feature extraction networks.
Table 4. Network parameters of the deep feature extraction networks.
Spatial Feature ExtractionTemporal Feature Extraction Network
CNNTCN BlockRD-TCN Block
Conv (6, (5*5), (1*1))Conv (6, 5, 1)Conv (6, 5, 1)
Pooling ((2*2), (2*2))Conv (6, 5, 1)Conv (6, 5, 1)
Conv (16, (5*5), (1*1))addConv (6, 5, 1)
Pooling ((2*2), (2*2))Conv (16, 5, 1)add
Conv (40, (5*5), (1*1))Conv (16, 5, 1)Conv (16, 5, 1)
Pooling ((2*2), (2*2))addConv (16, 5, 1)
FlattenConv (40, 3, 1)Conv (16, 5, 1)
FC (1024)Conv (40, 3, 1)add
/addConv (40, 3, 1)
//Conv (40, 3, 1)
//Conv (40, 3, 1)
//add
Table 5. Comparison of identification accuracy with baseline methods.
Table 5. Comparison of identification accuracy with baseline methods.
MethodACCTraining TimeInference Time
GMM − UBM31.57 ± 11.65%2.3 h677 ms
MFCC + SVM70.83 ± 4.88%1.3 h72 ms
I-vector + SVM90.33 ± 3.93%1.3 h182 ms
X-vector + SVM89.34 ± 4.26%1.8 h210 ms
BED + SVM91.61 ± 2.41%1.5 h26 ms
GSV + CNN92.14 ± 1.40%1.9 h34 ms
X-vector + TDNN95.91 ± 2.76%1.8 h51 ms
Multifeature fusion97.42 ± 2.05%2.6 h17 ms
Proposed method97.68 ± 0.47%1.2 h39 ms
Table 6. Comparison of the attention mechanism selection.
Table 6. Comparison of the attention mechanism selection.
ModelACC
Without temporal attention mechanism97.2%
Without spatial attention mechanism97.0%
Without branch attention mechanism97.2%
Model with three attention mechanisms97.6%
Table 7. Comparison experiment between RD-TCN and ordinary TCN.
Table 7. Comparison experiment between RD-TCN and ordinary TCN.
ModelACC
Model with ordinary TCN block97.4%
Model with RD-TCN block97.6%
Table 8. Experiments for the loss function setting.
Table 8. Experiments for the loss function setting.
Loss Function SettingProportional Coefficients ( α , β , γ )ACC
Single loss function0, 0, 196.7%
Joint loss function0.5, 0.25, 0.2597.1%
Joint loss function0.25, 0.5, 0.2597.6%
Joint loss function0.25, 0.25, 0.597.6%
Joint loss function0.4, 0.2, 0.497.3%
Joint loss function0.2, 0.6, 0.297.4%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, C.; Feng, S.; Zhu, D.; Wang, Z. Source Acquisition Device Identification from Recorded Audio Based on Spatiotemporal Representation Learning with Multi-Attention Mechanisms. Entropy 2023, 25, 626. https://doi.org/10.3390/e25040626

AMA Style

Zeng C, Feng S, Zhu D, Wang Z. Source Acquisition Device Identification from Recorded Audio Based on Spatiotemporal Representation Learning with Multi-Attention Mechanisms. Entropy. 2023; 25(4):626. https://doi.org/10.3390/e25040626

Chicago/Turabian Style

Zeng, Chunyan, Shixiong Feng, Dongliang Zhu, and Zhifeng Wang. 2023. "Source Acquisition Device Identification from Recorded Audio Based on Spatiotemporal Representation Learning with Multi-Attention Mechanisms" Entropy 25, no. 4: 626. https://doi.org/10.3390/e25040626

APA Style

Zeng, C., Feng, S., Zhu, D., & Wang, Z. (2023). Source Acquisition Device Identification from Recorded Audio Based on Spatiotemporal Representation Learning with Multi-Attention Mechanisms. Entropy, 25(4), 626. https://doi.org/10.3390/e25040626

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop