Next Article in Journal
Spatial-Temporal Speckle Variance in the En-Face View as a Contrast for Optical Coherence Tomography Angiography (OCTA)
Next Article in Special Issue
Pose2Sim: An End-to-End Workflow for 3D Markerless Sports Kinematics—Part 2: Accuracy
Previous Article in Journal
Possible Applications of Edge Computing in the Manufacturing Industry—Systematic Literature Review
Previous Article in Special Issue
Quantitative Analysis of Performance Recovery in Semi-Professional Football Players after the COVID-19 Forced Rest Period
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recognizing Solo Jazz Dance Moves Using a Single Leg-Attached Inertial Wearable Device

Faculty of Electrical Engineering, University of Ljubljana, 1000 Ljubljana, Slovenia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(7), 2446; https://doi.org/10.3390/s22072446
Submission received: 18 January 2022 / Revised: 15 March 2022 / Accepted: 19 March 2022 / Published: 22 March 2022
(This article belongs to the Special Issue Wearables and Computer Vision for Sports Motion Analysis)

Abstract

:
We present here a method for recognising dance moves in sequences using 3D accelerometer and gyroscope signals, acquired by a single wearable device, attached to the dancer’s leg. The recognition entails dance tempo estimation, temporal scaling, a wearable device orientation-invariant coordinate system transformation, and, finally, sliding correlation-based template matching. The recognition is independent of the orientation of the wearable device and the tempo of dancing, which promotes the usability of the method in a wide range of everyday application scenarios. For experimental validation, we considered the versatile repertoire of solo jazz dance moves. We created a database of 15 authentic solo jazz template moves using the performances of a professional dancer dancing at 120 bpm. We analysed 36 new dance sequences, performed by the professional and five recreational dancers, following six dance tempos, ranging from 120 bpm to 220 bpm with 20 bpm increment steps. The recognition F1 scores, obtained cumulatively for all moves for different tempos, ranged from 0.87 to 0.98. The results indicate that the presented method can be used to recognise repeated dance moves and to assess the dancer’s consistency in performance. In addition, the results confirm the potential of using the presented method to recognise imitated dance moves, supporting the learning process.

1. Introduction

Dancing is an engaging physical activity that requires high levels of body control, skill, and physical fitness. In addition, dancing is an activity cultivating the capacity for creative expression. Following the rhythmic structure, the dancer, especially the professional, expresses his or her interpretation of music by selecting, performing, and assembling various moves in dance sequences.
The main motivation of the research presented is to provide a mechanism for dance move recognition in performed sequences. In general, it has already been reported that dancers can greatly benefit from various assistive technologies [1,2,3,4,5]. A dance move recognition technology would benefit dancers engaged in the learning process. Namely, as one learns to dance, he or she tends to imitate performances demonstrated by instructors. These imitations are visually assessed, usually using a mirror. In addition, instructors provide their students with feedback on how their performances compare to the ideal template. Providing such feedback is usually more challenging in group sessions, in which more students engage in the learning process at the same time. A technological solution would benefit both the student and the instructor and would support performance assessment and progress monitoring.
More experienced dancers would also benefit. Recognising dance moves would enable further investigation and comparison of the sequences of moves, supporting overall creativity evaluation, individual style classification, and observations of differences when dancing alone, in pairs and/or in the presence of an audience. Additionally combining the recognised dance moves with information about the accompanying music would illuminate how the dancer interprets and responds to music. Relying on a dance move recognition technology during the creative process of producing a choreography, a professional dancer would also be able to create a transcript of the assembled moves for later reference.
A segment of dance motion analysis approaches made so far relies on signals obtained from optical motion capture systems, incorporating multiple cameras watching the stage and usually requiring reflective markers to be positioned on the subject’s body [6,7,8,9]. In [6], the authors present a framework for capturing the posture of the dancer’s body. Using the body posture parameters, signals are temporally segmented, motion patterns are extracted, and motion animation in generated. In [7], the authors present dance pose sequences similarity estimation.
Another segment of approaches relies on signals obtained from a Kinect device [1,2,5,10,11,12]. The Kinect device incorporates a depth camera together with a standard video camera into a single device, distinguishes human body parts and joints, and estimates their position and orientation. In [1], the authors present a novel framework for the real-time capture, assessment, and visualisation of ballet dance movements. The dance sequence captured is segmented and cross-referenced against a library of gestural components performed by a teacher. Recognition performance was evaluated on a database of six isolated basic ballet positions recordings made by a teacher and a student. Reported average recognition rates are between 90.5% and 99.5%. Bharatanatyam [2] and a salsa dancing coach application [5] have also been developed. In [10], another Kinect-based system is presented enabling Korean pop (K-pop) motion classification by extracting statistical features, reducing dimensionality with Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA) and incorporating extreme learning machine. Skeletal joint angles for 200 movement types obtained for four dancers were considered and the reported maximum classification rate is 96.5%. In [11], the authors analysed data on Greek traditional dance postures with the goal of extracting robust summaries and providing end users or dance experts a concise and meaningful abstract of dance movements.
Although valuable results have been reported, using equipment located in the surrounding area, either one or more video cameras or a Kinect device, limits the practical aspect of the methodology, making it difficult or altogether impossible to dance in the ballroom. Moreover, processing video signals is computationally intensive, and although the underlying technology and algorithms are steadily improving, computer vision is still limited by lighting and clutter constraints and subtraction of dynamic background.
The presented method capitalises on the benefits of small and lightweight microelectromechanical (MEMS) inertial sensors. Over the past few years, it has been consistently demonstrated that these sensors are an efficient tool in the broader research area of human motion analysis [13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29]. Their characteristic light weight, small size, low power consumption, portability, ease of use, and low cost pave the way for their omnipresence and indispensability in motion analysis.
Most of the research conducted in this field concerns daily activity recognition [13,14,15,16,17,18,19], gait analysis [20,21,22], gesture recognition [23], and sports activity performance assessment for activities such as golf, tennis, baseball, swimming, and skiing [24,25,26,27,28,29]. One of the first wearable systems for capturing dance gestures were developed by the MIT Media Lab [30,31], and various assistive solutions have been presented since [3,4,32,33,34,35,36,37,38,39]. In this context, wearable devices are usually used to detect dancer’s steps and provide them with feedback [3,4,32] or enable the dancer to interact or generate sounds [33,34,35]. In [36], a three-axis inertial measurement unit (IMU) is positioned on the subject’s right forearm and used to analyse one specific hip hop arm move. In particular, the aim is to divide the motion into phases, predict which phase is executed, and give feedback to the subject. In addition, accelerometers are used to estimate torso tilt [37] and assess performance [38] of basic routines in classical ballet.
We focus our analysis on the solo jazz dance style, a rhythmical and playful solo dance in which the dancer depicts jazz music through one’s one movement. Solo jazz evolved through the first half of the 20th century to include elements of both African and European dance and features the vocabulary and steps of the vernacular jazz tradition. To enable the dancer to follow the jazz song rhythmical structure, as a rule, a single solo jazz move is performed following eight music beats. The music beat is considered as the smallest time interval between two successive notes. Dance moves comprise various motion elements, performed in a specified order. Besides steps, motion elements can be, for example, kicks, taps, and jumps. The variety of motion elements and the order of their execution brings a pleiad of predefined, authentic solo jazz dance moves. Due to this variety, solo jazz fits perfectly into our dance move recognition research polygon.
In general, in dancing, the speed of move execution represents the dance tempo and is directly related to the musical tempo of the song. It is measured as the number of steps, or other motion elements, a dancer performs in a minute. As a rule, solo jazz is usually danced to jazz music with tempo anywhere between 120 and 250 beats per minute (bpm). Tempos between 140 and 160 bpm are considered medium. Tempos above 220 bpm are considered fast and generally prove to be too high for the recreational dancers to execute with accuracy and ease.
Aiming to provide ease of use, with the smallest amount of sensing equipment, our dance move recognition methodology relies on a single wearable unit, comprising a MEMS 3D accelerometer and gyroscope, positioned on the dancer’s leg, and enables dance move recognition, regardless of the device orientation and dance tempo. The methodology presented considers that different moves have different signal shapes in the time domain and that these shapes can be used as dance move fingerprints. Furthermore, since position and attitude estimates obtained using inertial sensors are known to be characterised by different inaccuracies, especially drift, the methodology resides on the shapes of the original acceleration and angular velocity signals instead.
The methodology presented is a continuation of our previous work and features solo jazz dance tempo estimation, as presented in [39]. In addition, the specific technique presented in [20], used for defining a wearable device orientation-invariant coordinate system using the 3D accelerometer signals acquired during walking, is adapted and applied in the research presented in this article.
The remainder of this article is organised as follows. In Section 2, we present the materials and methods used for creating the dance move template database, recognition, and experimental validation. In Section 3, we present and discuss the results. Finally, in Section 4, we summarise our findings and draw conclusions, implying further research directions. In all the subsequent sections, we use the following notation rules: large bold letters denote matrices, small bold letters denote vectors, and large or small italics denote scalars.

2. Materials and Methods

2.1. Data Acquisition

2.1.1. Materials

We captured dance motion using a single mbientlab MetaMotionR (MMR) wearable device [40], including a MEMS 3D accelerometer and gyroscope with respective measurement ranges ± 16 g and ± 2000°/s, placed directly above the dancer’s right leg ankle. We hypothesised that analysing the motion of a single leg is sufficient to distinguish between various dance moves. An alternative solution would be to place the device on the dancer’s lower back, since leg motion of high-quality dance performances is usually initiated from the pelvic region, torso, or even the shoulders. However, preliminary investigations have shown that this significantly reduces the recognition performance of the designed solution. The micro-position and orientation of the device are arbitrary.
We set the sampling frequency to 200 Hz, which proved to be sufficient for the problem at hand using empirical evidence. The 3D accelerometer and gyroscope together provided 6 inertial signals, each representing one variable in a 6D space.
In addition, a software metronome with an established beat frequency error of 1 bpm was used to simulate a steady quarter note music tempo and dictate the tempo of dancing.

2.1.2. Measurements

Six female dancers participated in the study—five recreational dancers (age: 33 ± 5 years, height: 163 ± 5 cm) with over three years of experience in solo jazz dancing and one professional (age: 35 years, height: 164 cm).
Altogether, I = 15 authentic solo jazz moves were considered: (1) Tackie Annie, (2) Fall of the log, (3) Kicks, (4) Half break, (5) Struttin’, (6) Savoy kick, (7) 20s Charleston, (8) Knee slaps, (9) Fishtails, (10) Apple Jacks, (11) Boogie back, (12) Boogie forward, (13) Crazy leg, (14) Cross step, and (15) Shorty George. For an informative overview of how these moves are performed, the reader is referred to various sources available online, e.g., [41]. Each move was performed on an 8-beat basis.
Measurements were performed in two sessions. Following the obtained instructions, participants attached the wearable device above their right leg ankle, voluntarily setting its micro-position and orientation.
The first measurement session was conducted only for the professional dancer. The dancer performed several consecutive repetitions for each of the 15 considered moves, as consistently as possible, following a fixed reference tempo υref = 120 bpm, dictated by the metronome. For each move, the dancer started dancing from a still position. We refer to the obtained set of six inertial signals as the learning sequence.
The second measurement session was conducted for all six participants. Each was given the task of performing 5 repetitions of each of the 15 considered moves in a prescribed order. The recreational dancers were instructed to mimic the professional’s execution of moves as closely as possible. This task was repeated for 6 different dance tempos, ranging from 120 bpm to 220 bpm with 20 bpm increments. For each dancer, we obtained 6 sets of inertial signals. We refer to these 36 sets as test sequences.
All measurements were supplemented with video recordings. The study was conducted in Ljubljana, Slovenia. It followed the Code of Ethics of the University of Ljubljana, which provides guidelines for studies involving human beings and is in accordance with the Declaration of Helsinki. All participants gave written informed consent.

2.2. Signal Processing Overview

The implemented signal processing workflow is presented in Figure 1. For each acquired sequence (1), signal pre-processing (2) is applied. The pre-processed learning sequence is used to extract template moves and create the template database (3), following a semi-automatic correlation-based segmentation procedure. The pre-processed test sequences are analysed and searched for template moves (4), according to the following steps:
4.1.
Dance tempo estimation and signal temporal scaling, achieved using a bank of enhanced comb filters as presented in [39];
4.2.
Initial template matching, performed on a sliding correlation basis, using the magnitudes of the temporally scaled acceleration and angular velocity;
4.3.
Signal transformation to the templates’ coordinate system; and
4.4.
Final template matching, performed again on a sliding correlation basis, but by using the acceleration and angular velocity 3D projections of on the template coordinate system axes instead of their magnitudes.
For both the initial and final template matching, correlation is used as the only feature for recognition. In the following, all the components of the proposed method are presented in detail. All signal processing was executed offline in the MATLAB 2021 environment [42].

2.3. Signal Pre-Processing

Following the calibration procedure presented in [43], we compensated all signals of the learning and test sequences for sensor inaccuracies. Since the device itself does not provide outputs at exactly equidistant time samples, we interpolated and decimated the acquired signals as necessary, considering the associated measurement timestamp values, to provide for uniform sampling at exactly 200 Hz. To remove motion artefacts and noise, we applied a low-pass filter with a cut-off frequency fco = 50 Hz and finally performed downsampling to fs = 100 Hz, obtaining 3D acceleration and angular velocities at equidistant time samples T = 1/fs = 0.01 s.

2.4. Templates’ Database

2.4.1. Template Extraction

Using all Ns acquired samples of the pre-processed learning sequence, we formed two Ns × 3 signal matrices, A s and Ω s , representing 3D acceleration and angular velocity, respectively. Columns of these two matrices are equal to the respective signal projections on the device-intrinsic coordinate system axes x, y, and z, while rows represent the time samples. The specific orientation of axes x, y, and z in a reference coordinate system, defined by the set position of the wearable device, is arbitrary. We further combined A s and Ω s in a common Ns × 6 signal matrix S:
S = [ A s Ω s ] .
Columns of S are equal to the 3D acceleration and angular velocity x, y, and z projections, while rows represent the time dimension. Comprising the inertial signals of the entire learning sequence, matrix S comprises all performed repetitions of all considered dance moves, used for creating the templates’ database.
We extracted the comprised dance move repetitions by partitioning S into submatrices in the vertical dimension. We achieved this by estimating the repetitions’ onsets and offsets, for each 1 ≤ iI = 15 dance move separately, following a semi-automatic, correlation-based segmentation procedure as follows. First, since for each dance move, the consecutive repetitions are performed from a still position, we defined the onset of the first repetition by visual inspection of the six inertial signals and the accompanying video. We denote the matrix S row index corresponding to this onset as n0.
For defining the offset of the first repetition, we consider, given the known reference dance tempo υref, the expected length, expressed as the number of samples NT, of a single dance move. Since each considered solo jazz move is performed on an 8-beat basis, at fs = 100 Hz sampling and precise 120 bpm dancing (Tbeat = 0.5 s), the expected length is NT = 8 × 0.5 s × 100 Hz = 400 samples. Since dancing is rarely this precise, we allow for a 2% length deviation and set the expected length to be in between 0.98 × NT = 392 and 1.02 × NT = 408 samples. Each length from this interval, i.e., Nn = N0 + n, where N0 = 392 and 0 ≤ n ≤ 0.04 × NT = 16, gives a candidate for the first repetition offset, n0 + Nn − 1.
Since the repetitions are consecutive, for each offset candidate for the first repetition, n0 + Nn 1, there is a single onset candidate for the second repetition, n0 + Nn. Finally considering both repetitions to be equal in length gives a single offset candidate for the second repetition, n0 + 2Nn − 1. For each n, we obtained one candidate pair of the first two repetitions of a move, represented with adjacent, same-size Nn × 6 submatrices of S.
We extracted and standardised to zero mean and unit standard deviation these two submatrices and performed column-wise vectorisation, obtaining, for each n, two 6(N0 + n) long column vectors. Pursuing the highest similarity, for each candidate pair n, we calculated the correlation coefficient between these two vectors. This calculation, supported with implementation equations, is described in detail in Appendix A.2.
We set the first and the second repetition of the considered move to be equal to the repetitions of the candidate pair with the highest correlation coefficient. We finally extracted all possible additional repetitions of the particular move on an equal correlation-based search basis using the onset of the last determined repetition as the new n0.
Due to slight variations in execution duration, we unified in length all extracted repetitions by temporal scaling to the exact expected length NT = 400 samples. For each move i, we composed a set of consistent repetitions, eliminating repetition outliers, determined by visual inspection. We averaged this set column-wise, obtaining finally, for each template move i, an associated single NT × 6 matrix T i . Denoting the first three columns, representing the 3D acceleration, with Ai and the second three, representing 3D angular velocity, with Ω i , for each 1 ≤ iI = 15, we can write:
T i = [ A i Ω i ] .
The 15 matrices T i of the acceleration and angular velocity projections on the device intrinsic coordinate system axes x, y, and z represent our database of template moves. The specific orientation of the axes x, y, and z in the reference coordinate system, defined by the arbitrarily set position of the wearable device during the learning sequence acquisition, defines the coordinate system of the templates.
In addition to Ti, for each move, i we also calculated an NT × 2 matrix T i ¯ ¯ , having the first and second column equal to the magnitudes of the 3D acceleration and angular velocity, respectively. Each k-th row of T i ¯ ¯ is obtained according to:
T i ¯ ¯ k , * = [ A i k , * Ω i k , * ] = [ A i k , 1 2 + A i k , 2 2 + A i k , 3 2 Ω i k , 1 2 + Ω i k , 2 2 + Ω i k , 3 2 ] .
In (3), * denotes all columns.

2.4.2. Templates Similarity Measures

To estimate the similarity between template moves, we used two correlation-based measures as follows. Firstly, we compared the template moves along the acceleration and angular velocity magnitudes. To achieve this, we standardised the columns of T i ¯ ¯ to zero mean and unit standard deviation for each 1 ≤ iI = 15 move and performed column-wise vectorisation of the resulting matrix. For each move, we obtained a 2NT-long vector of magnitudes. For each pair of template moves i and j, 1 ≤ i,jI = 15, we calculated the correlation between the associated vectors as a function of cyclical shift 0 ≤ τ < NT of the acceleration and angular velocity vector parts. We define the maximum value of this correlation for each i,j pair of template moves, denoted with r T i ¯ ¯ T j ¯ ¯ , as the magnitudes’ similarity. This calculation is supported in detail with implementation equations in Appendix A.3.
We also compared the template moves along the acceleration and angular velocity 3D projections. We standardised the columns of Ti to zero mean and unit standard deviation for each 1 ≤ iI = 15 move and performed column-wise vectorisation of the resulting matrix. For each move, we obtained a 6NT -long vector. For each pair of template moves i and j, 1 ≤ i,jI = 15, we again calculated the correlation between the associated vectors as a function of cyclical shift 0 ≤ τ < NT of the projections vector parts. We define the maximum value of this correlation for each i,j pair of template moves, denoted with r T i T j , as the projections’ similarity. This calculation is supported in detail with implementation equations in Appendix A.4.
In the same way as the correlation coefficient calculated along one dimension, r T i ¯ ¯ T j ¯ ¯ and r T i T j can take any value from the range [−1,1], where 1 indicates identical moves, 0 orthogonal moves, and –1 opposite moves. High r T i ¯ ¯ T j ¯ ¯ values suggest common leg activation patterns, regardless of the actual direction of motion in 3D. On the other hand, high r T i T j suggests that two moves also match with respect to the direction of the executing motion. Considering this, the similarity measure r T i T j is expected to have a higher discriminative potential. Note that by performing standardisation column-wise, equal weight is given to each dimension of comparison, x, y, and z, regardless of the intensity of motion along that dimension.

2.5. Dance Move Recognition

2.5.1. Dance Tempo Estimation and Temporal Scaling

For each pre-processed test sequence, we first estimate the dance tempo υest, using a method based on multiple resonators, implemented with enhanced comb feedback filters, as presented in [6].
We temporally scale all acquired inertial signals of the sequence by a factor of υref/υest to obtain a test sequence with dance moves that match the template moves in tempo. We denote with NX the final length of the test sequence.
Using all NX samples, we form two NX × 3 signal matrices of 3D acceleration and angular velocity, denoted with A X and Ω X , respectively. The columns of these two matrices correspond to the respective inertial signal projections on the intrinsic coordinate system axes of the wearable device. Since the orientation of the wearable device is different for each measurement session, these axes and the axes of the coordinate system of the template moves are generally not aligned.
We further form a common NX × 6 test sequence signal matrix X:
X = [ A X Ω X ] .
Columns of X correspond to the acceleration and angular velocity 3D projections while rows represent the time dimension.
We also form an NX × 2 matrix of acceleration and angular velocity magnitudes X ¯ ¯ . Each k-th row of X ¯ ¯ is calculated according to:
X ¯ ¯ k , * = [ A X k , * Ω X k , * ] = [ A X k , 1 2 + A X k , 2 2 + A X k , 3 2 Ω X k , 1 2 + Ω X k , 2 2 + Ω X k , 3 2 ] .
In (5), * denotes all columns.

2.5.2. Initial Template Matching

We search the time-scaled test sequence for template moves by applying sliding correlation-based template matching, initially considering the acceleration and angular velocity magnitudes. For the template i and the test sequence, these are comprised in matrices T i ¯ ¯ (3) and X ¯ ¯ (5), respectively.
For each template move 1 ≤ iI = 15, we first standardise the columns of T i ¯ ¯ to zero mean and unit standard deviation. By vectorising the standardised matrices column-wise, for each template move i, we obtain a 2NT-long vector t i ¯ ¯ of the acceleration and angular velocity magnitudes.
We further apply a 2D sliding window of size NT × 2 to the test sequence signal matrix X ¯ ¯ . For each 1 ≤ nNXNT + 1, we extract an NT × 2 submatrix of X ¯ ¯ , i.e., X ¯ ¯ n : n + N T 1 , * . After standardising and vectorising X ¯ ¯ n : n + N T 1 , * we obtain a 2NT-long vector of the associated acceleration and angular velocity magnitudes. We then calculate the correlation coefficient between this vector and t i ¯ ¯ . Sliding the 2D window vertically, we repeat this process for all n and obtain NXNT + 1 correlation coefficients. We store the results for each template move i in a vector denoted with r T i ¯ ¯ X ¯ ¯ . This calculation is supported in detail with implementation equations in Appendix A.5.
Indicating high similarity between T i ¯ ¯ and X ¯ ¯ , the local maximums of r T i ¯ ¯ X ¯ ¯ reveal the possible onsets of template move i executions in the test sequence. For each dance move i, we identify the local maximums of r T i ¯ ¯ X ¯ ¯ that exceed a certain threshold value, denoted with rmin. We define the indices of these local maximums as move i onsets in the sequence. We further define the associated values of r T i ¯ ¯ X ¯ ¯ local maximums as the recognition confidence.
After searching through the entire sequence for all template moves, we performed a final correction in the sense of false positives detection and elimination: if any two onsets of moves were closer than 0.7 NT, we discarded the one with the lower recognition confidence.

2.5.3. Signal Transformation

To transform the signals of the test sequence from the device intrinsic to the coordinate system of the templates, we first identify from all onsets found in the previous step, for all i and for all n, the one with the highest recognition confidence. This onset defines the strongest matching pair: for i and n, for which the recognition confidence is the highest, the magnitude matrices of the template and subsequence, i.e., T i ¯ ¯ and X ¯ ¯ n : n + N T 1 , * , respectively, match the most. We denote the corresponding matrices of the acceleration 3D projections with ATM and AXM and the combined matrices of the accelerations and angular velocity 3D projections with TM and XM.
Since this template–subsequence pair represents the same move, but in different coordinate systems, it is reasonable to assume that a rotation that aligns them can be defined. Moreover, if we determine such a rotation, we can apply it to the entire test sequence, obtaining the acceleration and angular velocity 3D projections in the coordinate system of the templates.
To define this rotation, we adapted and applied an orientation-independent transformation, as presented in [20]. The original procedure estimates a wearable device orientation-independent coordinate system by calculating three orthogonal unit vectors from the 3D accelerometer signals, acquired using a smartphone in the front pocket of the user’s trousers during walking. The first unit vector, denoted with ζ , is defined by the gravity direction, and is calculated as the mean acceleration within a motion (walking) cycle. The second unit vector, denoted with ξ , is considered to be aligned with the direction of motion and is determined using PCA as the direction with the highest variance once the projection on ζ has been subtracted. The third unit vector, denoted with ψ , is defined as orthogonal to the first two and as such covers lateral motion during the motion cycle.
In our dance analysis context, both the template and subsequence of the best-matching pair represent one full dance move, i.e., motion cycle. Considering this, following the abovementioned, from ATM, we obtain three orthogonal unit vectors ζ Τ , ξ T , and ψ T while from AXM, we obtain three orthogonal unit vectors ζ X , ξ X , and ψ X . Extracting unit vectors ζ , ξ , and ψ from the acceleration 3D projections has already been presented in detail in [20]; we provide details specific to this study in Appendix B.
Combining both triplets of unit vectors, ζ Τ , ξ T , ψ T and ζ X , ξ X , ψ X , we can now define a rotation matrix R that aligns the coordinate system of the test sequence with that of the templates as:
R = [ ζ X ξ X ψ X ] [ ζ T ξ T ψ T ] 1 .
Using R (6), we can transform the original acceleration and angular velocity 3D projections, comprised in matrices A X and Ω X , respectively, from the test sequence’s device intrinsic to the coordinate system of the template moves. Denoting the resulting matrices with A X ( R ) and Ω X ( R ) , we write:
A X ( R ) = A X R Ω X ( R ) = Ω X R .
The common NX × 6 matrix of the rotated acceleration and angular velocity 3D projections of the test sequence, denoted with X ( R ) , given in the coordinate system of the templates, is then:
X ( R ) = [ A X ( R ) Ω X ( R ) ] .

2.5.4. Final Template Matching

We can now search the test sequence for template moves, following a similar procedure as presented in Section 2.5.2., only this time considering the acceleration and angular velocity 3D projections instead of their magnitudes. For the template moves and the test sequence, these are comprised in matrices T i (2) and X ( R ) (6), respectively.
For each template move 1 ≤ iI = 15, we first standardise to zero mean and unit standard deviation the columns of T i . Vectorising the standardised matrices column-wise, for each template move i, we obtain a 6NT-long vector of acceleration and angular velocity projections ti.
We further use a 2D sliding window of size NT × 6, and for each 1 ≤ nNXNT + 1, we extract an NT × 6 submatrix of X ( R ) , i.e., X ( R ) n : n + N T + 1 , * . After standardisation and vectorisation, we obtain a 6NT-long vector of the associated acceleration and angular velocity projections. We then calculate the correlation coefficient between this vector and ti.
Sliding the 2D window vertically, we repeated this process for all n and obtain NXNT + 1 correlation coefficients. We store the results for each template move i in a vector denoted with r T i X ( R ) . This calculation is supported in detail with implementation equations in Appendix A.6.
Indicating high similarity between T i and X ( R ) , local maximums of r T i X ( R ) reveal the possible presence of move i in the test sequence. For each dance move i, we identify the local maximums of r T i X ( R ) that exceed a certain threshold value, rmin. We define the indices of these local maximums as move i onsets in the sequence and the associated r T i X ( R ) values as the recognition confidence.
After the entire sequence has been searched through, we performed a final correction for false positives detection and elimination: if any two onsets were closer than 0.7 NT, we discarded the one with the lower recognition confidence.

2.6. Recognition Performance Assessment

All performances of dancers for all test sequences were visually inspected by the professional using the accompanying videos. All moves determined to be improperly performed were discarded from analysis. All proper performances were appropriately labelled as one of the 15 moves and considered for assessing the recognition ability of the presented method.
The presented recognition method is non-binary and we assess its recognition ability by considering each move that is correctly recognised as a true positive (TP). Each move that is incorrectly recognised we consider as a false positive (FP). Using cumulative TP and FP rates for all I = 15 moves, we calculate the sensitivity and precision scores for each dance tempo considered. We calculate these scores for threshold values rmin ranging from 0.25 to 0.80 with 0.05 increment steps. We finally use the F1 score, calculated from sensitivity and precision, as the unified measure of recognition ability.
In the specific dance motion recognition context, the sensitivity scores are mainly influenced by the similarity between the analysed moves and the corresponding templates in the database. As such, the sensitivity scores obtained for the professional dancer indicate how consistent her performance is, while those obtained for the recreational dancers show how well they imitate the professional.
Precision scores are additionally influenced by the similarities between different moves, indicated by nondiagonal elements of r T i ¯ ¯ T j ¯ ¯ for the initial template matching and of r T i T j for the final template matching. Performed on acceleration and angular velocity 3D projections, final template matching is expected to have a higher discriminative potential than the initial template matching, performed using the acceleration and angular velocity magnitudes.

3. Results and Discussion

3.1. Database of Template Moves

Figure 2 shows the acquired acceleration and angular velocity signals for two illustrative template moves: (3) Kicks and (7) 20s Charleston. The first row depicts the acceleration signals, while the second row shows the angular velocities. Even from pure visual inspection, we can see that both the 3D projections and magnitudes have distinctive signal shapes. It is therefore reasonable to assume that these signal shapes can be used as fingerprints enabling dance move recognition.
The similarities between different template moves r T i ¯ ¯ T j ¯ ¯ , calculated using the magnitudes of acceleration and angular velocity, are presented in Table 1. As expected, some dance moves show greater pairwise similarity while others are more idiosyncratic and show greater distinction from the rest. The highest similarity is r T 7 ¯ ¯ T 8 ¯ ¯ = 0.61 , obtained for the pair of moves (7) 20s Charleston and (8) Knee slaps, reflecting a leg activation pattern these two moves have in common. Both of these moves are executed mainly with the right leg during the first four music beats and with the left leg during the second four music beats.
The move with the lowest similarity with respect to all other moves, i.e., with the lowest maximum r T i ¯ ¯ T j ¯ ¯ value, is Apple Jacks (10): r T 10 ¯ ¯ T j ¯ ¯ 0.37 for 1 ≤ j ≤ 15, j ≠ 10.
The presented values indicate the discriminative potential of using the acceleration and angular velocity magnitudes for recognising different solo jazz dance moves. In particular, for threshold values rmin > 0.61, high precision scores are expected; however, the higher rmin is, the lower is the expected sensitivity, diminishing the model’s overall recognition ability.
The similarities between different template moves r T i T j , calculated using the 3D projections, are presented in Table 2. We can see that these values are generally lower than the corresponding values of r T i ¯ ¯ T j ¯ ¯ , and for all pairs of moves, r T i T j is below 0.5.
The highest r T i T j value is obtained for the pair of moves (4) Half break and (11) Boogie back: r T 4 T 11 = 0.43 . The lowest maximum r T i T j value is 0.24 and is obtained for moves (8) Knee slaps and (14) Cross step. The presented values indicate the expected performance of the final template matching: for threshold values rmin > 0.43, high precision scores are expected.
Lower values of r T i T j obtained with respect to r T i ¯ ¯ T j ¯ ¯ confirm the natural assumption that comparing dance moves using the 3D projections of signals instead of their magnitudes, considering the direction of motion instead of only the general leg activation pattern, has a higher discriminative potential. Therefore, template matching achieved using the 3D projections instead of magnitudes is expected to provide for higher precision scores, better overall recognition ability, and lower values of rmin.

3.2. Dance Move Recognition

For all 36 test sequences, the estimated dance tempo is accurate up to 1 bpm, allowing us to temporally scale all sequences and perform correlation-based template matching, as presented in the previous section.

3.2.1. Validation Using the Professional Dancer’s Test Sequences

For the professional dancer, for each dance tempo 120–200 bpm, 75 moves were acquired. For tempo 220 bpm, 70 moves were acquired. The sensitivity and precision scores are presented in Figure 3 and Figure 4.
For the initial template matching, the highest sensitivity scores range from 0.80 to 0.91 and are obtained for rmin ≤ 0.4. For the final template matching, they range from 0.84 to 0.96 and are also obtained for rmin ≤ 0.4.
For all dance tempos considered, the precision scores are equal to 1 for rmin ≥ 0.65 for the initial template matching and rmin ≥ 0.50 for the final template matching, indicating that for rmin this high, there are no FPs. These results are consistent with the template moves’ pair-wise similarities presented in Table 1 and Table 2. The improvement in precision for the final template matching is a direct consequence of using the 3D projections of the inertial signals instead of their magnitudes for template matching. As expected, considering the direction of motion provides for a better discrimination of moves.
Figure 5 shows the F1 scores. For each dance tempo considered, the maximum F1 score indicates the best recognition ability. Lower F1 scores to the left of the maximum indicate higher FP rates and consequently lower precision. Lower F1 scores to the right of the maximum indicate lower TP rates and consequently, lower sensitivity.
For the initial template matching, F1 maximums are obtained for rmin ≤ 0.50. The highest maximum is 0.91, obtained for tempos 160 and 180 bpm for rmin = 0.30 and 0.35. The lowest maximum is 0.82, obtained for the slowest tempo considered, i.e., 120 bpm, for 0.30 ≤ rmin ≤ 0.40. We can also observe that for rmin > 0.60, F1 scores drop sharply.
Using the criteria of the highest minimum F1 score at a specific rmin, we can conclude that the overall best recognition performance is achieved for rmin = 0.30 and 0.35, for which the F1 scores range from 0.82 for 120 bpm to 0.91 for 180 bpm dance tempos.
As expected, the F1 scores for the final template matching exceed those obtained for the initial template matching. The F1 maximums range from 0.87 to 0.98. The overall best recognition performance is obtained for rmin = 0.35 and 0.40.
These results confirm the potential of using the presented method to recognise repeated dance moves and to assess dancers’ consistency in performance.

3.2.2. Validation Using the Recreational Dancer’s Test Sequences

For the recreational dancers, 293–311 dance moves were collected for each dance tempo considered. The sensitivity, precision, and F1 scores are presented in Figure 6, Figure 7 and Figure 8, respectively.
For the initial template matching, the sensitivity maximums range from 0.74 to 0.82 and are obtained for rmin ≤ 0.40. For the final template matching, they range from 0.89 to 0.98 and are obtained for rmin ≤ 0.30. For all dance tempos considered, the precision scores are equal to 1 for rmin ≥ 0.70 for the initial template matching and rmin ≥ 0.60 for the final template matching.
For the initial template matching, the F1 maximums range from 0.76 to 0.84 and are obtained for rmin ≤ 0.45. For the final template matching, the F1 maximums range from 0.89 to 0.98 and are obtained rmin ≤ 0.30.
The main difference when compared to the professional dancer’s results is that the maximum F1 scores are obtained for lower values of rmin. In addition, after rising to their maximum value, F1 scores fall more sharply. These observations are an expected consequence of individual style and subtle differences in move performances of different dancers.
The results confirm the potential of using the presented method to recognise imitated dance moves. Similar recognition ability can be achieved for the template professional dancer, but at lower values of rmin.
Finally, the recognition ability of the final template matching proves that extracting three device-independent coordinate system axes from the acquired dance motion acceleration signals as presented in the previous section is valid and enables reliable transformation of the inertial signals from the device-intrinsic to the coordinate system of the templates.

4. Conclusions

We have shown that a single wearable device, capturing 3D acceleration and angular velocity of a dancer’s right leg motion, provides for recognising 15 solo jazz dance moves. We have demonstrated that the analysed sequences can be reliably temporally scaled and that dance recognition can be achieved using the same templates, independent of the tempo of dancing. Considering this makes the creation of the template database easier.
Residing on a single device makes the solution highly practical. The practical aspect is further enhanced with the recognition’s independency of the device orientation. The results confirm that even for such a dynamic motion as solo jazz dancing, the acquired acceleration and angular velocity 3D projections can be reliably transformed to a device orientation-invariant system. Exploiting the 3D projections of the signals has a higher discriminative potential and provides for better recognition ability than does considering only the signals’ magnitudes.
The presented solution fills the gap of the existing state of the art: since the presented solution avoids using video cameras or IR imaging sensors, it is not computationally expensive or limited to specific locations; residing on a single wearable device and not being dependent on the orientation of the wearer makes the methodology suitable for unsupervised everyday use and applicable for a variety of dancing situations—whether the dancer is dancing alone, in the crowd, or in front of an audience.
Activity recognition methods relying on wearable devices, developed for ubiquitous measurement scenarios, usually try to balance the opposing requirements of device life autonomy and recognition performance. By relying on a limited set of features extracted from the inertial signals, these methods reduce the computational complexity and prolong the device life autonomy. Since the methodology presented in this article is primarily intended to be used for relatively short dancing sequences, e.g., per song basis, as opposed to throughout-the-day use, more computational power, needed for performing the correlation-based template matching, can be invested in the analysis.
The advantages for the dancer and/or dancing instructor are straightforward: for each performed dance sequence, dance moves can be extracted together with recognition confidence levels, indicating how well the execution matches the reference template. This way, the methodology supports progress monitoring. In addition, the recognised moves can be further analysed and compared to the reference moves in the database to evaluate the overall performance. Finally, while tested on solo jazz dance moves, the methodology presented in this article can be extended to other dance styles, as well.
The presented methodology demonstrates a high recognition ability for the preliminary set of measurements obtained for six dancers. However, a standalone study is needed to further explore in-depth the variability between dancers. Further analysis is also necessary to assess the methodology recognition performance in real dancing scenarios, when the dancer is dancing freely, combining various predefined and improvised moves in sequences. Further studies can also capitalise on this study and assess the dancer’s creativity and response to music, investigating the crucial connection between dancing and music. The benefits of using additional devices, positioned on other body segments, can be investigated. Finally, the presented implementation can be explored further with the aim of optimisation for real-time execution.

Author Contributions

Conceptualisation, S.S.; methodology, S.S.; software, S.S.; validation, S.S.; formal analysis, S.S. and S.T.; investigation, S.S. and S.T.; resources, S.S.; writing, S.S.; visualisation, S.S.; project administration, S.S.; funding acquisition, S.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Slovenian Research Agency, grant number P2-0246 ICT4QoL—Information and communication technologies for quality of life.

Institutional Review Board Statement

This study was conducted during regular dance training hours with the participants performing exactly their regular training activities. More than minimal risks were not identified, and ethical review and approval were waived. The study followed the Code of Ethics of the University of Ljubljana, which provides guidelines for studies involving human beings and is in accordance with the Declaration of Helsinki.

Informed Consent Statement

Informed written consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors give special thanks to Špela Premelč for contributing her professional dancing experience.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A

Appendix A.1. Signal Matrices Similarity

Let M be an arbitrary N × K matrix. Let M further be a matrix with columns equal to the columns of matrix M, standardised to zero mean and unit standard deviation:
M * , k = M * , k μ M * , k σ M * , k , 1 k K ,
where μ and σ denote the mean and standard deviation, respectively, while * denotes all rows.
Let now m be an LK-long vector obtained by vectorising matrix M column-wise:
m = [ M * , 1 T M * , 2 T M * , K T ] T .
In (A2), T denotes the transpose operator.
Following (A1) and (A2), for any two matrices M and N of equal size N × K, we can now define a similarity measure r M N , calculated as the correlation coefficient across the matrices’ K dimension as:
r M N = m T n / ( N K ) .
In particular, instead of M and N, we use signal matrices with columns representing 1 ≤ kK signals and rows representing 1 ≤ nN time samples. The calculated similarity measure (A3) represents the correlation coefficient between two sets of K signals.

Appendix A.2. Dance Move Repetition Extraction

Let S be an Ns × 6 signal matrix, with columns representing the 6 inertial signals and rows representing the signals’ Ns time samples. For a defined starting row n0 and an initial length Nn, let us extract two adjacent Nn × 6 submatrices of S. By standardising these two submatrices according to (A1) and applying (A2), we obtain their respective vector forms. Further applying (A3), we obtain the correlation coefficient, indicating the submatrices’ similarity.
Repeating the aforementioned process for different lengths of the submatrices, Nn = N0 + n, where, in particular for the problem at hand, N0 = 392 and 0 ≤ n ≤ 16, we obtain a vector of correlation coefficients rS. Finally, by identifying the index of the rS maximum, denoted with nmax, we obtain the number of rows Nnmax:
N n m a x = N 0 + n m a x
for which the adjacent submatrices match the most. These two submatrices, i.e., S n 0 : n 0 + N n m a x 1 , * and S n 0 + N n m a x : n 0 + 2 N n m a x 1 , * , are used as signal matrices of two consecutive repetitions of a dance move.

Appendix A.3. Magnitudes Similarity

Let T i ¯ ¯ and T j ¯ ¯ denote NT × 2 matrices of acceleration and angular velocity magnitudes for template moves i and j. By inserting T i ¯ ¯ instead of M in (A1)–(A2), we perform matrix normalisation and vectorisation. We denote the resulting 2NT-long vector with t i ¯ ¯ .
For each 0 ≤ τ < NT, we cyclically shift matrix T j ¯ ¯ rows:
T j ¯ ¯ τ = [ T j ¯ ¯ τ + 1 : N T , * T j ¯ ¯ 1 : τ , * ] .
By inserting T j ¯ ¯ τ instead of M in (A1)–(A2), we perform normalisation and vectorisation. For each τ, we denote the resulting 2NT-long vector with t j ¯ ¯ τ . By further inserting t i ¯ ¯ and t j ¯ ¯ τ instead of m and n in (A3) for each τ, we calculate a component of the similarity vector, denoted with r T i ¯ ¯ T j ¯ ¯ . We use the maximum value of this vector, denoted with r T i ¯ ¯ T j ¯ ¯ , as the two moves’ magnitude similarity measure.

Appendix A.4. Projections Similarity

Let Ti and Tj denote two NT × 6 matrices of 3D acceleration and angular velocity projections for template moves i and j. By inserting Ti instead of M in (A1)–(A2), we perform normalisation and vectorisation. We denote the resulting 6NT-long vector with t i .
For each 0 ≤ τ < NT, we cyclically shift matrix Tj rows:
T j τ = [ T j τ + 1 : N T , * T j 1 : τ , * ] .
By inserting T j τ instead of M in (A1)–(A2), we perform normalisation and vectorisation. For each τ, we denote the resulting 6NT-long vector with t j τ . By further inserting t i and t j τ instead of M and N in (A3) for each τ, we calculate a component of the similarity vector, denoted with r T i T j . We use the maximum value of this vector, denoted with r T i T j , as the two moves’ projection similarity measure.

Appendix A.5. Initial Template Matching

Let X ¯ ¯ denote an NT × 2 matrix of the test sequence 3D acceleration and angular velocity magnitudes. For each 1 ≤ nNXNT + 1, we extract a submatrix of X ¯ ¯ , i.e., X ¯ ¯ n : n + N T 1 , * . By inserting this submatrix instead of M in (A1)–(A2), we perform column-wise standardisation and matrix vectorisation, obtaining a 2NT-long vector denoted with x ¯ ¯ n for each n.
With the signals prepared, we can now calculate the vector of correlation coefficients between the magnitude signals of the sequence and the templates. For each template move i, we denote this vector with r T i ¯ ¯ X ¯ ¯ . For each time lag n, we obtain an element of vector r T i ¯ ¯ X ¯ ¯ by inserting t i ¯ ¯ and x n ¯ ¯ instead of M and N in (A3):
r X ¯ ¯ T i ¯ ¯ ( n ) = t i ¯ ¯ T x n ¯ ¯ / ( 2 N T ) .
Indicating high similarity between T i ¯ ¯ and X ¯ ¯ , local maximums of r T i ¯ ¯ X ¯ ¯ reveal the possible onsets of template move i executions in the test sequence.

Appendix A.6. Final Template Matching

Let X ( R ) denote an NT × 6 matrix of the test sequence 3D accelerations and angular velocities projections on the templates’ coordinate system axes. For each time lag 1 ≤ nNXNT + 1, we extract a submatrix of X ( R ) , i.e., X ( R ) n : n + N T 1 , * . By inserting this submatrix instead of M in (A1)–(A2), we perform column-wise standardisation and matrix vectorisation, obtaining a 6NT-long vector, denoted with x n ( R ) for each n.
We can now calculate the vector of the correlation coefficients between the projection signals of the sequence and the templates. For each template move i, we denote this vector with r T i X ( R ) . For each time lag n, we obtain an element of r T i X ( R ) by inserting t i and x n ( R ) instead of M and N in (A3):
r T i X ( R ) ( n ) = t i T x n ( R ) / ( 6 N T ) .
Indicating high similarity between T i and X ( R ) , local maximums of r T i X ( R ) reveal the possible presence of move i execution in the test sequence.

Appendix B

Orientation Independent Transformation

Let A be an NT × 3 matrix of 3D acceleration signals of a single dance move. The columns of A represent the acceleration projections on the device-intrinsic coordinate system axes and rows represent the time dimension. Following the procedure presented in [20], we extract three orthogonal 3 × 1 unit vectors from A as follows.
The orientation of the first unit vector, denoted with ζ , is defined as the gravity direction, being parallel to the user’s torso, and is calculated as the mean acceleration direction within the motion cycle, i.e., dance move. Denoting a as the mean direction of acceleration during one dance move, we can write:
a = [ μ A * , 1 μ A * , 2 μ A * , 3 ] T ζ = a / a .
In (A9), μ and denote the vector mean and magnitude calculation, respectively, while T denotes the transpose operator. Acceleration projection on axis ζ is then calculated:
A ζ = A ζ
and subtracted from the original 3D acceleration signals, giving acceleration in the plane orthogonal to the gravity direction:
A ζ f = A ζ A ζ T .
The orientation of the second unit vector ξ is defined as the direction covering most of A ζ f variance. Normalised eigenvectors of A ζ f are determined using Principal Component Analysis (PCA). The one associated with the maximum eigenvalue is set as ξ . Finally, the remaining unit vector is defined as being orthogonal to the first two:
ψ = ζ × ξ .
In (A12), x denotes the cross product.
We apply this method to our dance signals. By inserting the 3D acceleration signal matrices of the template and subsequence of the best matching pair, ATM and AXM, respectively, instead of A in (A9)–(A12), we obtain two triplets of orthogonal unit vectors. From ATM, we obtain unit vectors ζ T , ξ T , and ψ T , while from AXM, we obtain three orthogonal unit vectors ζ X , ξ X , and ψ X .

References

  1. Kyan, M.; Sun, G.; Li, H.; Zhong, L.; Muneesawang, P.; Dong, N.; Elder, B.; Guan, L. An Approach to Ballet Dance Training through MS Kinect and Visualization in a CAVE Virtual Reality Environment. ACM Trans. Intell. Syst. Technol. 2015, 6, 1–37. [Google Scholar] [CrossRef]
  2. Aich, A.; Mallick, T.; Bhuyan, H.B.G.S.; Das, P.; Majumdar, A.K. NrityaGuru: A Dance Tutoring System for Bharatanatyam using Kinect. In Computer Vision, Pattern Recognition, Image Processing, and Graphics; Rameshan, R., Arora, C., Dutta Roy, S., Eds.; Springer: Singapore, 2018; pp. 481–493. [Google Scholar] [CrossRef]
  3. Dos Santos, A.D.P.; Yacef, K.; Martinez-Maldonado, R. Let’s dance: How to build a user model for dance students using wearable technology. In Proceedings of the 25th Conference on User Modeling, Adaptation and Personalization, Bratislava, Slovakia, 9–12 July 2017; pp. 183–191. [Google Scholar] [CrossRef]
  4. Drobny, D.; Weiss, M.; Borchers, J. Saltate!: A sensor-based system to support dance beginners. In Proceedings of the 27th Annual CHI Conference on Human Factors in Computing Systems, Boston, MA, USA, 4–9 April 2009; pp. 3943–3948. [Google Scholar] [CrossRef]
  5. Romano, G.; Schneider, J.; Drachsler, H. Dancing Salsa with Machines—Filling the Gap of Dancing Learning Solutions. Sensors 2019, 19, 3661. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Ofli, F.; Erzin, E.; Yemez, Y.; Tekalp, A.M.; Erdem, C.E.; Erdem, A.T.; Abaci, T.; Ozkan, M.K. Unsupervised dance figure analysis from video for dancing Avatar animation. In Proceedings of the 15th IEEE International Conference on Image Processing, San Diego, CA, USA, 12–15 October 2008; pp. 1484–1487. [Google Scholar] [CrossRef] [Green Version]
  7. Yamane, R.; Shakunaga, T. Dance motion analysis by correlation matrix between pose sequences. In Proceedings of the 25th International Conference of Image and Vision Computing, Queenstown, New Zealand, 8–9 November 2010; pp. 1–6. [Google Scholar] [CrossRef]
  8. Shikanai, N.; Hachimura, K. The Effects of the Presence of an Audience on the Emotions and Movements of Dancers. Procedia Technol. 2014, 18, 32–36. [Google Scholar] [CrossRef] [Green Version]
  9. Shikanai, N. Relations between Femininity and the Movements in Japanese Traditional Dance. In Proceedings of the IEEE International Conference on Consumer Electronics–Asia (ICCE–Asia), Bangkok, Thailand, 12–14 June 2019; pp. 146–148. [Google Scholar] [CrossRef]
  10. Kim, D.; Kim, D.H.; Kwak, K.C. Classification of K-Pop Dance Movements Based on Skeleton Information Obtained by a Kinect Sensor. Sensors 2017, 17, 1261. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Bakalos, N.; Protopapadakis, E.; Doulamis, A.; Doulamis, N. Dance Posture/Steps Classification Using 3D Joints from the Kinect Sensors. In Proceedings of the IEEE 16th International Conference on Dependable, Autonomic and Secure Computing, 16th International Conference on Pervasive Intelligence and Computing, 4th International Conference on Big Data Intelligence and Computing and Cyber Science and Technology Congress (DASC/PiCom/DataCom/CyberSciTech), Athens, Greece, 12–15 August 2018; pp. 868–873. [Google Scholar] [CrossRef]
  12. Ho, C.; Tsai, W.; Lin, K.; Chen, H.H. Extraction and alignment evaluation of motion beats for street dance. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 2429–2433. [Google Scholar] [CrossRef]
  13. Cornacchia, M.; Ozcan, K.; Zheng, Y.; Velipasalar, S. A Survey on Activity Detection and Classification Using Wearable Sensors. IEEE Sens. J. 2017, 17, 386–403. [Google Scholar] [CrossRef]
  14. Siirtola, P.; Röning, J. Recognizing Human Activities User-independently on Smartphones Based on Accelerometer Data. Int. J. Artif. Intell. Interact. Multimed. 2012, 1, 38–45. [Google Scholar] [CrossRef]
  15. Long, X.; Yin, B.; Aarts, R.M. Single-accelerometer-based daily physical activity classification. In Proceedings of the 31st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC 2009), Minneapolis, MN, USA, 3–6 September 2009; pp. 6107–6110. [Google Scholar] [CrossRef] [Green Version]
  16. Lara, O.D.; Labrador, M.A. A Survey on Human Activity Recognition using Wearable Sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar] [CrossRef]
  17. Niemann, F.; Reining, C.; Moya Rueda, F.; Nair, N.R.; Steffens, J.A.; Fink, G.A.; ten Hompel, M. LARa: Creating a Dataset for Human Activity Recognition in Logistics Using Semantic Attributes. Sensors 2020, 20, 4083. [Google Scholar] [CrossRef]
  18. Sousa Lima, W.; Souto, E.; El-Khatib, K.; Jalali, R.; Gama, J. Human Activity Recognition Using Inertial Sensors in a Smartphone: An Overview. Sensors 2019, 19, 3213. [Google Scholar] [CrossRef] [Green Version]
  19. Shoaib, M.; Bosch, S.; Incel, O.D.; Scholten, H.; Havinga, P.J.M. Complex Human Activity Recognition Using Smartphone and Wrist-Worn Motion Sensors. Sensors 2016, 16, 426. [Google Scholar] [CrossRef] [PubMed]
  20. Gadaleta, M.; Rossi, M. IDNet: Smartphone-based gait recognition with convolutional neural networks. Pattern Recognit. 2018, 74, 25–37. [Google Scholar] [CrossRef] [Green Version]
  21. Sprager, S.; Juric, M.B. Inertial Sensor-Based Gait Recognition: A Review. Sensors 2015, 15, 22089–22127. [Google Scholar] [CrossRef]
  22. Cola, G.; Avvenuti, M.; Vecchio, A. Real-time identification using gait pattern analysis on a standalone wearable accelerometer. Comput. J. 2017, 60, 1173–1186. [Google Scholar] [CrossRef]
  23. Junker, H.; Amft, O.; Lukowicz, P.; Tröster, G. Gesture spotting with body-worn inertial sensors to detect user activities. Pattern Recognit. 2008, 41, 2010–2024. [Google Scholar] [CrossRef]
  24. Stančin, S.; Tomažič, S. Early Improper Motion Detection in Golf Swings Using Wearable Motion Sensors: The First Approach. Sensors 2013, 13, 7505–7521. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Martínez, A.; Jahnel, R.; Buchecker, M.; Snyder, C.; Brunauer, R.; Stöggl, T. Development of an Automatic Alpine Skiing Turn Detection Algorithm Based on a Simple Sensor Setup. Sensors 2019, 19, 902. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Kos, A.; Umek, A. Smart sport equipment: SmartSki prototype for biofeedback applications in skiing. Pers. Ubiquitous Comput. 2018, 22, 535–544. [Google Scholar] [CrossRef]
  27. Benages Pardo, L.; Buldain Perez, D.; Orrite Uruñuela, C. Detection of Tennis Activities with Wearable Sensors. Sensors 2019, 19, 5004. [Google Scholar] [CrossRef] [Green Version]
  28. Dadashi, F.; Millet, G.P.; Aminian, K. A Bayesian approach for pervasive estimation of breaststroke velocity using a wearable IMU. Pervasive Mob. Comput. 2015, 19, 37–46. [Google Scholar] [CrossRef]
  29. Ghasemzadeh, H.; Jafari, R. Coordination Analysis of Human Movements With Body Sensor Networks: A Signal Processing Model to Evaluate Baseball Swings. IEEE Sens. J. 2011, 3, 603–610. [Google Scholar] [CrossRef]
  30. Paradiso, J.A.; Hsiao, K.; Benbasat, A.Y.; Teegarden, Z. Design and implementation of expressive footwear. IBM Syst. J. 2000, 39, 511–529. [Google Scholar] [CrossRef] [Green Version]
  31. Aylward, R.; Lovell, S.D.; Paradiso, J.A. A Compact, Wireless, Wearable Sensor Network for Interactive Dance Ensembles. In Proceedings of the International Workshop on Wearable and Implantable Body Sensor Networks, Cambridge, MA, USA, 3–5 April 2006. [Google Scholar] [CrossRef] [Green Version]
  32. Dang, Q.K.; Pham, D.D.; Suh, Y.S. Dance training system using foot mounted sensors. In Proceedings of the 2015 IEEE/SICE International Symposium on System Integration (SII), Nagoya, Japan, 11–13 December 2015; pp. 732–737. [Google Scholar] [CrossRef]
  33. Kim, Y.; Jung, D.; Park, S.; Chi, J.; Kim, T.; Lee, S. The Shadow Dancer: A New Dance Interface with Interactive Shoes. In Proceedings of the 2008 International Conference on Cyberworlds, Hangzhou, China, 22–24 September 2008; pp. 745–748. [Google Scholar] [CrossRef]
  34. Tragtenberg, J.; Calegario, F.; Cabral, G.; Ramalho, G. TumTá and Pisada: Two Foot-controlled Digital Dance and Music Instruments Inspired by Popular Brazillian Traditions. In Proceedings of the 17th Brazilian Symposium on Computer Music (SBCM 2019), São João del-Rei, Brazil, 25–27 September 2019. [Google Scholar] [CrossRef]
  35. Yamaguchi, T.; Ariga, A.; Kobayashi, T.; Hashimoto, S. TwinkleBall: A Wireless Musical Interface for Embodied Sound Media. In Proceedings of the New Interfaces for Musical Expression (NIME 2010), Sydney, Australia, 15–18 June 2010; pp. 116–119. [Google Scholar]
  36. Samprita, S.; Koshy, A.S.; Megharjun, V.N.; Talasila, V. LSTM-Based Analysis of a Hip-Hop Movement. In Proceedings of the 2020 6th International Conference on Control, Automation and Robotics (ICCAR), Singapore, 20–23 April 2020; pp. 519–524. [Google Scholar] [CrossRef]
  37. Hinton-Lewis, C.W.; McDonough, E.; Moyle, G.M.; Thiel, D.V. An Assessment of Postural Sway in Ballet Dancers During First Position, Relevé and Sauté with Accelerometers. Procedia Eng. 2016, 147, 127–132. [Google Scholar] [CrossRef] [Green Version]
  38. Thiel, D.V.; Quandt, J.; Carter, S.J.L.; Moyle, G.M. Accelerometer based performance assessment of basic routines in classical ballet. Procedia Eng. 2014, 72, 14–19. [Google Scholar] [CrossRef] [Green Version]
  39. Stančin, S.; Tomažič, S. Dance Tempo Estimation Using a Single Leg-Attached 3D Accelerometer. Sensors 2021, 21, 8066. [Google Scholar] [CrossRef] [PubMed]
  40. Mbientlab MMR. Available online: https://mbientlab.com/metamotionr/ (accessed on 21 December 2021).
  41. Alphabetical Jazz Steps 3. Available online: https://www.youtube.com/watch?v=jAIwJd2tQo0&list=PLpLDojUPSMvcYMA7jEFPidEbSD2-vNz8m (accessed on 21 December 2021).
  42. MATLAB; Version 9.11.0.1769968 (R2021b); The MathWorks Inc.: Natick, MA, USA, 2021.
  43. Stančin, S.; Tomažič, S. Time-and Computation-Efficient Calibration of MEMS 3D Accelerometers and Gyroscopes. Sensors 2014, 14, 14885–14915. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Dance move recognition processing workflow.
Figure 1. Dance move recognition processing workflow.
Sensors 22 02446 g001
Figure 2. Examples of acceleration (first row) and angular velocity (second row) of right leg motion during two solo jazz moves. The first three columns show the 3D projections and the last column illustrates the magnitudes.
Figure 2. Examples of acceleration (first row) and angular velocity (second row) of right leg motion during two solo jazz moves. The first three columns show the 3D projections and the last column illustrates the magnitudes.
Sensors 22 02446 g002
Figure 3. Sensitivity scores obtained for the professional dancer’s test sequences.
Figure 3. Sensitivity scores obtained for the professional dancer’s test sequences.
Sensors 22 02446 g003
Figure 4. Precision scores obtained for the professional dancer’s test sequences.
Figure 4. Precision scores obtained for the professional dancer’s test sequences.
Sensors 22 02446 g004
Figure 5. F1 scores obtained for the professional dancer’s test sequences.
Figure 5. F1 scores obtained for the professional dancer’s test sequences.
Sensors 22 02446 g005
Figure 6. Sensitivity scores obtained for the recreational dancers’ test sequences.
Figure 6. Sensitivity scores obtained for the recreational dancers’ test sequences.
Sensors 22 02446 g006
Figure 7. Precision scores obtained for the recreational dancers’ test sequences.
Figure 7. Precision scores obtained for the recreational dancers’ test sequences.
Sensors 22 02446 g007
Figure 8. F1 scores obtained for the recreational dancers’ test sequences.
Figure 8. F1 scores obtained for the recreational dancers’ test sequences.
Sensors 22 02446 g008
Table 1. Solo jazz template moves’ pairwise similarities calculated using the magnitudes of acceleration and angular velocity. Depicted in red are the highest similarities obtained for each move, excluding self-similarity.
Table 1. Solo jazz template moves’ pairwise similarities calculated using the magnitudes of acceleration and angular velocity. Depicted in red are the highest similarities obtained for each move, excluding self-similarity.
Move(1)(2)(3)(4)(5)(6)(7)(8)(9)(10)(11)(12)(13)(14)(15)
(1)1.000.220.450.290.390.290.370.270.350.350.320.460.270.310.31
(2) 1.000.190.210.350.470.290.440.340.280.330.330.270.300.37
(3) 1.000.420.390.280.300.330.350.260.360.550.330.430.30
(4) 1.000.370.290.310.360.390.240.430.550.320.270.43
(5) 1.000.330.370.370.290.260.290.480.230.480.38
(6) 1.000.380.420.330.320.440.380.240.360.44
(7) 1.000.610.200.370.310.320.250.330.28
(8) 1.000.270.340.470.390.340.300.42
(9) 1.000.220.290.240.300.430.31
(10) 1.000.290.350.310.270.30
(11) 1.000.470.540.260.35
(12) 1.000.390.310.43
(13) 1.000.190.32
(14) 1.000.43
(15) 1.00
Table 2. Solo jazz template moves’ pairwise similarities calculated using the 3D projections of the acceleration and angular velocity. Depicted in red are the highest similarities obtained for each move, excluding self-similarity.
Table 2. Solo jazz template moves’ pairwise similarities calculated using the 3D projections of the acceleration and angular velocity. Depicted in red are the highest similarities obtained for each move, excluding self-similarity.
Move(1)(2)(3)(4)(5)(6)(7)(8)(9)(10)(11)(12)(13)(14)(15)
(1)1.000.180.280.210.230.150.230.150.260.320.250.190.100.230.25
(2) 1.000.220.100.220.230.200.230.120.210.150.220.140.290.19
(3) 1.000.210.400.180.250.100.190.1860.180.320.130.380.18
(4) 1.000.260.190.130.120.260.170.430.340.110.160.40
(5) 1.000.230.350.140.170.200.260.330.100.340.25
(6) 1.000.330.220.170.180.180.160.130.230.20
(7) 1.000.210.110.230.250.150.120.380.26
(8) 1.000.230.150.110.170.090.240.10
(9) 1.000.190.190.300.200.150.26
(10) 1.000.180.210.290.170.22
(11) 1.000.280.230.140.39
(12) 1.000.180.270.29
(13) 1.000.120.18
(14) 1.000.19
(15) 1.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Stančin, S.; Tomažič, S. Recognizing Solo Jazz Dance Moves Using a Single Leg-Attached Inertial Wearable Device. Sensors 2022, 22, 2446. https://doi.org/10.3390/s22072446

AMA Style

Stančin S, Tomažič S. Recognizing Solo Jazz Dance Moves Using a Single Leg-Attached Inertial Wearable Device. Sensors. 2022; 22(7):2446. https://doi.org/10.3390/s22072446

Chicago/Turabian Style

Stančin, Sara, and Sašo Tomažič. 2022. "Recognizing Solo Jazz Dance Moves Using a Single Leg-Attached Inertial Wearable Device" Sensors 22, no. 7: 2446. https://doi.org/10.3390/s22072446

APA Style

Stančin, S., & Tomažič, S. (2022). Recognizing Solo Jazz Dance Moves Using a Single Leg-Attached Inertial Wearable Device. Sensors, 22(7), 2446. https://doi.org/10.3390/s22072446

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop