Literature DB >> 35408060

Recognizing Solo Jazz Dance Moves Using a Single Leg-Attached Inertial Wearable Device.

Sara Stančin1, Sašo Tomažič1.   

Abstract

We present here a method for recognising dance moves in sequences using 3D accelerometer and gyroscope signals, acquired by a single wearable device, attached to the dancer's leg. The recognition entails dance tempo estimation, temporal scaling, a wearable device orientation-invariant coordinate system transformation, and, finally, sliding correlation-based template matching. The recognition is independent of the orientation of the wearable device and the tempo of dancing, which promotes the usability of the method in a wide range of everyday application scenarios. For experimental validation, we considered the versatile repertoire of solo jazz dance moves. We created a database of 15 authentic solo jazz template moves using the performances of a professional dancer dancing at 120 bpm. We analysed 36 new dance sequences, performed by the professional and five recreational dancers, following six dance tempos, ranging from 120 bpm to 220 bpm with 20 bpm increment steps. The recognition F1 scores, obtained cumulatively for all moves for different tempos, ranged from 0.87 to 0.98. The results indicate that the presented method can be used to recognise repeated dance moves and to assess the dancer's consistency in performance. In addition, the results confirm the potential of using the presented method to recognise imitated dance moves, supporting the learning process.

Entities:  

Keywords:  3D accelerometer; 3D gyroscope; dancing; inertial sensors; motion analysis; motion recognition; solo jazz

Mesh:

Year:  2022        PMID: 35408060      PMCID: PMC9003112          DOI: 10.3390/s22072446

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

Dancing is an engaging physical activity that requires high levels of body control, skill, and physical fitness. In addition, dancing is an activity cultivating the capacity for creative expression. Following the rhythmic structure, the dancer, especially the professional, expresses his or her interpretation of music by selecting, performing, and assembling various moves in dance sequences. The main motivation of the research presented is to provide a mechanism for dance move recognition in performed sequences. In general, it has already been reported that dancers can greatly benefit from various assistive technologies [1,2,3,4,5]. A dance move recognition technology would benefit dancers engaged in the learning process. Namely, as one learns to dance, he or she tends to imitate performances demonstrated by instructors. These imitations are visually assessed, usually using a mirror. In addition, instructors provide their students with feedback on how their performances compare to the ideal template. Providing such feedback is usually more challenging in group sessions, in which more students engage in the learning process at the same time. A technological solution would benefit both the student and the instructor and would support performance assessment and progress monitoring. More experienced dancers would also benefit. Recognising dance moves would enable further investigation and comparison of the sequences of moves, supporting overall creativity evaluation, individual style classification, and observations of differences when dancing alone, in pairs and/or in the presence of an audience. Additionally combining the recognised dance moves with information about the accompanying music would illuminate how the dancer interprets and responds to music. Relying on a dance move recognition technology during the creative process of producing a choreography, a professional dancer would also be able to create a transcript of the assembled moves for later reference. A segment of dance motion analysis approaches made so far relies on signals obtained from optical motion capture systems, incorporating multiple cameras watching the stage and usually requiring reflective markers to be positioned on the subject’s body [6,7,8,9]. In [6], the authors present a framework for capturing the posture of the dancer’s body. Using the body posture parameters, signals are temporally segmented, motion patterns are extracted, and motion animation in generated. In [7], the authors present dance pose sequences similarity estimation. Another segment of approaches relies on signals obtained from a Kinect device [1,2,5,10,11,12]. The Kinect device incorporates a depth camera together with a standard video camera into a single device, distinguishes human body parts and joints, and estimates their position and orientation. In [1], the authors present a novel framework for the real-time capture, assessment, and visualisation of ballet dance movements. The dance sequence captured is segmented and cross-referenced against a library of gestural components performed by a teacher. Recognition performance was evaluated on a database of six isolated basic ballet positions recordings made by a teacher and a student. Reported average recognition rates are between 90.5% and 99.5%. Bharatanatyam [2] and a salsa dancing coach application [5] have also been developed. In [10], another Kinect-based system is presented enabling Korean pop (K-pop) motion classification by extracting statistical features, reducing dimensionality with Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA) and incorporating extreme learning machine. Skeletal joint angles for 200 movement types obtained for four dancers were considered and the reported maximum classification rate is 96.5%. In [11], the authors analysed data on Greek traditional dance postures with the goal of extracting robust summaries and providing end users or dance experts a concise and meaningful abstract of dance movements. Although valuable results have been reported, using equipment located in the surrounding area, either one or more video cameras or a Kinect device, limits the practical aspect of the methodology, making it difficult or altogether impossible to dance in the ballroom. Moreover, processing video signals is computationally intensive, and although the underlying technology and algorithms are steadily improving, computer vision is still limited by lighting and clutter constraints and subtraction of dynamic background. The presented method capitalises on the benefits of small and lightweight microelectromechanical (MEMS) inertial sensors. Over the past few years, it has been consistently demonstrated that these sensors are an efficient tool in the broader research area of human motion analysis [13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29]. Their characteristic light weight, small size, low power consumption, portability, ease of use, and low cost pave the way for their omnipresence and indispensability in motion analysis. Most of the research conducted in this field concerns daily activity recognition [13,14,15,16,17,18,19], gait analysis [20,21,22], gesture recognition [23], and sports activity performance assessment for activities such as golf, tennis, baseball, swimming, and skiing [24,25,26,27,28,29]. One of the first wearable systems for capturing dance gestures were developed by the MIT Media Lab [30,31], and various assistive solutions have been presented since [3,4,32,33,34,35,36,37,38,39]. In this context, wearable devices are usually used to detect dancer’s steps and provide them with feedback [3,4,32] or enable the dancer to interact or generate sounds [33,34,35]. In [36], a three-axis inertial measurement unit (IMU) is positioned on the subject’s right forearm and used to analyse one specific hip hop arm move. In particular, the aim is to divide the motion into phases, predict which phase is executed, and give feedback to the subject. In addition, accelerometers are used to estimate torso tilt [37] and assess performance [38] of basic routines in classical ballet. We focus our analysis on the solo jazz dance style, a rhythmical and playful solo dance in which the dancer depicts jazz music through one’s one movement. Solo jazz evolved through the first half of the 20th century to include elements of both African and European dance and features the vocabulary and steps of the vernacular jazz tradition. To enable the dancer to follow the jazz song rhythmical structure, as a rule, a single solo jazz move is performed following eight music beats. The music beat is considered as the smallest time interval between two successive notes. Dance moves comprise various motion elements, performed in a specified order. Besides steps, motion elements can be, for example, kicks, taps, and jumps. The variety of motion elements and the order of their execution brings a pleiad of predefined, authentic solo jazz dance moves. Due to this variety, solo jazz fits perfectly into our dance move recognition research polygon. In general, in dancing, the speed of move execution represents the dance tempo and is directly related to the musical tempo of the song. It is measured as the number of steps, or other motion elements, a dancer performs in a minute. As a rule, solo jazz is usually danced to jazz music with tempo anywhere between 120 and 250 beats per minute (bpm). Tempos between 140 and 160 bpm are considered medium. Tempos above 220 bpm are considered fast and generally prove to be too high for the recreational dancers to execute with accuracy and ease. Aiming to provide ease of use, with the smallest amount of sensing equipment, our dance move recognition methodology relies on a single wearable unit, comprising a MEMS 3D accelerometer and gyroscope, positioned on the dancer’s leg, and enables dance move recognition, regardless of the device orientation and dance tempo. The methodology presented considers that different moves have different signal shapes in the time domain and that these shapes can be used as dance move fingerprints. Furthermore, since position and attitude estimates obtained using inertial sensors are known to be characterised by different inaccuracies, especially drift, the methodology resides on the shapes of the original acceleration and angular velocity signals instead. The methodology presented is a continuation of our previous work and features solo jazz dance tempo estimation, as presented in [39]. In addition, the specific technique presented in [20], used for defining a wearable device orientation-invariant coordinate system using the 3D accelerometer signals acquired during walking, is adapted and applied in the research presented in this article. The remainder of this article is organised as follows. In Section 2, we present the materials and methods used for creating the dance move template database, recognition, and experimental validation. In Section 3, we present and discuss the results. Finally, in Section 4, we summarise our findings and draw conclusions, implying further research directions. In all the subsequent sections, we use the following notation rules: large bold letters denote matrices, small bold letters denote vectors, and large or small italics denote scalars.

2. Materials and Methods

2.1. Data Acquisition

2.1.1. Materials

We captured dance motion using a single mbientlab MetaMotionR (MMR) wearable device [40], including a MEMS 3D accelerometer and gyroscope with respective measurement ranges ± 16 g and ± 2000°/s, placed directly above the dancer’s right leg ankle. We hypothesised that analysing the motion of a single leg is sufficient to distinguish between various dance moves. An alternative solution would be to place the device on the dancer’s lower back, since leg motion of high-quality dance performances is usually initiated from the pelvic region, torso, or even the shoulders. However, preliminary investigations have shown that this significantly reduces the recognition performance of the designed solution. The micro-position and orientation of the device are arbitrary. We set the sampling frequency to 200 Hz, which proved to be sufficient for the problem at hand using empirical evidence. The 3D accelerometer and gyroscope together provided 6 inertial signals, each representing one variable in a 6D space. In addition, a software metronome with an established beat frequency error of 1 bpm was used to simulate a steady quarter note music tempo and dictate the tempo of dancing.

2.1.2. Measurements

Six female dancers participated in the study—five recreational dancers (age: 33 ± 5 years, height: 163 ± 5 cm) with over three years of experience in solo jazz dancing and one professional (age: 35 years, height: 164 cm). Altogether, I = 15 authentic solo jazz moves were considered: (1) Tackie Annie, (2) Fall of the log, (3) Kicks, (4) Half break, (5) Struttin’, (6) Savoy kick, (7) 20s Charleston, (8) Knee slaps, (9) Fishtails, (10) Apple Jacks, (11) Boogie back, (12) Boogie forward, (13) Crazy leg, (14) Cross step, and (15) Shorty George. For an informative overview of how these moves are performed, the reader is referred to various sources available online, e.g., [41]. Each move was performed on an 8-beat basis. Measurements were performed in two sessions. Following the obtained instructions, participants attached the wearable device above their right leg ankle, voluntarily setting its micro-position and orientation. The first measurement session was conducted only for the professional dancer. The dancer performed several consecutive repetitions for each of the 15 considered moves, as consistently as possible, following a fixed reference tempo υ 120 bpm, dictated by the metronome. For each move, the dancer started dancing from a still position. We refer to the obtained set of six inertial signals as the learning sequence. The second measurement session was conducted for all six participants. Each was given the task of performing 5 repetitions of each of the 15 considered moves in a prescribed order. The recreational dancers were instructed to mimic the professional’s execution of moves as closely as possible. This task was repeated for 6 different dance tempos, ranging from 120 bpm to 220 bpm with 20 bpm increments. For each dancer, we obtained 6 sets of inertial signals. We refer to these 36 sets as test sequences. All measurements were supplemented with video recordings. The study was conducted in Ljubljana, Slovenia. It followed the Code of Ethics of the University of Ljubljana, which provides guidelines for studies involving human beings and is in accordance with the Declaration of Helsinki. All participants gave written informed consent.

2.2. Signal Processing Overview

The implemented signal processing workflow is presented in Figure 1. For each acquired sequence (1), signal pre-processing (2) is applied. The pre-processed learning sequence is used to extract template moves and create the template database (3), following a semi-automatic correlation-based segmentation procedure. The pre-processed test sequences are analysed and searched for template moves (4), according to the following steps:
Figure 1

Dance move recognition processing workflow.

Dance tempo estimation and signal temporal scaling, achieved using a bank of enhanced comb filters as presented in [39]; Initial template matching, performed on a sliding correlation basis, using the magnitudes of the temporally scaled acceleration and angular velocity; Signal transformation to the templates’ coordinate system; and Final template matching, performed again on a sliding correlation basis, but by using the acceleration and angular velocity 3D projections of on the template coordinate system axes instead of their magnitudes. For both the initial and final template matching, correlation is used as the only feature for recognition. In the following, all the components of the proposed method are presented in detail. All signal processing was executed offline in the MATLAB 2021 environment [42].

2.3. Signal Pre-Processing

Following the calibration procedure presented in [43], we compensated all signals of the learning and test sequences for sensor inaccuracies. Since the device itself does not provide outputs at exactly equidistant time samples, we interpolated and decimated the acquired signals as necessary, considering the associated measurement timestamp values, to provide for uniform sampling at exactly 200 Hz. To remove motion artefacts and noise, we applied a low-pass filter with a cut-off frequency f = 50 Hz and finally performed downsampling to f = 100 Hz, obtaining 3D acceleration and angular velocities at equidistant time samples T = 1/f = 0.01 s.

2.4. Templates’ Database

2.4.1. Template Extraction

Using all N acquired samples of the pre-processed learning sequence, we formed two N × 3 signal matrices, and representing 3D acceleration and angular velocity, respectively. Columns of these two matrices are equal to the respective signal projections on the device-intrinsic coordinate system axes x, y, and z, while rows represent the time samples. The specific orientation of axes x, y, and z in a reference coordinate system, defined by the set position of the wearable device, is arbitrary. We further combined and in a common N × 6 signal matrix S: Columns of S are equal to the 3D acceleration and angular velocity x, y, and z projections, while rows represent the time dimension. Comprising the inertial signals of the entire learning sequence, matrix S comprises all performed repetitions of all considered dance moves, used for creating the templates’ database. We extracted the comprised dance move repetitions by partitioning S into submatrices in the vertical dimension. We achieved this by estimating the repetitions’ onsets and offsets, for each 1 ≤ i ≤ I = 15 dance move separately, following a semi-automatic, correlation-based segmentation procedure as follows. First, since for each dance move, the consecutive repetitions are performed from a still position, we defined the onset of the first repetition by visual inspection of the six inertial signals and the accompanying video. We denote the matrix S row index corresponding to this onset as n. For defining the offset of the first repetition, we consider, given the known reference dance tempo υ, the expected length, expressed as the number of samples NT, of a single dance move. Since each considered solo jazz move is performed on an 8-beat basis, at f = 100 Hz sampling and precise 120 bpm dancing (T = 0.5 s), the expected length is NT = 8 × 0.5 s × 100 Hz = 400 samples. Since dancing is rarely this precise, we allow for a 2% length deviation and set the expected length to be in between 0.98 × NT = 392 and 1.02 × NT = 408 samples. Each length from this interval, i.e., N = N + n, where N = 392 and 0 ≤ n ≤ 0.04 × NT = 16, gives a candidate for the first repetition offset, n + N − 1. Since the repetitions are consecutive, for each offset candidate for the first repetition, n + N − 1, there is a single onset candidate for the second repetition, n + N. Finally considering both repetitions to be equal in length gives a single offset candidate for the second repetition, n + 2N − 1. For each n, we obtained one candidate pair of the first two repetitions of a move, represented with adjacent, same-size N × 6 submatrices of S. We extracted and standardised to zero mean and unit standard deviation these two submatrices and performed column-wise vectorisation, obtaining, for each n, two 6(N + n) long column vectors. Pursuing the highest similarity, for each candidate pair n, we calculated the correlation coefficient between these two vectors. This calculation, supported with implementation equations, is described in detail in Appendix A.2. We set the first and the second repetition of the considered move to be equal to the repetitions of the candidate pair with the highest correlation coefficient. We finally extracted all possible additional repetitions of the particular move on an equal correlation-based search basis using the onset of the last determined repetition as the new n. Due to slight variations in execution duration, we unified in length all extracted repetitions by temporal scaling to the exact expected length NT = 400 samples. For each move i, we composed a set of consistent repetitions, eliminating repetition outliers, determined by visual inspection. We averaged this set column-wise, obtaining finally, for each template move i, an associated single NT × 6 matrix Denoting the first three columns, representing the 3D acceleration, with A and the second three, representing 3D angular velocity, with for each 1 ≤ i ≤ I = 15, we can write: The 15 matrices of the acceleration and angular velocity projections on the device intrinsic coordinate system axes x, y, and z represent our database of template moves. The specific orientation of the axes x, y, and z in the reference coordinate system, defined by the arbitrarily set position of the wearable device during the learning sequence acquisition, defines the coordinate system of the templates. In addition to T, for each move, i we also calculated an NT × 2 matrix having the first and second column equal to the magnitudes of the 3D acceleration and angular velocity, respectively. Each k-th row of is obtained according to: In (3), * denotes all columns.

2.4.2. Templates Similarity Measures

To estimate the similarity between template moves, we used two correlation-based measures as follows. Firstly, we compared the template moves along the acceleration and angular velocity magnitudes. To achieve this, we standardised the columns of to zero mean and unit standard deviation for each 1 ≤ i ≤ I = 15 move and performed column-wise vectorisation of the resulting matrix. For each move, we obtained a 2NT-long vector of magnitudes. For each pair of template moves i and j, 1 ≤ i,j ≤ I = 15, we calculated the correlation between the associated vectors as a function of cyclical shift 0 ≤ τ < NT of the acceleration and angular velocity vector parts. We define the maximum value of this correlation for each i,j pair of template moves, denoted with as the magnitudes’ similarity. This calculation is supported in detail with implementation equations in Appendix A.3. We also compared the template moves along the acceleration and angular velocity 3D projections. We standardised the columns of T to zero mean and unit standard deviation for each 1 ≤ i ≤ I = 15 move and performed column-wise vectorisation of the resulting matrix. For each move, we obtained a 6NT -long vector. For each pair of template moves i and j, 1 ≤ i,j ≤ I = 15, we again calculated the correlation between the associated vectors as a function of cyclical shift 0 ≤ τ < NT of the projections vector parts. We define the maximum value of this correlation for each i,j pair of template moves, denoted with as the projections’ similarity. This calculation is supported in detail with implementation equations in Appendix A.4. In the same way as the correlation coefficient calculated along one dimension, and can take any value from the range [−1,1], where 1 indicates identical moves, 0 orthogonal moves, and –1 opposite moves. High values suggest common leg activation patterns, regardless of the actual direction of motion in 3D. On the other hand, high suggests that two moves also match with respect to the direction of the executing motion. Considering this, the similarity measure is expected to have a higher discriminative potential. Note that by performing standardisation column-wise, equal weight is given to each dimension of comparison, x, y, and z, regardless of the intensity of motion along that dimension.

2.5. Dance Move Recognition

2.5.1. Dance Tempo Estimation and Temporal Scaling

For each pre-processed test sequence, we first estimate the dance tempo υ, using a method based on multiple resonators, implemented with enhanced comb feedback filters, as presented in [6]. We temporally scale all acquired inertial signals of the sequence by a factor of υ/υ to obtain a test sequence with dance moves that match the template moves in tempo. We denote with N the final length of the test sequence. Using all N samples, we form two N × 3 signal matrices of 3D acceleration and angular velocity, denoted with and respectively. The columns of these two matrices correspond to the respective inertial signal projections on the intrinsic coordinate system axes of the wearable device. Since the orientation of the wearable device is different for each measurement session, these axes and the axes of the coordinate system of the template moves are generally not aligned. We further form a common N × 6 test sequence signal matrix X: Columns of X correspond to the acceleration and angular velocity 3D projections while rows represent the time dimension. We also form an N × 2 matrix of acceleration and angular velocity magnitudes Each k-th row of is calculated according to: In (5), * denotes all columns.

2.5.2. Initial Template Matching

We search the time-scaled test sequence for template moves by applying sliding correlation-based template matching, initially considering the acceleration and angular velocity magnitudes. For the template i and the test sequence, these are comprised in matrices (3) and (5), respectively. For each template move 1 ≤ i ≤ I = 15, we first standardise the columns of to zero mean and unit standard deviation. By vectorising the standardised matrices column-wise, for each template move i, we obtain a 2NT-long vector of the acceleration and angular velocity magnitudes. We further apply a 2D sliding window of size NT × 2 to the test sequence signal matrix . For each 1 ≤ n ≤ N − NT + 1, we extract an NT × 2 submatrix of , i.e., . After standardising and vectorising we obtain a 2NT-long vector of the associated acceleration and angular velocity magnitudes. We then calculate the correlation coefficient between this vector and . Sliding the 2D window vertically, we repeat this process for all n and obtain N − NT + 1 correlation coefficients. We store the results for each template move i in a vector denoted with This calculation is supported in detail with implementation equations in Appendix A.5. Indicating high similarity between , the local maximums of reveal the possible onsets of template move i executions in the test sequence. For each dance move i, we identify the local maximums of that exceed a certain threshold value, denoted with r. We define the indices of these local maximums as move i onsets in the sequence. We further define the associated values of local maximums as the recognition confidence. After searching through the entire sequence for all template moves, we performed a final correction in the sense of false positives detection and elimination: if any two onsets of moves were closer than 0.7 NT, we discarded the one with the lower recognition confidence.

2.5.3. Signal Transformation

To transform the signals of the test sequence from the device intrinsic to the coordinate system of the templates, we first identify from all onsets found in the previous step, for all i and for all n, the one with the highest recognition confidence. This onset defines the strongest matching pair: for i and n, for which the recognition confidence is the highest, the magnitude matrices of the template and subsequence, i.e., and respectively, match the most. We denote the corresponding matrices of the acceleration 3D projections with A and A and the combined matrices of the accelerations and angular velocity 3D projections with T and X. Since this template–subsequence pair represents the same move, but in different coordinate systems, it is reasonable to assume that a rotation that aligns them can be defined. Moreover, if we determine such a rotation, we can apply it to the entire test sequence, obtaining the acceleration and angular velocity 3D projections in the coordinate system of the templates. To define this rotation, we adapted and applied an orientation-independent transformation, as presented in [20]. The original procedure estimates a wearable device orientation-independent coordinate system by calculating three orthogonal unit vectors from the 3D accelerometer signals, acquired using a smartphone in the front pocket of the user’s trousers during walking. The first unit vector, denoted with is defined by the gravity direction, and is calculated as the mean acceleration within a motion (walking) cycle. The second unit vector, denoted with is considered to be aligned with the direction of motion and is determined using PCA as the direction with the highest variance once the projection on has been subtracted. The third unit vector, denoted with is defined as orthogonal to the first two and as such covers lateral motion during the motion cycle. In our dance analysis context, both the template and subsequence of the best-matching pair represent one full dance move, i.e., motion cycle. Considering this, following the abovementioned, from A, we obtain three orthogonal unit vectors and while from A, we obtain three orthogonal unit vectors and Extracting unit vectors from the acceleration 3D projections has already been presented in detail in [20]; we provide details specific to this study in Appendix B. Combining both triplets of unit vectors, and we can now define a rotation matrix R that aligns the coordinate system of the test sequence with that of the templates as: Using R (6), we can transform the original acceleration and angular velocity 3D projections, comprised in matrices and respectively, from the test sequence’s device intrinsic to the coordinate system of the template moves. Denoting the resulting matrices with and we write: The common N × 6 matrix of the rotated acceleration and angular velocity 3D projections of the test sequence, denoted with given in the coordinate system of the templates, is then:

2.5.4. Final Template Matching

We can now search the test sequence for template moves, following a similar procedure as presented in Section 2.5.2., only this time considering the acceleration and angular velocity 3D projections instead of their magnitudes. For the template moves and the test sequence, these are comprised in matrices (2) and (6), respectively. For each template move 1 ≤ i ≤ I = 15, we first standardise to zero mean and unit standard deviation the columns of . Vectorising the standardised matrices column-wise, for each template move i, we obtain a 6NT-long vector of acceleration and angular velocity projections t. We further use a 2D sliding window of size NT × 6, and for each 1 ≤ n ≤ N − NT + 1, we extract an NT × 6 submatrix of i.e., After standardisation and vectorisation, we obtain a 6NT-long vector of the associated acceleration and angular velocity projections. We then calculate the correlation coefficient between this vector and t. Sliding the 2D window vertically, we repeated this process for all n and obtain N − NT + 1 correlation coefficients. We store the results for each template move i in a vector denoted with This calculation is supported in detail with implementation equations in Appendix A.6. Indicating high similarity between and , local maximums of reveal the possible presence of move i in the test sequence. For each dance move i, we identify the local maximums of that exceed a certain threshold value, r. We define the indices of these local maximums as move i onsets in the sequence and the associated values as the recognition confidence. After the entire sequence has been searched through, we performed a final correction for false positives detection and elimination: if any two onsets were closer than 0.7 NT, we discarded the one with the lower recognition confidence.

2.6. Recognition Performance Assessment

All performances of dancers for all test sequences were visually inspected by the professional using the accompanying videos. All moves determined to be improperly performed were discarded from analysis. All proper performances were appropriately labelled as one of the 15 moves and considered for assessing the recognition ability of the presented method. The presented recognition method is non-binary and we assess its recognition ability by considering each move that is correctly recognised as a true positive (TP). Each move that is incorrectly recognised we consider as a false positive (FP). Using cumulative TP and FP rates for all I = 15 moves, we calculate the sensitivity and precision scores for each dance tempo considered. We calculate these scores for threshold values r ranging from 0.25 to 0.80 with 0.05 increment steps. We finally use the F1 score, calculated from sensitivity and precision, as the unified measure of recognition ability. In the specific dance motion recognition context, the sensitivity scores are mainly influenced by the similarity between the analysed moves and the corresponding templates in the database. As such, the sensitivity scores obtained for the professional dancer indicate how consistent her performance is, while those obtained for the recreational dancers show how well they imitate the professional. Precision scores are additionally influenced by the similarities between different moves, indicated by nondiagonal elements of for the initial template matching and of for the final template matching. Performed on acceleration and angular velocity 3D projections, final template matching is expected to have a higher discriminative potential than the initial template matching, performed using the acceleration and angular velocity magnitudes.

3. Results and Discussion

3.1. Database of Template Moves

Figure 2 shows the acquired acceleration and angular velocity signals for two illustrative template moves: (3) Kicks and (7) 20s Charleston. The first row depicts the acceleration signals, while the second row shows the angular velocities. Even from pure visual inspection, we can see that both the 3D projections and magnitudes have distinctive signal shapes. It is therefore reasonable to assume that these signal shapes can be used as fingerprints enabling dance move recognition.
Figure 2

Examples of acceleration (first row) and angular velocity (second row) of right leg motion during two solo jazz moves. The first three columns show the 3D projections and the last column illustrates the magnitudes.

The similarities between different template moves , calculated using the magnitudes of acceleration and angular velocity, are presented in Table 1. As expected, some dance moves show greater pairwise similarity while others are more idiosyncratic and show greater distinction from the rest. The highest similarity is obtained for the pair of moves (7) 20s Charleston and (8) Knee slaps, reflecting a leg activation pattern these two moves have in common. Both of these moves are executed mainly with the right leg during the first four music beats and with the left leg during the second four music beats.
Table 1

Solo jazz template moves’ pairwise similarities calculated using the magnitudes of acceleration and angular velocity. Depicted in red are the highest similarities obtained for each move, excluding self-similarity.

Move(1)(2)(3)(4)(5)(6)(7)(8)(9)(10)(11)(12)(13)(14)(15)
(1)1.000.220.450.290.390.290.370.270.350.350.320.460.270.310.31
(2) 1.000.190.210.350.470.290.440.340.280.330.330.270.300.37
(3) 1.000.420.390.280.300.330.350.260.360.550.330.430.30
(4) 1.000.370.290.310.360.390.240.430.550.320.270.43
(5) 1.000.330.370.370.290.260.290.480.230.480.38
(6) 1.000.380.420.330.320.440.380.240.360.44
(7) 1.000.610.200.370.310.320.250.330.28
(8) 1.000.270.340.470.390.340.300.42
(9) 1.000.220.290.240.300.430.31
(10) 1.000.290.350.310.270.30
(11) 1.000.470.540.260.35
(12) 1.000.390.310.43
(13) 1.000.190.32
(14) 1.000.43
(15) 1.00
The move with the lowest similarity with respect to all other moves, i.e., with the lowest maximum value, is Apple Jacks (10): for 1 ≤ j ≤ 15, j ≠ 10. The presented values indicate the discriminative potential of using the acceleration and angular velocity magnitudes for recognising different solo jazz dance moves. In particular, for threshold values r > 0.61, high precision scores are expected; however, the higher r is, the lower is the expected sensitivity, diminishing the model’s overall recognition ability. The similarities between different template moves calculated using the 3D projections, are presented in Table 2. We can see that these values are generally lower than the corresponding values of , and for all pairs of moves, is below 0.5.
Table 2

Solo jazz template moves’ pairwise similarities calculated using the 3D projections of the acceleration and angular velocity. Depicted in red are the highest similarities obtained for each move, excluding self-similarity.

Move(1)(2)(3)(4)(5)(6)(7)(8)(9)(10)(11)(12)(13)(14)(15)
(1)1.000.180.280.210.230.150.230.150.260.320.250.190.100.230.25
(2) 1.000.220.100.220.230.200.230.120.210.150.220.140.290.19
(3) 1.000.210.400.180.250.100.190.1860.180.320.130.380.18
(4) 1.000.260.190.130.120.260.170.430.340.110.160.40
(5) 1.000.230.350.140.170.200.260.330.100.340.25
(6) 1.000.330.220.170.180.180.160.130.230.20
(7) 1.000.210.110.230.250.150.120.380.26
(8) 1.000.230.150.110.170.090.240.10
(9) 1.000.190.190.300.200.150.26
(10) 1.000.180.210.290.170.22
(11) 1.000.280.230.140.39
(12) 1.000.180.270.29
(13) 1.000.120.18
(14) 1.000.19
(15) 1.00
The highest value is obtained for the pair of moves (4) Half break and (11) Boogie back: The lowest maximum value is 0.24 and is obtained for moves (8) Knee slaps and (14) Cross step. The presented values indicate the expected performance of the final template matching: for threshold values r > 0.43, high precision scores are expected. Lower values of obtained with respect to confirm the natural assumption that comparing dance moves using the 3D projections of signals instead of their magnitudes, considering the direction of motion instead of only the general leg activation pattern, has a higher discriminative potential. Therefore, template matching achieved using the 3D projections instead of magnitudes is expected to provide for higher precision scores, better overall recognition ability, and lower values of r.

3.2. Dance Move Recognition

For all 36 test sequences, the estimated dance tempo is accurate up to 1 bpm, allowing us to temporally scale all sequences and perform correlation-based template matching, as presented in the previous section.

3.2.1. Validation Using the Professional Dancer’s Test Sequences

For the professional dancer, for each dance tempo 120–200 bpm, 75 moves were acquired. For tempo 220 bpm, 70 moves were acquired. The sensitivity and precision scores are presented in Figure 3 and Figure 4.
Figure 3

Sensitivity scores obtained for the professional dancer’s test sequences.

Figure 4

Precision scores obtained for the professional dancer’s test sequences.

For the initial template matching, the highest sensitivity scores range from 0.80 to 0.91 and are obtained for r ≤ 0.4. For the final template matching, they range from 0.84 to 0.96 and are also obtained for r ≤ 0.4. For all dance tempos considered, the precision scores are equal to 1 for r ≥ 0.65 for the initial template matching and r ≥ 0.50 for the final template matching, indicating that for r this high, there are no FPs. These results are consistent with the template moves’ pair-wise similarities presented in Table 1 and Table 2. The improvement in precision for the final template matching is a direct consequence of using the 3D projections of the inertial signals instead of their magnitudes for template matching. As expected, considering the direction of motion provides for a better discrimination of moves. Figure 5 shows the F1 scores. For each dance tempo considered, the maximum F1 score indicates the best recognition ability. Lower F1 scores to the left of the maximum indicate higher FP rates and consequently lower precision. Lower F1 scores to the right of the maximum indicate lower TP rates and consequently, lower sensitivity.
Figure 5

F1 scores obtained for the professional dancer’s test sequences.

For the initial template matching, F1 maximums are obtained for r ≤ 0.50. The highest maximum is 0.91, obtained for tempos 160 and 180 bpm for r = 0.30 and 0.35. The lowest maximum is 0.82, obtained for the slowest tempo considered, i.e., 120 bpm, for 0.30 ≤ r ≤ 0.40. We can also observe that for r > 0.60, F1 scores drop sharply. Using the criteria of the highest minimum F1 score at a specific r, we can conclude that the overall best recognition performance is achieved for r = 0.30 and 0.35, for which the F1 scores range from 0.82 for 120 bpm to 0.91 for 180 bpm dance tempos. As expected, the F1 scores for the final template matching exceed those obtained for the initial template matching. The F1 maximums range from 0.87 to 0.98. The overall best recognition performance is obtained for r = 0.35 and 0.40. These results confirm the potential of using the presented method to recognise repeated dance moves and to assess dancers’ consistency in performance.

3.2.2. Validation Using the Recreational Dancer’s Test Sequences

For the recreational dancers, 293–311 dance moves were collected for each dance tempo considered. The sensitivity, precision, and F1 scores are presented in Figure 6, Figure 7 and Figure 8, respectively.
Figure 6

Sensitivity scores obtained for the recreational dancers’ test sequences.

Figure 7

Precision scores obtained for the recreational dancers’ test sequences.

Figure 8

F1 scores obtained for the recreational dancers’ test sequences.

For the initial template matching, the sensitivity maximums range from 0.74 to 0.82 and are obtained for r ≤ 0.40. For the final template matching, they range from 0.89 to 0.98 and are obtained for r ≤ 0.30. For all dance tempos considered, the precision scores are equal to 1 for r ≥ 0.70 for the initial template matching and r ≥ 0.60 for the final template matching. For the initial template matching, the F1 maximums range from 0.76 to 0.84 and are obtained for r ≤ 0.45. For the final template matching, the F1 maximums range from 0.89 to 0.98 and are obtained r ≤ 0.30. The main difference when compared to the professional dancer’s results is that the maximum F1 scores are obtained for lower values of r. In addition, after rising to their maximum value, F1 scores fall more sharply. These observations are an expected consequence of individual style and subtle differences in move performances of different dancers. The results confirm the potential of using the presented method to recognise imitated dance moves. Similar recognition ability can be achieved for the template professional dancer, but at lower values of r. Finally, the recognition ability of the final template matching proves that extracting three device-independent coordinate system axes from the acquired dance motion acceleration signals as presented in the previous section is valid and enables reliable transformation of the inertial signals from the device-intrinsic to the coordinate system of the templates.

4. Conclusions

We have shown that a single wearable device, capturing 3D acceleration and angular velocity of a dancer’s right leg motion, provides for recognising 15 solo jazz dance moves. We have demonstrated that the analysed sequences can be reliably temporally scaled and that dance recognition can be achieved using the same templates, independent of the tempo of dancing. Considering this makes the creation of the template database easier. Residing on a single device makes the solution highly practical. The practical aspect is further enhanced with the recognition’s independency of the device orientation. The results confirm that even for such a dynamic motion as solo jazz dancing, the acquired acceleration and angular velocity 3D projections can be reliably transformed to a device orientation-invariant system. Exploiting the 3D projections of the signals has a higher discriminative potential and provides for better recognition ability than does considering only the signals’ magnitudes. The presented solution fills the gap of the existing state of the art: since the presented solution avoids using video cameras or IR imaging sensors, it is not computationally expensive or limited to specific locations; residing on a single wearable device and not being dependent on the orientation of the wearer makes the methodology suitable for unsupervised everyday use and applicable for a variety of dancing situations—whether the dancer is dancing alone, in the crowd, or in front of an audience. Activity recognition methods relying on wearable devices, developed for ubiquitous measurement scenarios, usually try to balance the opposing requirements of device life autonomy and recognition performance. By relying on a limited set of features extracted from the inertial signals, these methods reduce the computational complexity and prolong the device life autonomy. Since the methodology presented in this article is primarily intended to be used for relatively short dancing sequences, e.g., per song basis, as opposed to throughout-the-day use, more computational power, needed for performing the correlation-based template matching, can be invested in the analysis. The advantages for the dancer and/or dancing instructor are straightforward: for each performed dance sequence, dance moves can be extracted together with recognition confidence levels, indicating how well the execution matches the reference template. This way, the methodology supports progress monitoring. In addition, the recognised moves can be further analysed and compared to the reference moves in the database to evaluate the overall performance. Finally, while tested on solo jazz dance moves, the methodology presented in this article can be extended to other dance styles, as well. The presented methodology demonstrates a high recognition ability for the preliminary set of measurements obtained for six dancers. However, a standalone study is needed to further explore in-depth the variability between dancers. Further analysis is also necessary to assess the methodology recognition performance in real dancing scenarios, when the dancer is dancing freely, combining various predefined and improvised moves in sequences. Further studies can also capitalise on this study and assess the dancer’s creativity and response to music, investigating the crucial connection between dancing and music. The benefits of using additional devices, positioned on other body segments, can be investigated. Finally, the presented implementation can be explored further with the aim of optimisation for real-time execution.
  12 in total

1.  Time- and computation-efficient calibration of MEMS 3D accelerometers and gyroscopes.

Authors:  Sara Stančin; Sašo Tomažič
Journal:  Sensors (Basel)       Date:  2014-08-13       Impact factor: 3.576

2.  Classification of K-Pop Dance Movements Based on Skeleton Information Obtained by a Kinect Sensor.

Authors:  Dohyung Kim; Dong-Hyeon Kim; Keun-Chang Kwak
Journal:  Sensors (Basel)       Date:  2017-06-01       Impact factor: 3.576

3.  LARa: Creating a Dataset for Human Activity Recognition in Logistics Using Semantic Attributes.

Authors:  Friedrich Niemann; Christopher Reining; Fernando Moya Rueda; Nilah Ravi Nair; Janine Anika Steffens; Gernot A Fink; Michael Ten Hompel
Journal:  Sensors (Basel)       Date:  2020-07-22       Impact factor: 3.576

4.  Development of an Automatic Alpine Skiing Turn Detection Algorithm Based on a Simple Sensor Setup.

Authors:  Aaron Martínez; Rüdiger Jahnel; Michael Buchecker; Cory Snyder; Richard Brunauer; Thomas Stöggl
Journal:  Sensors (Basel)       Date:  2019-02-21       Impact factor: 3.576

5.  Human Activity Recognition Using Inertial Sensors in a Smartphone: An Overview.

Authors:  Wesllen Sousa Lima; Eduardo Souto; Khalil El-Khatib; Roozbeh Jalali; Joao Gama
Journal:  Sensors (Basel)       Date:  2019-07-21       Impact factor: 3.576

6.  Dancing Salsa with Machines-Filling the Gap of Dancing Learning Solutions.

Authors:  Gianluca Romano; Jan Schneider; Hendrik Drachsler
Journal:  Sensors (Basel)       Date:  2019-08-23       Impact factor: 3.576

7.  Detection of Tennis Activities with Wearable Sensors.

Authors:  Luis Benages Pardo; David Buldain Perez; Carlos Orrite Uruñuela
Journal:  Sensors (Basel)       Date:  2019-11-16       Impact factor: 3.576

8.  Early improper motion detection in golf swings using wearable motion sensors: the first approach.

Authors:  Sara Stančin; Sašo Tomažič
Journal:  Sensors (Basel)       Date:  2013-06-10       Impact factor: 3.576

9.  Complex Human Activity Recognition Using Smartphone and Wrist-Worn Motion Sensors.

Authors:  Muhammad Shoaib; Stephan Bosch; Ozlem Durmaz Incel; Hans Scholten; Paul J M Havinga
Journal:  Sensors (Basel)       Date:  2016-03-24       Impact factor: 3.576

Review 10.  Inertial Sensor-Based Gait Recognition: A Review.

Authors:  Sebastijan Sprager; Matjaz B Juric
Journal:  Sensors (Basel)       Date:  2015-09-02       Impact factor: 3.576

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.