Literature DB >> 34104813

Reliability and concurrent validity of TRAZER compared to three-dimensional motion capture.

Jennifer A Hogg1, Lynette M Carlson1, Abigail Rogers2, Mason W Briles3, Shellie N Acocello1, Gary B Wilkerson1.   

Abstract

BACKGROUND: Efficient neural processing of visuospatial and proprioceptive input appears to be crucial for avoidance of sport injury. As such, clinically-feasible tests are needed to identify deficiencies found by advanced neuroimaging and electrophysiological tests. Three-dimensional motion capture in a laboratory setting is currently the gold standard for measurement of human movement parameters but is costly and requires extensive training. Non-immersive virtual reality systems with body motion tracking, such as TRAZER, may provide a clinically-feasible and portable means of acquiring similar variables. Test-retest reliability and concurrent validity of these systems are currently lacking. AIM: The aim of the study was to assess the concurrent validity of the TRAZER single-camera system with 3D motion capture system and to assess the test-retest reliability of TRAZER's whole-body reactive agility metrics.
METHODS: Participants - For validity, 13 healthy individuals (24.8±3.1 years, 170.0±7.7 cm, 70.0±14.2 kg); for reliability, 18 healthy individuals (23.3±2.5 years, 168.2±11.2 cm, 78.2±17.8 kg). Design - Validity was a single-session cross-sectional study. Reliability was a 3 consecutive day test-retest study. Setting-Controlled laboratory study. Intervention - Assessments utilized randomized movements in eight directions for forty total repetitions as designated by the TRAZER system. TRAZER protocol was simultaneously tracked by Vicon Motion Capture and the TRAZER system. Reliability data were captured on three consecutive days by the TRAZER system. Main Outcome Measures - Maximum acceleration, maximum velocity, and total distance were recorded for validation. In addition to these measures, maximum deceleration, average velocity, average acceleration, average deceleration, and average reaction time were collected for reliability.
RESULTS: Overall, a lack of agreement exists between maximum outputs for TRAZER and 3D motion capture (velocity r=0.808, acceleration r=-0.090), but total distance correlation was high (r =.961). ICC values between days 1-2-3 for average measures were high (average velocity=0.847, average acceleration=0.919, and average deceleration=0.948) with the exception of average reaction time being fair (ICC=0.536). ICCs for maximum measures showed a much smaller correlation between days (velocity=0.654, acceleration=0.171, and deceleration=0.416).
CONCLUSIONS: Even though there is a lack of strong concurrent validity between measures obtained from TRAZER and 3D motion capture systems, there is strong test-retest reliability of the TRAZER system. The applicability of these findings makes TRAZER clinically relevant in scenarios requiring pre- and post-testing for return to play decisions, or monitoring of a training regimen where demonstration of validation to a gold standard measurement is not relevant. RELEVANCE FOR PATIENTS: When test-retest capability is desired, such as in return-to-play protocols following an injury, Trazer is a reliable option. Copyright: © Whioce Publishing Pte. Ltd.

Entities:  

Keywords:  functional performance; kinect; reactive agility; virtual reality

Year:  2021        PMID: 34104813      PMCID: PMC8177031     

Source DB:  PubMed          Journal:  J Clin Transl Res        ISSN: 2382-6533


1. Introduction

Functional movement assessments are used by sports medicine clinicians to identify movement dysfunction for purposes of injury prevention, evaluation, rehabilitation, and return to activity decisions. Movement screenings, along with proper corrective interventions, are essential components of injury prevention [1]. Prevention strategies are critical to enhance physical performance, optimize health, minimize health-care expenses, and avoid chronic dysfunction and disability [2]. Portable and convenient acquisition of accurate movement assessment data is necessary for effective injury risk identification and ultimate prevention of sport-related injury. Lab-based three-dimensional (3D) motion capture systems are the gold standard in functional movement analysis, with reported excellent reliability (ICC3,k > 0.93) [3] and validity (±0.198 mm) (Vicon.com). However, they have limited clinical application due to financial, spatial, and temporal costs [4] and expertise needed to collect and interpret data. Other means of assessing quality of human movement have been developed and studied extensively, including the Functional Movement Screen [5], Star Excursion Balance Test [6], and the Landing Error Scoring System [7]. These tools, however, lack ecological validity, as their required movements are anticipated, whereas responding to unanticipated events and simultaneous performance of cognitive and motor tasks are typically required during athletic activities. Unanticipated events can lead to sensory prediction errors and improper muscle co-contractions, potentially resulting in musculoskeletal injury. Identifying deficits in the simultaneous processing of environmental stimuli and task constraints and in the ability to preplan correct motor sequences (feed-forward) is important for injury prevention [8-10]. Virtual reality systems that track body movements in response to visual stimuli may be valuable for assessing neuromechanical responsiveness, or the ability to optimally integrate neurocognitive and neuromuscular processes [11], and integrated perception-motor neural processes [12], which appear to be crucial for preventing athletic injuries [11]. Virtual reality is an emerging multidisciplinary tool [13] and is typically categorized as either immersive or non-immersive [14,15]. An immersive virtual reality system is one in which the user dons a head-mounted display to interact with their environment, while non-immersive virtual reality is interacted with through a television or computer monitor. TRAZER (Traq Global Ltd, Westlake, OH) is a commercially available non-immersive virtual reality system that utilizes a camera and software gesture recognition. It is marketed as a tool for prevention, rehabilitation, balance training, fall prevention, and concussion management. Its novelty lay in its output metrics. Namely, the measures of total distance, maximum velocity, maximum acceleration, and reaction time are indicative of overall functional performance and as such, may be useful for clinicians and coaches aiming to determine one’s ability to return to sport or their improvement in overall function [11,16]. Briefly, the TRAZER system is a non-immersive virtual reality system that utilizes a Kinect camera. The system employs an infrared camera to create a two-dimensional (2D) representation on a video monitor whereby a participant responds to visual stimuli and is recorded by a Kinect camera. All data captured by the TRAZER system are processed by embedded proprietary algorithms, making it unique to a stand-alone Kinect camera. The Kinect camera has been evaluated for reliability and validity compared to 3D motion capture systems during postural control and balance tasks [17-19], dynamic side-cut maneuvers [20], squatting [21], and single leg squatting [22]. Studies report good to excellent concurrent validity for kinematic angles during a postural balance task (Pearson’s r>0.90) [18], (ICC>0.75) [17], side-cutting maneuvers (absolute agreement ICC range=0.77–0.99) [20], and squatting (Pearson’s r>0.55) [21]. Test-retest reliability has variable results, from excellent reliability (ICC>0.90) reported by Schmitz et al., 2015 [21], to modest reliability (ICC≥0.70) reported by Clark et al., 2015 [19]. Although Kinect systems have been assessed for accuracy, the accuracy of proprietary algorithms overlaid on Kinect, such as those provided by TRAZER, has yet to be determined. One study compared TRAZER outputs with 3D Vicon motion capture system during balance and lateral shuffling tests [23]. Three separate and successive trials of four movement patterns were completed by one female and one male participant during simultaneous TRAZER and Vicon captures. The two systems yielded highly correlated results for joint coordinate positions (ICC range=0.75–0.99), joint marker distances (ICC range=0.93–1.0), marker velocities (ICC range=0.93–1.0), joint marker accelerations (ICC range=0.83–0.96), and joint marker decelerations (ICC range=0.91–0.97) [23]. However, due to the small sample size (n=2) and single day data collection, there is a need for further investigation. Therefore, the present study assessed the test-retest reliability of the TRAZER system and its concurrent validity against a 3D Vicon motion capture system. We hypothesized that TRAZER would display moderate test-retest reliability and good to excellent concurrent validity. This study represents a logical step towards the ability to make an informed decision about the fidelity of TRAZER’s measurement of functional movement and its use in the clinic and the laboratory. If TRAZER demonstrates agreement against the gold standard 3D motion capture, the utilization of a more clinically-feasible alternative may expand clinicians’ resources. The ability to accurately measure movement through a commercially available system has the potential for widespread adoption. Reliability is important in the clinic as well as in the laboratory, as reproducing consistent outcomes allow for accurate comparisons of patient progress.

2. Methods

2.1. Participants

The data acquired to assess reliability and concurrent validity were obtained from separate cohorts. The reliability cohort consisted of a convenience sample of 18 healthy individuals (11 female, 23.3±2.5 years, 168.2±11.2 cm, 78.2±17.8 kg). The validity cohort consisted of a convenience sample of 13 healthy individuals (eight females, 24.8±3.1 years, 170.0±7.7 cm, and 70.0±14.2 kg). Participants were included if they were between the ages of 18 and 35. Participants were excluded if they had sustained a lower extremity injury in the previous 6 months or if they had a history of vestibular disorders, balance disorders, or cardiac conditions or limitations. History of concussion was not considered exclusionary. The reliability and validity components of this study were approved by the university’s Institutional Review Board.

2.2. Instrumentation

TRAZER uses a depth-sensing Microsoft Kinect camera to create a three dimensional map of a 1.75 × 1.75 m capture area. Once within the field of view, a simulated person (avatar) appears on the monitor and mirrors the participant as they respond to visual targets randomly appearing on the perimeter of the capture area. Anatomical landmarks (e.g., joint centers) are determined with a randomized decision forest algorithm with a reported 1 ms latency [23-25]. Before the protocol, each participant stands in the center of the capture area facing the TRAZER television screen for a brief (~5 s) calibration, during which the Kinect camera recognizes and identifies the participant. Following calibration, a visual target randomly appears at one of eight possible locations on the perimeter of the capture area (forward, backward, left, right, forward left diagonal, forward right diagonal, backward left diagonal, or backward right diagonal) (Figure 1). Once the indicator appears, the participant moves as quickly as possible to the location. Once TRAZER detects the participant in the correct location, the indicator disappears, and the participant returns to the start position to prepare for the next repetition. The protocol consists of forty repetitions (five at each of the eight possible locations), each repetition entailing seven to eight feet of travel, and with a complete trial taking approximately 3 min to complete. TRAZER does not output raw data, but provides pre-defined performance metrics such as reaction time, average/maximum velocity, average/maximum acceleration, and deceleration, and total distance traveled (see Nyman, 2017a and https://trazer.com/science for comprehensive background information on TRAZER).
Figure 1

Depiction of TRAZER protocol. Top row is the screen as viewed by the participant. Bottom row is the participant reacting to the location of the virtual target.

2.3. Procedures

Following written informed consent, height and weight were obtained from each participant. To measure reliability, operationally defined as test-retest consistency, participants were not digitized with any 3D markers or instrumentation. They attended data collection sessions on three consecutive days at similar times each day, during which each participant performed the TRAZER protocol. Based on pilot testing, 3 consecutive days were used to account for a possible learning effect on the 1st day. For validity, operationally defined as absolute agreement, each participant attended a single session in which a reflective marker placed on the S2 spinous process was digitized with Vicon Nexus (Vicon, Oxford Metrics, London, England) software. A static capture was recorded. Participants then performed the TRAZER protocol while being tracked concurrently with the TRAZER system sampling at 30 Hz and an 8-camera Vicon 3D Motion Capture system sampling at 60 Hz, consistent with previous validity studies that have used higher sampling rates for 3D motion capture [4,18,23].

2.4. Data handling

TRAZER obtains its metrics by tracking the “base of the spine,” which was operationalized as the digitized S2 spinous process marker. To obtain measurements from three-dimensional data, raw coordinates for the S2 marker were exported from Vicon Nexus. Data were processed in R (R Core Team (2013). R: A language and environment for statistical computing) using the packages “signal,” [26] “imputeTS,” [27] “zoo,” [28] and “purr” [29]. Raw marker coordinate data were interpolated and filtered with a 12Hz low-pass 4th order Butterworth filter. The cutoff of 12 Hz was chosen after a residual analysis indicated 13 Hz to be the optimum cut point to maximize the signal to noise ratio. Total distance, maximum velocity, maximum acceleration, average velocity, average acceleration, average deceleration, and reaction time were the outcomes of interest. However, due to the proprietary nature of TRAZER calculations, only total distance, maximum velocity, and maximum acceleration were able to be confidently computed independently from TRAZER; thus, for validity, only these three variables were used. They were calculated as follows: Total distance (m) = ∑❘pi - pi-1❘, where p = position and i = capture frame Maximum velocity (m/s) = , where p = position, i = capture frame, and t = time Maximum acceleration (m/s2) = , where v = velocity, i = capture frame, and t = time

2.5. Statistics

To assess reliability, a repeated-measures MANOVA was used to determine differences between days. Interclass correlation coefficients (ICC(3, 3)) and standard errors of the measurement (SEM) were computed for total distance, maximum velocity, maximum acceleration, maximum deceleration, average velocity, average acceleration, average deceleration, and reaction time. To assess validity, a repeated-measures MANOVA was conducted to compare means between the TRAZER and Vicon systems for total distance, maximum velocity, and maximum acceleration. Paired t-tests were conducted as post-hoc follow-ups. Pearson correlations, ICC(2, 2), and SEM were also computed to determine levels of absolute agreement. ICCs were computed according to McGraw and Wong [30] and interpreted as follows: Less than 0.5, between 0.5 and 0.75, between 0.75 and 0.9, and greater than 0.9 were considered poor, moderate, good, and excellent, respectively [31]. Bland-Altman plots were constructed with the R package “BlandAltmanLeh” [32] to visually inspect validity and reliability data. Alpha was set at 0.05 for all statistical tests. Analyses were conducted in SPSS (IBM Corp, Armonk, NY).

3. Results

3.1. Reliability

A significant deviation from normality was evident for day one’s reaction time (Shapiro–Wilk P=0.002), but neither logarithmic nor square root transformation provided any substantial improvement in distribution normality. A repeated measures MANOVA including all eight dependent variables demonstrated a significant difference among trials (Wilks’ Lambda = 0.36; F15,54 = 2.23; P=0.015). With the exception of the three variables of average reaction time, average deceleration, and maximum acceleration, ICCs(3, 3) across all three testing sessions were moderate to excellent, ranging from 0.65 to 0.95. Test-retest consistency among trials, Shapiro–Wilk test of normality result for each trial, and univariable repeated measure analysis of variance results for differences among trials are presented in Tables 1 and 2.
Table 1

Test-retest consistency among trials (ICC3,3), mean, standard deviation, and Shapiro-Wilk test of normality result for each trial (PS-W), unvariable repeated measure analysis of variance result for difference among trials (PTrials), and standard error of measurement (SEM).

Variable3-Day ICC3,3 (SEM)Days 1 & 2 ICC3,3 (SEM)Days 2 & 3 ICC3,3 (SEM)Trial 1Trial 2Trial 3



Mean±SDPS-WMean±SDPS-WMean±SDPS-WPTrials
Total Distance (m)0.745 (3.36)0.794 (3.02)0.761 (3.25)82.01±7.260.15181.77±7.100.79579.39±7.530.7140.076
Average Velocity (m/s)0.847 (0.03)0.819 (0.03)0.864 (0.03)0.68±0.110.5750.68±0.090.6210.68±0.080.7710.937
Average Acceleration (m/s2)0.919 (0.13)0.852 (0.03)0.957 (0.10)2.11±0.470.7402.14±0.510.3382.13±0.530.2080.915
Average Deceleration (m/s2)0.948 (0.07)0.927 (0.08)0.907 (0.09)1.73±0.330.8061.78±0.280.7351.86±0.350.3670.008
Maximum Velocity (m/s)0.654 (0.06)0.665 (0.06)0.559 (0.07)0.82±0.120.2820.85±0.110.1730.84±0.110.2530.444
Maximum Acceleration (m/s2)0.171 (0.51)0.067 (0.54)0.032 (0.55)2.83±0.610.9153.33±1.120.0012.82±0.530.2630.121*
Maximum Deceleration (m/s2)0.212 (0.13)0.255 (0.14)-0.446 (0.14)2.30±0.44…4642.61±0.77<.0012.36±0.420.7420.222*
Reaction Time (ms)0.536 (31)0.447 (40)0.493 (40)320±1010.002326±410.291316±260.9460.773*

Mauchly’s test of sphericity identified a significant difference among trials (P <.05); Greenhouse-Geisser df adjustment

Table 2

Mean, standard deviation, Shapiro-Wilk test of normality result (PS-W), Pearson correlation, and associated P value for measurements derived from different motion analysis systems.

VariableTRAZERViconPearson rPSpearman’s ρP


Mean±SDPS-WMean±SDPS-W
Total Distance (m)89.46±17.850.108108.59±24.470.0150.961<.0010.956<.001
Maximum Velocity (m/s)1.00±0.120.3122.46±0.520.0040.808<.0010.6630.014
Maximum Acceleration (m/s2)4.32±0.470.9031.34±0.420.778-0.0900.7700.0110.972
Mauchly’s test of sphericity identified a significant difference among trials (P <.05); Greenhouse-Geisser df adjustment

3.2. Validity

Vicon distributions for total distance and maximum velocity deviated significantly from normality, but neither logarithmic nor square root transformation provided any substantial improvements. Thus, the results of non-parametric correlations (Spearman’s rho) were also calculated and are presented in Table 2. A repeated measures MANOVA including all three dependent variables demonstrated a significant difference between systems (Wilks’ Lambda = 0.03; F3,10 = 94.00; P<0.001). Means, standard deviations, and Pearson r values for total distance, maximum velocity, and maximum acceleration obtained from TRAZER and Vicon are presented in Table 2. ICCs (SEM) for total distance, maximum velocity, and maximum acceleration were 0.79 (11.1m), 0.08 (0.5 m/s), and 0.01 (0.5 m/s2), respectively. Bland Altman plots are presented as visual representations of the absolute agreement between TRAZER and Vicon (Figures 2-4).
Figure 2

Bland-Altman plot depicting limits of agreement between TRAZER and three-dimensional motion capture for total distance. The central dotted line represents systematic error. The outer dotted lines represent ±2 standard deviations.

Figure 4

Bland-Altman plot depicting limits of agreement between TRAZER and three-dimensional motion capture for maximum acceleration. The central dotted line represents systematic error. The outer dotted lines represent ±2 standard deviations.

4. Discussion

We hypothesized that TRAZER would display moderate test-retest reliability and good to excellent concurrent validity. Our hypothesis was partially supported as evidenced by good to excellent reliability, but poor concurrent validity. To the best of our knowledge, such a comparison has not been previously performed. Although we observed a lack of concurrent validity between TRAZER and 3D motion capture, there was strong test-retest reliability of the TRAZER system particularly with regards to average acceleration, average velocity, and average deceleration. Test-retest reliability of TRAZER was particularly strong with averaged measures such as average acceleration (ICC=0.919), average velocity (ICC=0.847), and average deceleration (ICC=0.948) in contrast to reliability of peak measures of acceleration, velocity, and deceleration (ICC range = 0.171–0.654). This is reasonable, as averaging a variable over the length of the TRAZER protocol is inherently more stable than extracting the peak from its path. These results align with Lopes’s meta-analysis [33], reporting high ICCs for 2D intra-rater reliability. Lopes’s meta-analysis reported an ICC of 0.99 for intra-rater reliability in studies that observed frontal plane projection angles with 2D video systems. Although this meta-analysis assessed joint angle reliability, while the current study inspected X-Y coordinate data, this lends further confidence to the reliability of 2D video systems such as TRAZER. As such, TRAZER may be appropriate for use in a setting in which serial measures (test-retest) are obtained to monitor progress, such as a physical therapy clinic or strength and conditioning facility. For instance, periodic measurements could be obtained following a lower extremity injury to quantify the extent to which a patient achieves superior functional performance. Furthermore, baseline TRAZER testing could be appropriate as a post-injury comparison to determine the point at which a patient reaches pre-injury level of function. Observing high absolute agreement would have provided evidence to support the use of the more clinically-feasible TRAZER system as a clinical injury screening tool in comparison to the more burdensome gold standard 3D motion capture. The lack of strong concurrent validity between TRAZER and 3D motion capture is in stark contrast to a previous study (white paper) conducted by Nyman (Nyman, 2017). The findings of Nyman’s comprehensive evaluation testing TRAZER’s validity against 3D motion capture demonstrated high correlation for aggregate joint marker accelerations, decelerations, velocities, and distances (ICC range = 0.83–0.96). While both studies possessed similarities, the number of participants differed, Nyman had only two participants, whereas 18 participants were tested in the current study. For validation, we analyzed peak velocity and acceleration while Nyman reported average velocity and average acceleration. Average measures will be more stable than isolated peak measures and may explain the differences between the two sets of findings. Without access to TRAZER’s proprietary algorithms for calculation of these metrics, we were unable to assess TRAZER’s metric calculations and instead computed the metrics using data obtained from 3D motion capture. It should be noted that Nyman’s work was a white paper, for which raw coordinate data were available. As such, TRAZER time series trajectories of the S2 marker were overlaid on Vicon-obtained trajectories and then submitted to a validity analysis, whereas we assessed the validity of TRAZER-reported discrete metrics with metrics calculated with Vicon coordinate data. Although Nyman reported excellent validity ICCs for time series data, coupling this with the current data suggests that gross metrics (e.g., trajectories, averages) are more valid than discrete (e.g., peaks, maximums, and minimums) measures. In addition, the systematic error as depicted in the Bland Altman plots for TRAZER and 3D motion capture validity showed the lack of between-system agreement. Over the course of the 40 repetition protocol, TRAZER systematically captured 19.13 m less in total distance traveled than 3D motion capture, equating to an 18% discrepancy. A possible explanation for lack of agreement could be the difference in capture rate. It should be noted that the Vicon system sampled data at 60 Hz. This, coupled with an array of eight cameras, naturally allows more accuracy in detecting peaks and maximums, compared with the TRAZER system operating with a single camera and a capture rate of 30 Hz. While TRAZER primarily picks up gross movements, Vicon is designed to record more finely tuned peaks that TRAZER may gloss over, thus producing unagreeable measures. This may explain the systematic differences between the two systems. The authors elected for a 60 Hz Vicon sampling rate because this is more representative of a gold-standard motion capture collection. This is consistent with other researchers who have used different sampling frequencies when comparing two-dimensional with three-dimensional motion capture [4,18,23]. In fact, in the white paper reported by Nyman, three-dimensional motion capture was sampled at 120 Hz, while TRAZER was sampled at 30 Hz. Thus, although different sampling rates likely partially explain the systematic differences observed between the TRAZER and Vicon in the current study, it does not fully account for the discrepancy.

4.1. Clinical and translational impact

A key advantage of the TRAZER system is the ability to closely replicate sport demands by presenting visual-cognitive virtual reality challenges that elicit quantifiable whole-body movement responses. Confidence in the system’s measurements of reactive responses for clinical documentation of pre- and post-injury performance capabilities and assessment of injury risk is supported by some of the validity and reliability coefficients derived from our testing. Exceptionally close agreement of the total distance measurement derived from TRAZER with that from the Vicon system, along with very good test-retest reliability; clearly support its use as an indicator of whole-body movement precision in the deactivation of virtual reality targets. These data indicate that TRAZER is appropriate for use as a baseline measure, in addition to post-injury quantification of return to pre-injury levels of functional capacity; thus, allowing for more individualized return-to-play protocols. An important limitation of this study was the different sampling rates of each system; TRAZER captured at 30 Hz, while Vicon captured at 60 Hz. We acknowledge that this difference may partially account for discrepancies in accuracy. However, as gold standard three-dimensional motion capture typically samples at higher frequencies, this is a more externally valid comparison. Although participants were instructed to conduct similar activities prior to each day’s testing and were tested at similar times each day, we cannot discount the fact that outside factors may have influenced the participants’ energy levels and reactions each day. History of concussion was not considered exclusionary for participation. While this does not affect test-retest reliability or concurrent validity, the reader should caution against using these data as reflective of the larger population. Finally, validation efforts were significantly hampered by the inability to access proprietary algorithms utilized by TRAZER to calculate the mean data for validation. Relying on maximum outputs likely accounted for poor correlation across platforms with different capture rates and camera angles. As evidenced by a higher inter-rater reliability coefficient and r value for total distance, when measures are taken over time, instead of at maximal points in time, the systems display higher congruency. TRAZER is designed to be clinically feasible, transportable, and user-friendly compared to the expensive, training-intensive, and sedentary 3D motion capture system. Although the poor validity of TRAZER against a gold standard 3D motion capture hinders that ability to directly compare values obtained from each system, lack of validation does not inhibit TRAZER’s ability to provide reliable test-retest measures. Findings for the test-retest reliability of the TRAZER system may prove beneficial in future studies to determine best uses of TRAZER system in a clinical setting. Additional studies focusing on examining the validity of mean values with a larger sample size could provide further evidence of research and clinical usefulness.
  23 in total

1.  Validity of the Microsoft Kinect for assessment of postural control.

Authors:  Ross A Clark; Yong-Hao Pua; Karine Fortin; Callan Ritchie; Kate E Webster; Linda Denehy; Adam L Bryant
Journal:  Gait Posture       Date:  2012-05-23       Impact factor: 2.840

2.  Reliability of landing 3D motion analysis: implications for longitudinal analyses.

Authors:  Kevin R Ford; Gregory D Myer; Timothy E Hewett
Journal:  Med Sci Sports Exerc       Date:  2007-11       Impact factor: 5.411

3.  Reliability and concurrent validity of the Microsoft Xbox One Kinect for assessment of standing balance and postural control.

Authors:  Ross A Clark; Yong-Hao Pua; Cristino C Oliveira; Kelly J Bower; Shamala Thilarajah; Rebekah McGaw; Ksaniel Hasanki; Benjamin F Mentiplay
Journal:  Gait Posture       Date:  2015-04-20       Impact factor: 2.840

4.  Reliability and Validity of Frontal Plane Kinematics of the Trunk and Lower Extremity Measured With 2-Dimensional Cameras During Athletic Tasks: A Systematic Review With Meta-analysis.

Authors:  Thiago Jambo Alves Lopes; Deisi Ferrari; Joshua Ioannidis; Milena Simic; Fábio Mícolis de Azevedo; Evangelos Pappas
Journal:  J Orthop Sports Phys Ther       Date:  2018-06-12       Impact factor: 4.751

5.  Functional movement screening: the use of fundamental movements as an assessment of function - part 1.

Authors:  Gray Cook; Lee Burton; Barbara J Hoogenboom; Michael Voight
Journal:  Int J Sports Phys Ther       Date:  2014-05

6.  Determining sensitivity/specificity of virtual reality-based neuropsychological tool for detecting residual abnormalities following sport-related concussion.

Authors:  Elizabeth Teel; Michael Gay; Brian Johnson; Semyon Slobounov
Journal:  Neuropsychology       Date:  2016-01-25       Impact factor: 3.295

7.  A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research.

Authors:  Terry K Koo; Mae Y Li
Journal:  J Chiropr Med       Date:  2016-03-31

8.  Detection of Persisting Concussion Effects on Neuromechanical Responsiveness.

Authors:  Gary B Wilkerson; Dustin C Nabhan; Chad J Prusmack; William J Moreau
Journal:  Med Sci Sports Exerc       Date:  2018-09       Impact factor: 5.411

9.  Concussion History and Neuromechanical Responsiveness Asymmetry.

Authors:  Gary B Wilkerson; Dustin C Nabhan; Ryan T Crane
Journal:  J Athl Train       Date:  2020-06-23       Impact factor: 2.860

10.  Reliability and validity of a novel Kinect-based software program for measuring a single leg squat.

Authors:  John Ressman; Eva Rasmussen-Barr; Wilhelmus Johannes Andreas Grooten
Journal:  BMC Sports Sci Med Rehabil       Date:  2020-05-11
View more
  1 in total

1.  Effects of caffeine, methylliberine, and theacrine on vigilance, marksmanship, and hemodynamic responses in tactical personnel: a double-blind, randomized, placebo-controlled trial.

Authors:  Harry P Cintineo; Marissa L Bello; Alexa J Chandler; Thomas D Cardaci; Bridget A McFadden; Shawn M Arent
Journal:  J Int Soc Sports Nutr       Date:  2022-08-18       Impact factor: 4.948

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.