Next Article in Journal
Mean-Line Design of a Supercritical CO2 Micro Axial Turbine
Next Article in Special Issue
Fixation Stability and Stiffness of Two Implant Systems for Proximal Femoral Varization Osteotomy
Previous Article in Journal
Real-Time Visual Tracking of Moving Targets Using a Low-Cost Unmanned Aerial Vehicle with a 3-Axis Stabilized Gimbal System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intra- and Interobserver Reliability Comparison of Clinical Gait Analysis Data between Two Gait Laboratories

1
Department of Orthopaedic and Trauma Surgery, Martin-Luther-University Halle-Wittenberg, Ernst-Grube-Str. 40, 06120 Halle, Germany
2
Private Practice for Neurology Dr. Wegener, Lienaustr. 2, 23730 Neustadt, Holstein, Germany
3
Laboratory for Biomechanics and Biomaterials, Hannover Medical School, Anna-von-Borries-Str. 1-7, 30625 Hannover, Germany
4
Department of Health Sciences, University of Colorado, Colorado Springs, CO 80918, USA
5
Department of Orthopaedic Surgery, Friedrich-Alexander-University Erlangen-Nürnberg, Rathsberger Str. 57, 91054 Erlangen, Germany
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(15), 5068; https://doi.org/10.3390/app10155068
Submission received: 7 July 2020 / Revised: 21 July 2020 / Accepted: 22 July 2020 / Published: 23 July 2020

Abstract

:

Featured Application

Clinical gait analysis (CGA) is an in vivo method used to measure the movement behavior/gait patterns of patients before and after an orthopedic treatment. It monitors joint kinematics and kinetics under dynamic conditions within the musculoskeletal system. This study contributes to a better understanding of the comparability and significance of motion analysis data recorded in different gait laboratories using different technical qualities.

Abstract

Comparing clinical gait analysis (CGA) data between clinical centers is critical in the treatment and rehabilitation progress. However, CGA protocols and system configurations, as well as choice of marker sets and individual variability during marker attachment, may affect the comparability of data. The aim of this study was to evaluate reliability of CGA data collected between two gait analysis laboratories. Three healthy subjects underwent a standardized CGA protocol at two separate centers. Kinematic data were captured using the same motion capturing systems (two systems, same manufacturer, but different analysis software and camera configurations). The CGA data were analyzed by the same two observers for both centers. Interobserver reliability was calculated using single measure intraclass correlation coefficients (ICC). Intraobserver as well as between-laboratory intraobserver reliability were assessed using an average measure ICC. Interobserver reliability for all joints (ICCtotal = 0.79) was found to be significantly lower (p < 0.001) than intraobserver reliability (ICCtotal = 0.93), but significantly higher (p < 0.001) than between-laboratory intraobserver reliability (ICCtotal = 0.55). Data comparison between both centers revealed significant differences for 39% of investigated parameters. Different hardware and software configurations impact CGA data and influence between-laboratory comparisons. Furthermore, lower intra- and interobserver reliability were found for ankle kinematics in comparison to the hip and knee, particularly for interobserver reliability.

1. Introduction

Three-dimensional clinical gait analysis (CGA) is an important diagnostic tool within the field of movement disorders [1]. Optical marker tracking systems are considered the gold standard for this type of assessment [2,3]. However, discrepancies in CGA data across multiple analyses can be caused by differences in marker sets [4], observer error [5], marker placement [6] and patient variability [7]. Despite applied standardized protocols (e.g., marker-set, measurement pipeline), incompatibilities and discrepancies within CGA data of the same patient cohorts collected by different institutions are often observed, and have caused reservations concerning the applicability of CGA among multicenter studies [3,4,8]. Ferrari et al. [4] compared the trunk, pelvis and lower limb kinematics of five separate measurement protocols (different marker sets), using a single data pool of subjects. These authors reported good correlations between kinematic variables in the sagittal plane (flexion/extension), but poorer correlations for out-of-sagittal-plane rotations, such as knee abduction/adduction. They also reported good correlations among protocols with similar biomechanical models [4]. Ferrari et al. [4] concluded that the comparison of different measurement protocols results in higher data variability when compared to interobserver and interlaboratory comparisons for most gait characteristics. Obviously, model conventions and definitions are crucial for data comparison. Gorton et al. [8] identified the marker placement procedure of examiners as the largest source of error. The use of a standardized protocol for marker placement decreased data variability by up to 20%. Based on clinical experience, the authors of the current study hypothesized that significant discrepancies would exist between CGA data sets collected using different hardware and software infrastructures or configurations.
Due to the observed limitations in the reproducibility of CGA data, efforts have focused on the reduction of measurement error, with the aim of producing data with the highest possible degree of reproducibility. To improve our understanding of the accuracy and reproducibility of CGA data, cross-laboratory studies must be performed using a single cohort of subjects. Previous research has reported on the repeatability of motion capture data using separate trials, sessions and observers under various conditions for a specific CGA setup or laboratory [5,6,7,9,10]. However, reports for multi-center repeatability are limited and have focused on differences in hardware configurations, marker placement, as well as between trials and days of measurement [3,8,11]. Unfortunately, a validation using the same motion capturing technology, but different hardware and software infrastructure/configurations (e.g., capturing software, camera type), among a single cohort of subjects is lacking.
It is critical for clinicians and researchers to have reliable examination tools to accurately and objectively assess the functional status of a joint [12,13,14,15,16]. A high level of intraobserver reliability is imperative to accurately evaluate the longitudinal effects of the rehabilitation process and to identify differences between subjects [17,18]. A central research question is whether the findings of clinical gait analyses conducted by multiple laboratories are consistent and reliable enough for making clinical decisions—or is the dependence on observers, repeat measurements or measurement and analysis protocols too large?
The specific aims of this study were to evaluate the intra- and interobserver reliability and the equivalence of CGA data between two gait analysis laboratories (between-laboratory intraobserver reliability) using the same motion capturing technology, but different hardware and software infrastructure/configurations among a single cohort of subjects. The following two hypotheses were specifically investigated:
(1)
Intraobserver reliability of CGA data for the lower extremity obtained for the same cohort of subjects and captured at one laboratory will be good-to-excellent, whereas interobserver reliability will be fair-to-good.
(2)
CGA data for the lower extremity obtained for the same cohort of subjects, captured at two separate laboratories by the same observers, using different hardware and software systems, will be equivalent.

2. Materials and Methods

2.1. Subjects

A standardized measurement protocol was performed repeatedly on three adult subjects (one female, two males) who were asymptomatic at the time of testing and did not display any movement disorders. This cohort had a mean (±SD) age of 33.3 (±4.0) years, mean body weight of 77.3 (±16.3) kg, mean body height of 176.7 (±9.1) cm, and a mean body mass index of 24.3 (±2.9) kg∙m-2 (Table A1). The study was approved by the ethical committee of Martin-Luther-University Halle-Wittenberg (approval number: 217/08.03.10/10). Written consent from study participants was obtained prior to data collection.

2.2. Measurement Set-Up

Gait analyses were performed in two gait laboratories (Figure 1). Both laboratories used an optical infrared camera-based motion capturing system provided by one manufacturer (Vicon Motion System Ltd., Oxford, UK).
The first laboratory (GL1) was equipped with a six-camera system (460 Vcam cameras, Workstation 4.6 build 142 software). The second laboratory (GL2) used a six-camera system (MXF-20 cameras, Nexus 1.3 software). These systems thus differed in their respective configurations, both in terms of analysis software (Workstation vs. Nexus) and camera type (Vcam vs. MXF). Both laboratories applied the same sampling rate (200 Hz) and the same reflective markers (14 mm diameter). Capture space represented the full dimensions (length, width and height) of the motion capture room for GL1 (7 m, 4 m and 3 m) and for GL2 (15.5 m, 8.8 m and 4.8 m). Capture volume was defined as the area within the capture space, for which motion capturing cameras were able to capture the motion task of each subject. The GL1 capture volume was 6 m, 2 m and 2 m. The capture volume for GL2 was 10.0 m, 8.5 m and 2.9 m. Captured marker data were processed and trajectories were labeled using the PlugInGait model under a standardized protocol at both laboratories. All kinematic data were Woltring filtered using a mean squared error setting of 10. These gait data were reduced to 100% of one gait cycle using gait cycle event detection, based on the available force plates (threshold: 20 N). All subsequent measurement conditions were consistent at both laboratories.

2.3. Measurement Protocol

Calibration of the optical infrared camera-based motion capturing systems at both laboratories were performed according to the manufacturer’s guidelines. This calibration process consisted of two main steps: (i) a static calibration to calculate the origin of the capture volume and define the 3D workspace orientation (x, y, z directions) and (ii) a dynamic calibration to calculate the relative position and orientation of each camera within the capture volume. Calibration quality was checked according to the manufacturer’s guidelines.
The PlugInGait marker setup for the lower extremity (kinematic model V 2.3) was used in both laboratories, based on the work of Kadaba et al. [19]. Markers were attached to each subject by the same experienced staff member (Observer 1) at both gait analysis laboratories, according to a standardized protocol for anthropometric measurements, landmark identification and marker mounting.
Each of the three subjects performed 20 barefoot gait trials with a self-selected walking speed resulting in a total of 60 CGA trials per gait laboratory. Individual gait speed was controlled and standardized between data collection sessions. The same two staff members (Observer 1 and Observer 2) performed data processing and analysis at both laboratories, using a standardized protocol for data processing, labeling and gait event detection. For each reliability analysis variation, the observer started with the original raw data in their data processing routine. These raw data were labeled, gait events were detected and possible gaps of reconstructed marker trajectories were filled. The kinematic parameters were extracted using the same template and again used a standardized workflow for both observers. Kinematic data used for reliability analysis, which consisted of specific movement parameters in the sagittal and frontal plane for the hip, knee and ankle. Parameters were selected according to the specifications of Benedetti et al. [20] (Figure 2).
Data from the right leg of each subject were used for the analyses of reliability. Three different measures of reliability were performed (Figure 1):
  • Interobserver Reliability: The reliability of the two observers, using the same data set, was assessed using separate analysis software (Observer 1: GL1 (Workstation); Observer 2: GL2 (Nexus)).
  • Intraobserver Reliability: The reliability of the same data set among the same observer was tested. This assessment was performed with a time interval of 12 months between analyses using the same software (Workstation). The CGA data for this assessment were collected at the GL2 site.
  • Between-Laboratory Intraobserver Reliability: To compare the effect of laboratory environment and instrumentation while excluding observer-dependent influences, CGA data collected at both laboratories were analyzed by a single observer using the same analysis software.
The following reliability variables were assessed. ICCmean was the average of all parameters for hip, knee and ankle. ICCtotal (Equation 1) was the average of hip ICCmean, knee ICCmean and ankle ICCmean for all three types of reliability (intraobserver, interobserver, between-laboratory intraobserver).
ICCtotal = hip ICCmean + knee ICCmean + ankle ICCmean / 3

2.4. Statistics

Descriptive statistics (mean, standard deviation) were based on 20 barefoot trials and were calculated for 31 kinematic parameters in the sagittal and frontal plane for the hip, knee and ankle joints. Reliability analyses were divided into five parts:
  • For interobserver reliability, a single measure intraclass correlation coefficient (ICC) was calculated. The number of measures (k) was 60 (n = 20 barefoot trials from three subjects).
  • For intraobserver reliability, an average measure ICC was calculated. The number of measures (k) was 60.
  • To assess between-laboratory intraobserver reliability, an average measure ICC was calculated and again referenced to the same ICC value classification [21]. The ICC indicated excellent reliability if the value was above 0.75, fair-to-good reliability between 0.40 and 0.75 and poor reliability when less than 0.40. A two-way mixed-effects model (definition: absolute agreement) was used for all calculations. For all ICC values, 95% confidence intervals were reported.
  • To estimate experimental errors of a joint angle, the standard error was calculated (for intra- (σrepeated), inter- (σobserver) and between-laboratory intraobserver reliability (σsess(lab))) as described by Schwartz et al. [10], as well as the magnitude of the interobserver error and its ratio to intra-subject error r (Equations 2–4).
    r = σrepeatedtrial
    r = σobservertrial
    r = σsesss(lab)trial
  • A scatter-plot technique suggested by Bland and Altman [22] was used to assess interchangeability (equivalence) of CGA data between laboratories. Calculated differences for joint angles were plotted against their average value for each subject. The interchangeability of CGA data was tested by a bounding criterion defined as the mean ± two standard deviations of the measured differences (approximately 95% of all measured values).
  • To evaluate the variability within and between subjects, the standard error of measurement (SEM) was calculated in conjunction with the ICCs, using the following equation from Portney and Watkins [23]:
    SEM = SD × 1 ICC
    ICC values may be influenced by inter-subject variability of scores, because a large ICC may be reported despite poor trial-to-trial consistency if the inter-subject variability is too high [23,24]. However, the SEM is not affected by inter-subject variability [24].
  • Mean differences for multicenter comparisons were tested using variance analysis. A one-factor, univariate general linear model (GLM; dependent variable: hip flexion during stance; independent variable: CGA center; covariate: walking speed) was performed. Prior to inference statistical analyses, all variables were tested for normal distribution (Kolmogorov–Smirnov test). To estimate practical relevance and to quantify the differences between GL1 and GL2, effect sizes (partial eta squared, ηp2 [25,26]) were calculated for the main effects (ηp2) and the mean differences divided by the pooled standard deviations (d). To evaluate effect sizes, d or ηp2 were classified, with d ≥ 0.2, d ≥ 0.5, d ≥ 0.8 or ηp2 ≥ 0.01, ηp2 ≥ 0.06, ηp2 ≥ 0.14 indicating small, medium or large effects, respectively [27]. Level of significance was adjusted to p ≤ 0.002 (0.05/31 = 0.002) for multiple tests by means of a Bonferroni correction.
All statistical analyses were performed using SPSS version 25.0 for Windows (SPSS Inc., IBM, Armonk, NY, USA).

3. Results

Because of the controlled and standardized CGA data collection, walking speed and stride length parameters did not differ between laboratory measurement sessions (walking speed: 1.40 ± 0.06 m∙s-1 vs. 1.40 ± 0.04 m∙s-1; stride length: 1.38 ± 0.09 m vs. 1.40 ± 0.10 m).

3.1. Interobserver Reliability

All interobserver reliability variables fulfilled the assumption of normality. Interobserver reliability at one laboratory (GL2), showed excellent reliability for 71% of the parameters observed (22/31). Based on the lower limit of the 95% CI, 58% (18/31) displayed an ICC value of at least 0.75. Nineteen percent (6/31) of the parameters showed fair-to-good reliability, whereas 10% (3/31) were poor (Table 1).
The highest ICC value was observed for knee extension in the stance phase (ICC = 0.97, excellent reliability). In contrast, dorsiflexion during the stance phase showed poor reliability (ICC = 0.22) (Table 1). After averaging all parameters for interobserver reliability, an ICCtotal = 0.79 (95% CI: 0.67–0.86) was observed (Table 2).
For interobserver reliability, the largest standard error was observed for hip adduction (% gait cycle) during stance phase, as well as for ankle dorsiflexion (% gait cycle) during swing phase, with each presenting with a standard error of σobserver = 4.4% (Table 3). Plantarflexion during the stance phase had the worst detected angle (σobserver = 3.8°). The largest ratio of interobserver to intrasubject error was observed for knee abduction angle in the stance phase (σobserver = 2.4°, r = 3.2; Table 3).

3.2. Intraobserver Reliability

All intraobserver reliability variables fulfilled the assumption of normality. Intraobserver reliability (ICCtotal = 0.93) was significantly (p < 0.001, ηp2 = 0.399) higher than interobserver reliability (ICCtotal = 0.79). When considering the total intraobserver ICC, excellent values were observed for 94% (29/31) of parameters, 3% (1/31) were fair to good and 3% (1/31) were poor (Table 1). Based on the lower limit of the 95% CI, 87% (27/31) of the parameters showed an ICC value of at least 0.75 (excellent reliability). For intraobserver reliability, the calculated standard errors (σrepeated) and SEM were smaller than those for interobserver reliability (σrepeated) (Table 3, Tabel A2); however, these values were not compared statistically. The highest error for intraobserver reliability existed for ankle dorsiflexion during the swing phase (% gait cycle) (σrepeated = 4.2%), as well as plantarflexion during the swing phase (mean σobserver = 2.5°).
Intraobserver reliability across laboratories (between-laboratory intraobserver) was excellent in 29% (9/31) of all parameters, fair to good for 48% (15/31), and poor for 26% (8/31) (Table 1). The highest ICC (0.90) was observed for hip flexion during the stance phase (% gait cycle). The worst ICC (0) was calculated for knee abduction angle in the swing phase.

3.3. Between-Laboratory Intraobserver Reliability

All between-laboratory intraobserver reliability variables fulfilled the assumption of normality. Between-laboratory intraobserver reliability for the entire lower limb (hip, knee, ankle) was poor (ICCtotal = 0.56, Table 2). Excellent ICC values were only calculated for 29% (9/31) of parameters (Table 1). Considering the lower limit of the 95% CI, no parameter reached an ICC level of 0.75. Calculated standard errors between-laboratory intraobserver reliability (σsess(lab)) were larger than those for interobserver and intraobserver reliability (Table 3). The largest standard errors were the same as those found for interobserver reliability, which included hip adduction (% gait cycle) at the stance phase with a mean σsess(lab) = 5.4%, and for ankle dorsiflexion (% gait cycle) at the stance phase with a mean σsess(lab) = 4.8%. The largest ratio was again observed for knee abduction in the swing phase (σsess(lab) = 4.2°, r = 5.1) whereas its smallest inter-trial error was 0.8°.

3.4. Variance Analysis

When comparing the between-site CGA data from GL1 and GL2, the general linear model revealed significant differences (adjusted significance level: p ≤ 0.002) for 39% of all parameters (12/31) (Table 4). Only three parameters (hip abduction swing phase (degree), hip extension stance phase (% gait cycle), knee flexion swing phase (% gait cycle)) fulfilled all three statistical criteria (p ≤ 0.002 and ηp2 ≥ 0.14 and d ≥ 0.80).

3.5. Scatter-Plot Technique Suggested by Bland and Altman [22]

The Bland and Altman [22] scatter-plot technique revealed that the largest (worst) bounding criterion was for frontal plane knee angle (−2.3 ± 16.1°) (Figure 3d). The hip joint angles (Figure 3a,b), as well as the sagittal plane angles of the knee joint (Figure 3c,d), had the lowest bounding range. Ankle angles in general had a small bounding range (Figure 3e,f), with the exception of the maximum plantarflexion angle during the swing phase (−3.6 ± 10.7°).

3.6. Joint-Related Reliability

Inter- and intraobserver reliability for the ankle was in the fair-to-good and excellent ranges, in contrast to the knee and hip, which showed excellent reliability only (Table 2, Table A2). The within-site interobserver reliability for the ankle was ICCmean = 0.58 (95% CI: 0.33–0.73) compared with ICCmean = 0.81 (95% CI: 0.66–0.88) for intraobserver reliability. This was much lower than the reliability for the knee (interobserver ICCmean = 0.83 (95% CI: 0.74–0.89), intraobserver ICCmean = 0.98 (95% CI: 0.94–0.99)) and hip (interobserver ICCmean = 0.90 (95% CI: 0.84–0.94), intraobserver ICCmean = 0.97 (95% CI: 0.95–0.99)) (Table 2).
Between-laboratory intraobserver reliability showed fair-to-good reliability for all joints. Within this reliability level, small differences were found for the knee (ICCmean = 0.46 (95% CI: 0.05–0.68)) in comparison to the ankle (between-laboratory intraobserver ICCmean = 0.58) and hip (between-laboratory intraobserver ICCmean = 0.65 (95% CI: 0.27–0.81)).
Based on the ICCtotal, significant differences were calculated between all three types of reliability (p < 0.001) (Table 2).

4. Discussion

The ability to exchange and compare clinical gait analysis data between laboratories is a valuable tool to improve the monitoring of clinical treatments and rehabilitation progressions for patients with musculoskeletal disease. Gait analysis data was thus compared in three ways in the present study. First, to determine interobserver reliability; second, to determine intraobserver reliability; and finally, to determine between-laboratory intraobserver reliability of CGA data between two gait laboratories.
In support of our first hypothesis, the total mean reliability was excellent for interobserver (ICCtotal = 0.79, 95% CI: 0.67–0.86) and intraobserver reliability (ICCtotal = 0.93, 95% CI: 0.87–0.96) (Table 2). For a single laboratory, we found fair-to-good or poor interobserver reliability, which suggests an observer influence dependent on subjective observer variability and the different software used for analysis (Workstation vs. Nexus) (Table 1). These results support those of previous work by Schwartz et al. [10] (Table 3). We found excellent intraobserver reliability for one observer processing the data using the same software (excluding observer and software error). Nonetheless, ICC values for inter- and intraobserver reliability still fall under the excellent classification for a majority of the parameters observed (interobserver: 71% parameters; intraobserver: 94% parameters).
The ankle displayed fair-to-good and excellent interobserver reliability, whereas the knee and hip showed excellent reliability only (Table 2). This outcome is in general agreement with the systematic review of McGinley et al. [3], who reported excellent intra- and interobserver reliability for the parameters of the hip and knee joints in the sagittal-plane, and a lower reliability for the ankle joint. However, the observed differences within ankle kinematics may in part be attributed to technical improvements in the applied software (Nexus, version 1.3 vs. Workstation, version 4.6 build 142). However, future research is needed to investigate this possibility.
For total mean reliability, we observed fair-to-good between-laboratory intraobserver reliability (ICCtotal = 0.56, 95% CI: 0.16–0.74) and excellent inter- (ICCtotal = 0.79, 95% CI: 0.67–0.86) and intraobserver reliability (ICCtotal = 0.93, 95% CI: 0.87–0.96). Only 26% of the between-laboratory intraobserver parameters revealed excellent ICC values, whereas the remaining 74% fell within the fair-to-good and poor ranges. Significant differences (p ≤ 0.002) were found in 39% of the compared parameters (Table 4). This fair-to-good and poor reliability may be caused by differences in the measurement system hardware configurations and by slight differences in marker placement between data collection sessions. Both laboratories employed electro-optical motion capturing from the same manufacturer, but use a different hardware configuration (capturing software, camera type).
According to our second hypothesis, we believed that CGA data could be accurately collected between two gait laboratories, making such data interchangeable. An acceptance of this hypothesis could be affirmed for 61% of the parameters, with significant differences (p ≤ 0.002) in the remaining 39% (Table 4). Between-laboratory intraobserver reliability of CGA data captured at two different gait analysis laboratories had total ICC values at the fair-to-good level, which was lower than the excellent inter- and intraobserver reliability values (Table 2). In general, differences were found at each joint (ankle, knee, hip) for angle as well as for time-dependent (% gait cycle) analysis. The Bland–Altman plots indicated a detectable amount of variation for specific parameters, such as the knee in the frontal plane (Figure 3d), when compared between laboratories. These plots also showed a small data distribution for each individual observer, when compared between laboratories and the largest (worst) bounding criteria for frontal plane knee angle (2.3 ± 16.1°) (Figure 3d).
The standard errors for between-laboratory intraobserver analysis (maximum abduction at stance phase σsess(lab) = 4.2, maximum adduction at swing phase σsess(lab) = 3.4, maximum second abduction swing phase σsess(lab) = 3.2) supports the low between-laboratory intraobserver reliability found for the knee (ICCmean = 0.46, Table 2). These values are twice as large as for inter- (maximum first abduction stance phase σobserver = 2.4; maximum adduction/abduction swing phase σobserver = 1.5; maximum second abduction swing phase σobserver = 1.3) and intraobserver reliability (maximum first abduction stance phase σrepeated = 0.8; maximum adduction/abduction swing phase σrepeated = 1.0; maximum second abduction swing phase σrepeated = 0.7). The lower between-laboratory intraobserver reliability and significant differences observed indicate an influence of the applied motion capturing camera technology on the captured CGA data between testing sites. In contrast to our measure of intra- and interobserver variability in the between-laboratory comparison, the effect of different hardware (cameras, lenses, Analog-to-digital (A/D) converter, etc.), as well as marker removal and replacement between data sets can also contribute to lower reliability. The reduction in reliability suggests that the detection of tracked markers may be affected by the different camera resolutions investigated (Vcam with 0.3 million pixels, 659 x 439 black/white pixel sensor resolution vs. MXF with 2 million pixels, 1600 x 1280 grayscale pixel sensor resolution). We believe the effect of the measurement protocol and observer dependence are probably small in comparison to available technical requirements, existing in the different generations of cameras (i.e., sensor technology) because the ICC values of the one-site interobserver reliability were much higher, falling well within the excellent classification. These findings were somewhat different than those reported by Bucknall et al. [11], who observed apparent differences during motion data captured simultaneously using three camera systems (612, MX-13 and MX-F40).
Results of this study support the findings of Gorton et al. [8] and Bucknall et al. [11], which showed a dependence of CGA data quality on system resources, as well as applying standardized measurement protocols. The fact that CGA data quality could be affected by different camera types, even ones from the same manufacturer, is problematic, not only in the context of multicenter investigations, but it could also be a problem when comparing CGA data within institutions that have recently updated their laboratory equipment. Based on these findings when multicenter investigations are considered, it is important that both laboratories use the same or comparable camera equipment and software.
A limitation of this investigation was the sample size (n = 3). Each subject performed 20 barefoot gait trials using a self-selected walking speed. Subjects had no gait pathologies, resulting in similar gait patterns. This approach included the risk of correlated observations. On the other hand, it was not feasible to conduct our investigation with actual patients (with gait pathologies) and a powered sample, due to time restrictions in subject preparation (e.g., time required for marker application) and transportation limitations between testing sites.

5. Conclusions

The results of this study showed higher intraobserver reliability than interobserver reliability for CGA data conducted in a single gait analysis laboratory. There was weaker intra- and interobserver reliability ankle kinematics when compared to the hip and knee. Inter- or intraobserver reliability for data collected at a single laboratory was much stronger compared to data collected between laboratories. The outcomes of this study indicate that CGA results are probably influenced by testing conditions, such as laboratory equipment and software (capturing software and camera type). Peer-reviewed literature has reported an effect on the repeatability of motion capture data using separate trials, sessions and observers, etc. [5,6,7,9,10], and for multicenter repeatability focused on different applied hardware configurations, marker placement, as well as between trials and days of measurement [3,8,11]. The results of the current study support these previous investigations, and suggest that when multicenter investigations are considered, it is important that both laboratories use the same or comparable camera equipment and software. The reduced between-laboratory intraobserver reliability implies that comparisons of CGA data between centers with varying measurement equipment are generally not recommended.

Author Contributions

Conceptualization, R.S., F.S., C.H. and R.W.; methodology, R.W., R.S., F.S. and C.H.; formal analysis, R.W., R.S. and F.S.; investigation, R.W., R.S. and F.S.; resources, C.H.; data curation, R.S.; writing—original draft preparation, R.W., R.S. and F.S.; writing—review and editing, R.S., F.S., C.H., K.L. and R.W.; visualization, F.S.; supervision, C.H.; project administration, R.S. and F.S.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors wish to thank Gavin D. Olender for his assistance during preparing this manuscript. We would although like to thank Martin Scott-Löhrer from prophysics AG (Zürich, Switzerland) for his assistance and support. We thank Siegfried Leuchte and Henning Windhagen for the supervision and support of this project. This manuscript is dedicated to Siegfried Leuchte, who passed away unexpectedly at the end of 2018. We honor him as a committed and insightful Professor and keep him in honorable memory. This research was made possible out of research cooperation between members of the Musculosceletal Biomechanics Network (www.msb-net.org) of the DGOOC (German Association of Orthopedics and Orthopedic Surgery).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Anthropometric data for each individual test subject.
Table A1. Anthropometric data for each individual test subject.
Age
[years]
Body Weight
[kg]
Body Height
[cm]
Body Mass Index
[kg∙m-2]
Subject 1295916721
Subject 2378317826
Subject 3349018526
Mean33.377.317724.3
SD4.0416.39.072.89
Table A2. Descriptive statistics (mean ± standard deviation) and SEM values reported for intra- and interobserver reliability to describe the variability within and between subjects.
Table A2. Descriptive statistics (mean ± standard deviation) and SEM values reported for intra- and interobserver reliability to describe the variability within and between subjects.
ParameterIntraobserverInterobserver
Exam 1Exam 2SEMObserver 1Observer 2SEM
Hip (degree)
Flexion ST38.2 ± 1.7338.1 ± 1.68038.1 ± 1.6838.1 ± 1.690.29
Extension ST-6.58 ± 3.68-6.57 ± 3.670-6.57 ± 3.67-6.24 ± 4.492.00
Flexion SW37.6 ± 1.6237.6 ± 1.58037.6 ± 1.5837.7 ± 1.610.16
Adduction ST6.69 ± 3.716.71 ± 3.7206.71 ± 3.726.53 ± 3.740.53
Abduction SW-7.01 ± 1.38-7.02 ± 1.370-7.02 ± 1.37-6.96 ± 1.450.37
Hip (% gait cycle)
Flexion ST3.55 ± 3.193.45 ± 3.1703.45 ± 3.173.73 ± 3.380.66
Extension ST51.0 ± 1.1750.7 ± 1.100.3250.7 ± 1.1051.0 ± 1.130.49
Flexion SW86.6 ± 3.0586.6 ± 3.941.4486.6 ± 3.9486.2 ± 3.261.94
Adduction ST25.2 ± 8.3825.2 ± 8.15025.2 ± 8.1524.8 ± 7.811.78
Abduction SW65.7 ± 3.3965.4 ± 3.200.3365.4 ± 3.2065.1 ± 3.530.82
Knee (degree)
Flexion ST27.2 ± 3.0327.1 ± 3.02027.1 ± 3.0227.1 ± 3.160.93
Extension ST7.91 ± 2.397.96 ± 2.3507.96 ± 2.357.86 ± 2.440
Flexion SW68.3 ± 2.3368.3 ± 2.36068.3 ± 2.3667.7 ± 4.582.53
Range of motion60.4 ± 2.2760.4 ± 2.27060.4 ± 2.2759.9 ± 4.482.61
First abduction SW8.61 ± 4.038.62 ± 4.0608.62 ± 4.068.44 ± 3.810.39
Adduction SW0.44 ± 2.410.39 ± 2.4500.39 ± 2.450.60 ± 2.330.41
Second abduction SW5.79 ± 1.585.86 ± 1.6005.86 ± 1.605.75 ± 1.470.27
Knee (% gait cycle)
Flexion ST11.9 ± 1.8811.8 ± 1.820.1911.8 ± 1.8211.9 ± 1.690.61
Extension ST38.8 ± 1.7538.7 ± 1.720.2538.7 ± 1.7238.4 ± 1.890.44
Flexion SW72.4 ± 0.9772.0 ± 1.030.3372.0 ± 1.0371.9 ± 1.090.42
First abduction SW65.0 ± 3.2264.5 ± 3.010.5464.5 ± 3.0164.8 ± 2.961.19
Adduction SW76.6 ± 2.2076.1 ± 2.230.3176.1 ± 2.2375.5 ± 3.111.41
Second abduction SW86.4 ± 3.3185.8 ± 3.300.4785.8 ± 3.3085.8 ± 3.781.06
Ankle (degree)
Plantarflexion ST0.37 ± 1.321.34 ± 1.111.071.34 ± 1.11-0.17 ± 1.241.11
Dorsiflexion ST16.6 ± 1.1117.5 ± 2.000.9717.5 ± 2.0016.0 ± 1.681.34
Plantarflexion SW-8.66 ± 6.08-7.68 ± 6.690.90-7.68 ± 6.69-9.01 ± 5.641.63
Dorsiflexion SW8.92 ± 1.569.76 ± 2.770.929.76 ± 2.778.45 ± 2.251.33
Ankle (% gait cycle)
Plantarflexion ST5.35 ± 1.495.22 ± 1.650.225.22 ± 1.655.57 ± 1.310.42
Dorsiflexion ST47.0 ± 1.4146.9 ± 1.270.3046.9 ± 1.2745.7 ± 4.462.53
Plantarflexion SW63.4 ± 1.1263.1 ± 1.210.3363.1 ± 1.2162.9 ± 1.040.45
Dorsiflexion SW87.6 ± 5.1687.7 ± 5.371.1887.7 ± 5.3785.7 ± 3.813.25
Remarks: ST = stance phase; SW = swing phase; Exam = examination.

References

  1. Sander, K.; Rosenbaum, D.; Bohm, H.; Layher, F.; Lindner, T.; Wegener, R.; Wolf, S.I.; Seehaus, F. Instrumented gait and movement analysis of musculoskeletal diseases. Orthopade 2012, 41, 802–819. [Google Scholar] [CrossRef] [PubMed]
  2. Baker, R. Gait analysis methods in rehabilitation. J. Neuroeng. Rehabil. 2006, 3, 4. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. McGinley, J.L.; Baker, R.; Wolfe, R.; Morris, M.E. The reliability of three-dimensional kinematic gait measurements: A systematic review. Gait Posture 2009, 29, 360–369. [Google Scholar] [CrossRef] [PubMed]
  4. Ferrari, A.; Benedetti, M.G.; Pavan, E.; Frigo, C.; Bettinelli, D.; Rabuffetti, M.; Crenna, P.; Leardini, A. Quantitative comparison of five current protocols in gait analysis. Gait Posture 2008, 28, 207–216. [Google Scholar] [CrossRef] [PubMed]
  5. Noonan, K.J.; Halliday, S.; Browne, R.; O’Brien, S.; Kayes, K.; Feinberg, J.R. Interobserver variability of gait analysis in patients with cerebral palsy. J. Pediatr. Orthop. 2003, 23, 279–287. [Google Scholar] [CrossRef]
  6. Kadaba, M.P.; Ramakrishnan, H.K.; Wootten, M.E.; Gainey, J.; Gorton, G.; Cochran, G.V. Repeatability of kinematic, kinetic, and electromyographic data in normal adult gait. J. Orthop. Res. 1989, 7, 849–860. [Google Scholar] [CrossRef]
  7. Steinwender, G.; Saraph, V.; Scheiber, S.; Zwick, E.B.; Uitz, C.; Hackl, K. Intrasubject repeatability of gait analysis data in normal and spastic children. Clin. Biomech. 2000, 15, 134–139. [Google Scholar] [CrossRef]
  8. Gorton, G.E., 3rd; Hebert, D.A.; Gannotti, M.E. Assessment of the kinematic variability among 12 motion analysis laboratories. Gait Posture 2009, 29, 398–402. [Google Scholar] [CrossRef]
  9. Growney, E.; Meglan, D.; Johnson, M.; Cahalan, T.; An, K.-N. Repeated measures of adult normal walking using a video tracking system. Gait Posture 1997, 6, 147–162. [Google Scholar] [CrossRef]
  10. Schwartz, M.H.; Trost, J.P.; Wervey, R.A. Measurement and management of errors in quantitative gait data. Gait Posture 2004, 20, 196–203. [Google Scholar] [CrossRef] [Green Version]
  11. Bucknall, V.; Gibbs, S.; Arnold, G.P. Comparison of the kinematic data from three different Vicon systems. Gait Posture. Gait Posture 2008, 28, S33–S34. [Google Scholar] [CrossRef]
  12. Kolber, M.J.; Beekhuizen, K.; Cheng, M.S.; Fiebert, I.M. The reliability of hand-held dynamometry in measuring isometric strength of the shoulder internal and external rotator musculature using a stabilization device. Physiother. Theory Pract. 2007, 23, 119–124. [Google Scholar] [CrossRef] [PubMed]
  13. Kolber, M.J.; Hanney, W.J. The reliability and concurrent validity of shoulder mobility measurements using a digital inclinometer and goniometer: A technical report. Int. J. Sports Phys. Ther. 2012, 7, 306–313. [Google Scholar]
  14. Cools, A.M.; De Wilde, L.; Van Tongel, A.; Ceyssens, C.; Ryckewaert, R.; Cambier, D.C. Measuring shoulder external and internal rotation strength and range of motion: Comprehensive intrarater and interrater reliability study of several testing protocols. J. Shoulder Elbow Surg. 2014, 23, 1454–1461. [Google Scholar] [CrossRef] [PubMed]
  15. Mullaney, M.J.; McHugh, M.P.; Johnson, C.P.; Tyler, T.F. Reliability of shoulder range of motion comparing a goniometer to a digital level. Physiother. Theory Pract. 2010, 26, 327–333. [Google Scholar] [CrossRef] [PubMed]
  16. Muir, S.W.; Corea, C.L.; Beaupre, L. Evaluating change in clinical status: Reliability and measures of agreement for the assessment of glenohumeral range of motion. N. Am. J. Sports Phys. Ther. 2010, 5, 98–110. [Google Scholar]
  17. May, S.; Littlewood, C.; Bishop, A. Reliability of procedures used in the physical examination of non-specific low back pain: A systematic review. Aust. J. Physiother. 2006, 52, 91–102. [Google Scholar] [CrossRef] [Green Version]
  18. Wilhelmsen, K.; Strand, L.I.; Nordahl, S.H.G.; Eide, G.E.; Ljunggren, A.E. Psychometric properties of the Vertigo symptom scale: Short form. BMC Ear Nose Throat Disord. 2008, 8, 2. [Google Scholar] [CrossRef] [Green Version]
  19. Kadaba, M.P.; Ramakrishnan, H.K.; Wootten, M.E. Measurement of lower extremity kinematics during level walking. J. Orthop. Res. 1990, 8, 383–392. [Google Scholar] [CrossRef]
  20. Benedetti, M.G.; Catani, F.; Leardini, A.; Pignotti, E.; Giannini, S. Data management in gait analysis for clinical applications. Clin. Biomech. 1998, 13, 204–215. [Google Scholar] [CrossRef]
  21. Shrout, P.E.; Fleiss, J.L. Intraclass correlations: Uses in assessing rater reliability. Psychol. Bull. 1979, 86, 420–428. [Google Scholar] [CrossRef] [PubMed]
  22. Bland, J.M.; Altman, D.G. Statistical methods for assessing agreement between two methods of clinical measurement. Lancet 1986, 1, 307–310. [Google Scholar] [CrossRef]
  23. Portney, L.G.; Watkins, M.P. Foundations of Clinical Research: Applications to Practice, 3rd ed.; Pearson Prentice Hall: Upper Saddle River, NJ, USA, 2009. [Google Scholar]
  24. Weir, J.P. Quantifying Test-Retest Reliability Using the Intraclass Correlation Coefficient and the SEM. J. Strength Cond. Res. 2005, 19. [Google Scholar] [CrossRef]
  25. Richardson, J.T.E. Eta squared and partial eta squared as measures of effect size in educational research. Edu. Res. Rev. 2011, 6, 135–147. [Google Scholar] [CrossRef]
  26. Hartmann, A.; Herzog, T.; Drinkmann, A. Psychotherapy of bulimia nervosa: What is effective? A meta-analysis. J. Psychosom. Res. 1992, 36, 159–167. [Google Scholar] [CrossRef]
  27. Cohen, J. Statistical Power Analysis for the Behavioural Sciences, 2nd ed.; Lawrence Erlbaum: Hillsdale, NJ, USA, 1998. [Google Scholar]
Figure 1. Data processing chart—clinical gait analysis (CGA) data collected on three healthy test-subjects (n = 20 barefoot trials per person). The data were captured using the same measurement protocols in both the first (GL1) and the second laboratories (GL2). Three reliability tests were performed with the datasets from GL1 and GL2. Interobserver reliability was tested for two observers using the CGA data collected at GL2. Observer 1 used the data processing software from GL 1 (Workstation 4.6), which was different to that of Observer 2 (GL2: Nexus 1.3). To assess intraobserver reliability, the same data set was analyzed a second time by observer 1 after a time interval of 12 months, using the CGA data collected at GL2 and the software from GL1. To test between-laboratory intraobserver reliability, the CGA data collected at both sites (GL1 and GL2) were analyzed by Observer 1 using software from GL1.
Figure 1. Data processing chart—clinical gait analysis (CGA) data collected on three healthy test-subjects (n = 20 barefoot trials per person). The data were captured using the same measurement protocols in both the first (GL1) and the second laboratories (GL2). Three reliability tests were performed with the datasets from GL1 and GL2. Interobserver reliability was tested for two observers using the CGA data collected at GL2. Observer 1 used the data processing software from GL 1 (Workstation 4.6), which was different to that of Observer 2 (GL2: Nexus 1.3). To assess intraobserver reliability, the same data set was analyzed a second time by observer 1 after a time interval of 12 months, using the CGA data collected at GL2 and the software from GL1. To test between-laboratory intraobserver reliability, the CGA data collected at both sites (GL1 and GL2) were analyzed by Observer 1 using software from GL1.
Applsci 10 05068 g001
Figure 2. Parameters extracted (No. 1 to 16) out of processed kinematic data for sagittal (a) and frontal (b) plane of the hip joint, for sagittal (c) and frontal (d) plane of the knee joint, as well as data for sagittal (e) plane of the ankle joint.
Figure 2. Parameters extracted (No. 1 to 16) out of processed kinematic data for sagittal (a) and frontal (b) plane of the hip joint, for sagittal (c) and frontal (d) plane of the knee joint, as well as data for sagittal (e) plane of the ankle joint.
Applsci 10 05068 g002
Figure 3. Bland and Altman plots presenting the computed differences between the clinical gait analysis data from the first laboratory (GL1) and the second laboratory (GL2) for (ab) hip joint, (cd) knee joint and (ef) ankle joint. Each data point (test subject 1 = circle, test subject 2 = triangle, test subject 3 = square) represents the computed difference between CGA data of GL1 and GL2 (ordinate) which was plotted vs. the mean difference (abscissa) respectively, from the same healthy test subjects. The solid horizontal line represents the mean of all the differences plotted, and the two dashed lines represent the mean ± 2 SD (standard deviation) agreement interval, as defined by Bland and Altman [22].
Figure 3. Bland and Altman plots presenting the computed differences between the clinical gait analysis data from the first laboratory (GL1) and the second laboratory (GL2) for (ab) hip joint, (cd) knee joint and (ef) ankle joint. Each data point (test subject 1 = circle, test subject 2 = triangle, test subject 3 = square) represents the computed difference between CGA data of GL1 and GL2 (ordinate) which was plotted vs. the mean difference (abscissa) respectively, from the same healthy test subjects. The solid horizontal line represents the mean of all the differences plotted, and the two dashed lines represent the mean ± 2 SD (standard deviation) agreement interval, as defined by Bland and Altman [22].
Applsci 10 05068 g003
Table 1. Single site inter- and intraobserver intraclass correlation coefficient (ICC) values and results of repeated single observer analysis at both laboratories (between-laboratory intraobserver values) of an ICC greater than 0.75 were considered excellent and marked in bold.
Table 1. Single site inter- and intraobserver intraclass correlation coefficient (ICC) values and results of repeated single observer analysis at both laboratories (between-laboratory intraobserver values) of an ICC greater than 0.75 were considered excellent and marked in bold.
ParameterIntraobserverInterobserverBetween-Laboratory
Intraobserver
ICC95% CIICC95% CIICC95% CI
lowerupper lowerupper lowerupper
Hip (degree)
Flexion ST1.001.001.000.970.950.980.600.100.80
Extension ST1.001.001.000.760.630.850.720.540.83
Flexion SW1.001.001.000.990.970.990.7100.89
Adduction ST1.001.001.000.980.970.990.880.550.95
Abduction SW1.001.001.000.930.890.960.590.060.80
Hip (% gait cycle)
Flexion ST1.000.991.000.960.920.970.900.830.94
Extension ST0.920.770.970.810.640.900.460.000.71
Flexion SW0.830.720.900.710.560.810.310.000.57
Adduction ST1.001.001.000.950.920.970.780.630.87
Abduction SW0.990.980.990.940.900.970.530.230.72
Knee (degree)
Flexion ST1.001.001.000.910.860.950.570.080.78
Extension ST1.001.001.001.000.991.000.740.560.84
Flexion SW1.001.001.000.470.250.650.820.700.89
Range of motion1.001.001.000.400.170.590.460.100.68
First abduction SW1.001.001.000.990.980.990.000.000.35
Adduction SW1.001.001.000.970.940.980.410.000.67
Second abduction SW1.001.001.000.970.950.960.410.000.67
Knee (% gait cycle)
Flexion ST0.990.980.990.880.810.930.810.180.93
Extension ST0.980.970.990.940.830.970.210.000.53
Flexion SW0.890.660.950.840.750.900.130.000.42
First abduction SW0.970.930.980.840.740.900.480.120.69
Adduction SW0.980.920.990.720.550.830.190.000.52
Second abduction SW0.980.811.000.910.850.950.750.550.86
Ankle (degree)
Plantarflexion ST0.230.000.510.1100.3140.330.000.60
Dorsiflexion ST0.610.280.780.470.020.7210.740.570.85
Plantarflexion SW0.980.940.990.930.790.9680.860.620.94
Dorsiflexion SW0.820.630.900.720.250.8750.830.650.91
Ankle (% gait cycle)
Plantarflexion ST0.980.960.990.820.690.900.770.070.92
Dorsiflexion ST0.950.920.970.220.000.440.360.000.62
Plantarflexion SW0.920.820.960.840.750.900.270.000.55
Dorsiflexion SW0.950.920.970.500.260.680.450.060.68
ICC ≥ 0.75 (%)94% (29/31)71% (22/31)29% (9/31)
Remarks: ST = stance phase; SW = swing phase.
Table 2. Calculated mean ICC values (ICCmean = mean of all parameters for hip, knee and ankle) and 95% confidence intervals (95% CI) for the ankle, knee and hip, as well as total mean ICC values (ICCtotal = ICCmean hip + ICCmean knee + ICCmean ankle). Values for ICC above 0.75 were considered excellent and marked in bold.
Table 2. Calculated mean ICC values (ICCmean = mean of all parameters for hip, knee and ankle) and 95% confidence intervals (95% CI) for the ankle, knee and hip, as well as total mean ICC values (ICCtotal = ICCmean hip + ICCmean knee + ICCmean ankle). Values for ICC above 0.75 were considered excellent and marked in bold.
JointIntraobserver (95% CI)Interobserver (95% CI)Between-laboratory Intraobserver (95% CI)
Hip ∅ ICCmean0.97 (0.95–0.99)0.90 (0.84–0.94)0.65 (0.27–0.81)
Knee ∅ ICCmean0.98 (0.94–0.99)0.83 (0.74–0.89)0.46 (0.05–0.68)
Ankle ∅ ICCmean0.81 (0.66–0.88)0.58 (0.33–0.73)0.58 (0.21–0.76)
ICCtotal0.93 (0.87–0.96)0.79 (0.67–0.86)0.56 (0.16–0.74)
Analysis of Variance
(ICCtotal)
Comparison of all three types of reliability: p < 0.001; ηp2 = 0.533
Intraobserver vs. Interobserver: p < 0.001; ηp2 = 0.399
Intraobserver vs. Between-laboratory intraobserver: p < 0.001; ηp2 = 0.676
Interobserver vs. Between-laboratory intraobserver: p < 0.001; ηp2 = 0.374
Remarks: ICCmean = mean of all parameters for hip, knee and ankle; ICCtotal = ICCmean hip + ICCmean knee + ICCmean ankle divided by 3 (for all three types of reliability).
Table 3. Standard errors for intra- (σrepeated), inter- (σobserver) and between-laboratory intraobserver (σsess(lab)) reliability. The inter-trial error is represented by r (r = σrepeated / σtrial; r = σobserver / σtrial; r = σsess(lab)/ σtrial). Kinematic parameters for ankle-, knee- and hip-joints are presented by peak values (amplitude) and associated time point during the gait cycle.
Table 3. Standard errors for intra- (σrepeated), inter- (σobserver) and between-laboratory intraobserver (σsess(lab)) reliability. The inter-trial error is represented by r (r = σrepeated / σtrial; r = σobserver / σtrial; r = σsess(lab)/ σtrial). Kinematic parameters for ankle-, knee- and hip-joints are presented by peak values (amplitude) and associated time point during the gait cycle.
ParameterIntraobserverInterobserverBetween-Laboratory Intraobserver
σobserverrσrepeatedrσsess(lab)r
Joint Angles [degree]HipFlexion ST1.21.01.31.01.61.4
Extension ST0.81.00.91.01.61.7
Flexion SW0.71.00.91.01.21.5
Adduction ST1.01.01.21.01.51.6
Abduction SW0.71.00.91.21.31.4
KneeFlexion ST2.11.02.41.02.31.2
Extension ST1.31.01.91.31.71.4
Flexion SW1.11.01.81.41.61.5
Range of motion1.61.01.81.01.71.1
First abduction SW0.81.02.43.24.25.1
Adduction SW1.01.01.51.63.43.7
Second abduction SW0.71.01.31.63.23.8
AnklePlantarflexion ST1.61.41.51.51.61.4
Dorsiflexion ST1.71.31.51.31.41.1
Plantarflexion SW2.51.13.81.73.61.5
Dorsiflexion SW1.51.51.61.41.31.4
Time at % Gait CycleHipFlexion ST1.41.01.91.21.31.2
Extension ST0.81.00.91.21.21.2
Flexion SW2.91.03.21.04.21.1
Adduction ST3.81.04.41.05.41.0
Abduction SW1.51.02.51.55.01.0
KneeFlexion ST0.81.01.11.11.01.2
Extension ST1.41.01.71.11.41.1
Flexion SW0.91.01.01.10.91.2
First abduction SW2.71.03.01.02.11.0
Adduction SW2.01.03.11.02.40.9
Second abduction SW1.71.03.01.22.30.9
AnklePlantarflexion ST1.01.01.31.41.11.3
Dorsiflexion ST1.01.01.21.21.21.0
Plantarflexion SW0.91.01.01.11.11.3
Dorsiflexion SW4.21.04.41.14.81.1
Analysis of VarianceIntraobserver vs. Interobserver:
p < 0.001; ηp2 = 0.498
Intraobserver vs. Between-Laboratory Intraobserver:
p = 0.001; ηp2 = 0.325
Interobserver vs. Between-Laboratory Intraobserver:
p = 0.083; ηp2 = 0.097
Remarks: ST = stance phase; SW = swing phase.
Table 4. Single observer comparison between laboratories (means, standard deviations, general linear model (GLM), standard error of measurement (SEM)). Significance level was set at 0.002. Large effect sizes (d ≥ 0.8; ηp2 ≥ 0.14) marked in bold.
Table 4. Single observer comparison between laboratories (means, standard deviations, general linear model (GLM), standard error of measurement (SEM)). Significance level was set at 0.002. Large effect sizes (d ≥ 0.8; ηp2 ≥ 0.14) marked in bold.
ParameterGL1GL2Analysis of VariancedSEM
meanSDmeanSDFp ηp2
Hip (degree)
Flexion ST36.52.7938.11.682.440.1210.020.721.41
Extension ST-7.192.00-6.573.6720.0<0.0010.150.221.50
Flexion SW35.92.3437.61.580.030.8550.000.871.06
Adduction ST5.043.946.713.723.320.0710.030.441.33
Abduction SW-8.442.12-7.021.37141<0.0010.550.811.12
Hip (% gait cycle)
Flexion ST3.003.653.453.1779.2<0.0010.400.131.08
Extension ST52.01.6450.71.1035.2<0.0010.320.951.01
Flexion SW88.75.3286.63.940.130.7210.000.453.85
Adduction ST27.110.425.28.1525.7<0.0010.180.214.35
Abduction SW67.28.2365.43.200.670.4160.010.323.92
Knee (degree)
Flexion ST24.73.3027.13.021.620.2060.010.762.07
Extension ST7.253.817.962.3540.4<0.0010.260.231.57
Flexion SW67.94.4968.32.3616.0<0.0010.120.121.45
Range of motion60.71.7260.42.273.350.0700.030.151.47
First abduction SW6.336.778.624.061.900.1710.020.425.42
Adduction SW-0.077.730.462.489.260.0030.080.103.92
Second abduction SW4.967.335.831.600.000.9600.000.203.43
Knee (% gait cycle)
Flexion ST12.91.7011.81.8212.10.0010.090.630.77
Extension ST39.11.4938.71.720.090.7690.000.251.43
Flexion SW72.90.7872.01.0319.2<0.0010.141.000.84
First abduction SW64.51.8064.53.010.110.7400.0001.73
Adduction SW75.12.2176.22.240.800.3740.010.492.00
Second abduction SW85.33.5285.73.374.450.0350.040.121.72
Ankle (degree)
Plantarflexion ST0.062.791.341.1139.1<0.0010.250.661.60
Dorsiflexion ST18.02.6617.52.003.060.0830.030.221.19
Plantarflexion SW-11.310.7-7.686.69110<0.0010.490.423.25
Dorsiflexion SW8.692.969.762.779.060.0030.070.371.18
Ankle (% gait cycle)
Plantarflexion ST6.431.835.221.6512.70.0010.100.700.84
Dorsiflexion ST47.31.4846.91.270.010.9400.000.291.10
Plantarflexion SW64.41.0263.11.213.660.0580.031.170.95
Dorsiflexion SW91.45.3787.75.370.200.6560.000.693.98
Remarks: ST = stance phase; SW = swing phase; GL1 = laboratory 1; GL2 = laboratory 2.

Share and Cite

MDPI and ACS Style

Schwesig, R.; Wegener, R.; Hurschler, C.; Laudner, K.; Seehaus, F. Intra- and Interobserver Reliability Comparison of Clinical Gait Analysis Data between Two Gait Laboratories. Appl. Sci. 2020, 10, 5068. https://doi.org/10.3390/app10155068

AMA Style

Schwesig R, Wegener R, Hurschler C, Laudner K, Seehaus F. Intra- and Interobserver Reliability Comparison of Clinical Gait Analysis Data between Two Gait Laboratories. Applied Sciences. 2020; 10(15):5068. https://doi.org/10.3390/app10155068

Chicago/Turabian Style

Schwesig, René, Regina Wegener, Christof Hurschler, Kevin Laudner, and Frank Seehaus. 2020. "Intra- and Interobserver Reliability Comparison of Clinical Gait Analysis Data between Two Gait Laboratories" Applied Sciences 10, no. 15: 5068. https://doi.org/10.3390/app10155068

APA Style

Schwesig, R., Wegener, R., Hurschler, C., Laudner, K., & Seehaus, F. (2020). Intra- and Interobserver Reliability Comparison of Clinical Gait Analysis Data between Two Gait Laboratories. Applied Sciences, 10(15), 5068. https://doi.org/10.3390/app10155068

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop