Accuracy and Reliability of Visual Inspection and Smartphone Applications for Measuring Finger Range of Motion

Accuracy and Reliability of Visual Inspection and Smartphone Applications for Measuring Finger Range of Motion

Reliable and accurate measurement of range of motion (ROM) of a joint is an integral part of the physical examination in hand surgery. Handheld goniometer measurements are often considered the gold standard. Goniometers are the most widely used, often studied, economical, and portable devices for measuring joint ROM. However, neither goniometer nor visual inspections has shown high levels of accuracy compared with radiographic measurements in finger and wrist joints.

Recently, several new smartphone-based applications (apps) allowing measurement of ROM have been introduced. With the widespread use of smartphones, these apps may offer new means of providing accurate and reliable measures of ROM, especially in clinical situations where the standard goniometer and/or radiographs may be unavailable. If these apps were proven to be reliable and accurate, apps could be designed to automate the recording of these measurements and possibly even place them into the electronic medical record, improving efficiency for both therapists and clinicians.

The purposes of this study were to compare different smartphone apps and visual inspection measurements of the finger joints with the goniometer measurements of radiographs and to determine interrater reliability for these measurement tools.

Finger flexion at the metacarpophalangeal joint (MCPJ), proximal interphalangeal joint (PIPJ), and distal interphalangeal joint (DIPJ) were independently measured by a hand fellowship–trained surgeon (J.R.F.) and 2 residents (H.H.L., K.S.L.). The index finger of a healthy subject with no known pathology was held in place using an aluminum-alloy splint (Figure 1A). A fluoroscopic image was acquired by a mini C-arm (Figure 1B). An independent observer measured each joint flexion angle on the fluoroscopic image using a universal handheld goniometer. This measurement was used as the reference standard. Finger flexion angle reference measurements. The index finger of a healthy subject with no known pathology was held in place by an aluminum-alloy splint (A). A fluoroscopic image was acquired by a mini C-arm. An independent observer measured each joint flexion angle using a universal handheld goniometer (B). Three different iPhone 6 (Apple, Cupertino, California) apps were used to measure the flexion angles of the finger according to manufacturer instructions: Goniometer (June Infrastructure Pvt Ltd, Viman Nagar, Pune, India), iPhone Compass (Apple), and PT-Tools Suite (by David Raney). Each observer measured the MCPJ, PIPJ, and DIPJ of the finger held in the aluminum-alloy splint once. In addition, visual inspection was used to estimate the flexion angles of the MCPJ, PIPJ, and DIPJ. A total of 9 trials were performed, each with a varied finger flexion configuration selected by the independent observer. Ease of use was rated by each observer. Statistical analyses were performed using SPSS software (IBM, Armonk, New York). The Bland–Altman method was used to evaluate the agreement of each measurement tool with the standard measurement. Comparisons of each of the measurement tools against the standard measurement were performed via 2-tailed Pearson correlation, as well as one-way analysis of variance and Kruskal–Wallis test for parametric and nonparametric data, respectively. Interrater reliability was quantified using interclass correlation coefficient.

The “user-friendliness” of each measurement tool is presented in Table 1. All 3 observers found visual inspection to be the easiest. Ease of Use of Measurement Methods Figure 2 documents all measured readings of finger joints for all 3 observers. Each data point represents an observer's measured angle of the MCPJ, PIPJ, or DIPJ using visual inspection, Goniometer, iPhone Compass, or PT-Tools Suite apps vs the measurement of fluoroscopic images of each joint. Measured angles vs the fluoroscopic image (XR) reference standard. Visual inspection (A), Goniometer (June Infrastructure Pvt Ltd, Viman Nagar, Pune, India) (B), iPhone Compass (Apple, Cupertino, California) (C), and PT-Tools Suite (by David Raney) (D). Overall agreement of the smartphone apps as well as visual inspection measurements of the finger joints with the fluoroscopic image measurement, as assessed by the Bland–Altman method, is illustrated in Figure 3 and Table 2. Visual inspection had 3 cases (3.7%), iPhone Compass had 4 cases (4.9%), and Goniometer and PT-Tools Suite had 5 cases (6.2%) outside of the 95% agreement range. Goniometer and iPhone Compass had the mean difference between the measurement and fluoroscopic image standard close to zero, and Goniometer had the smallest standard deviation of differences. When the subanalysis was performed for each joint, the MCPJ had the greatest agreement, with Goniometer and PT-Tools Suite having no cases outside of the agreement range (data not shown). Bland–Altman plot for agreement with fluoroscopic image (XR) reference standard. Visual inspection (A), Goniometer (June Infrastructure Pvt Ltd, Viman Nagar, Pune, India) (B), iPhone Compass (Apple, Cupertino, California) (C), and PT-Tools Suite (by David Raney) (D). Average differences between the fluoroscopic image measurements with the measured angles using the tools studied were 10.8° for visual inspection, 9.4° for Goniometer, 10.4° for iPhone Compass, and 12.2° for PT-Tools Suite. When the authors dropped the allowed difference between the fluoroscopic image measurements to the measured angles to ±10°, visual inspection had 65.4%, Goniometer had 59.3%, iPhone Compass had 49.4%, and PT-Tools Suite had 51.9% of cases within the range. No statistical differences were observed among the 4 measurement tools for either average differences (P=.214) or percentage of measures that are within 10% from the fluoroscopic image standard (P=.159). The Pearson r coefficients were 0.8149 for visual inspection (range, 0.7435–0.8660), 0.7061 for Goniometer (range, 0.6463–0.7388), 0.7201 for iPhone Compass (range, 0.5875–0.8283), and 0.7118 for PT-Tools Suite (range, 0.6481–0.7503) (Table 3). The data were subdivided into each joint using each of the measurement tools. The PIPJ was found to have the strongest correlation when compared with the other joints, except when measured with the iPhone Compass app. Overall interrater reliability was excellent—the interclass correlation coefficient was 0.950 (range, 0.931–0.964). Interrater reliability was further analyzed for each measurement tool as well as each joint (Table 4). Visual inspection and the Goniometer app had the highest correlation coefficients.

The purposes of this study were to (1) compare different smartphone apps as well as visual inspection measurements of the finger joints with goniometer measurements of the fluoroscopic image and (2) determine interrater reliability for these measurement tools. The results of this study suggest that all 3 smartphone measurement tools, as well as visual inspection, agree and correlate well with the reference standard fluoroscopic image measurement. Average differences between the fluoroscopic image measurements and the tools used in this study ranged from 9.4° to 12.2°, with 49.4% to 65.4% of measurements within ±10° of references. The mean correlation coefficients for each iPhone app exceeded 0.7, signifying strong correlation. The strongest correlations were between visual inspection and the fluoroscopic image reference standard. However, this was not the case for the Bland–Altman agreement. Overall interrater reliabilities were similar, with the interclass correlation coefficient being greater than 0.9 for all of the measurement tools. Visual inspection and the Goniometer app had the highest correlation coefficients. Although there was a strong correlation between visual inspection and the fluoroscopic image reference standard, agreement and accuracy between the 2 were not as strong. Rose et al estimated the angles of the MCPJ, PIPJ, and DIPJ of a resin cast of an adult hand and reported 25% inaccuracy using visual estimate. However, visual estimates of the surgeons were better than those of the physiotherapists. In addition, both hand surgery experience and stated interest in hand surgery correlated with the accuracy, which may be due to improvement with time and practice. Nonetheless, Rose et al concluded that visual assessment alone is imprecise and recommended that goniometers be used in evaluating finger angles. McVeigh et al found that neither visual inspection nor handheld goniometer measurements were consistently accurate when compared with radiographic measurement. This is similar to findings in the current study when the authors dropped the allowed difference range between the measured angles to the fluoroscopic image reference to ±10°. In addition, the authors found no difference in accuracy of visual inspection between hand surgeons and therapists. The accuracies of the wrist and MCPJ angles were similar between the visual estimate and Goniometer measurement, but the PIPJ angle was more accurate with the Goniometer measurement. The authors found a high degree of interobserver variability in measurements. The overall strong interrater reliability and significant correlation between the visual inspection and fluoroscopic image reference standard may be explained by all observers being orthopedic surgeons and one having hand fellowship training. Among the finger joints, the MCPJ showed the strongest agreement, whereas the PIPJ showed the strongest correlation. Ellis et al measured the finger joint angles using a goniometer and wire tracings. They showed that the MCPJ had the greatest repeatability when compared with the more distal joints, with the DIPJ having the greatest variability. They postulated that this was due to the shorter lever arms to align the measuring tools, a less definable joint line, and lack of a distal bony landmark for alignment. A study by Macionis confirmed this finding. This study had several limitations. First, only 3 observers were included in the study. Other studies have included a greater number of examiners, and this could have affected the interrater reliability and the strength of correlation. In addition, the subject had no known finger pathology, as this was a feasibility study regarding whether smartphone apps can be used as additional tools to measure ROM of healthy finger joints. How pain and pathology may affect these measurements in the setting of pathology is unknown. Furthermore, only flexion/extension angles of the finger joints were measured, without fully accounting for the overall hand's place in space. The wrist position inclination and supination/pronation that could have affected the smartphone app measurements were not controlled. Finally, the reference standard was the actual bony joint flexion angles as measured by radiographs, whereas the app measurements as well as visual inspection were applied over the soft tissue. This is an inherent shortcoming of clinical measure of ROM, as frequent radiographic measurements can be cost prohibitive.

Images Powered by Shutterstock