Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Monday, March 11, 2019 at 4:00 PM to 4:30 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Advertisement

Citing this Article

Right click to copy or hit: ctrl+c (cmd+c on mac)

Published on 06.06.19 in Vol 4, No 1 (2019): Jan-Dec

Preprints (earlier versions) of this paper are available at http://preprints.jmir.org/preprint/12291, first published Sep 21, 2018.

This paper is in the following e-collection/theme issue:

    Original Paper

    Determining the Accuracy of Oculus Touch Controllers for Motor Rehabilitation Applications Using Quantifiable Upper Limb Kinematics: Validation Study

    RREACH (Robotics for Rehabilitation Exercise and Assessment in Collaborative Healthcare) Lab, Department of Mechanical Engineering, The University of British Columbia, Vancouver, BC, Canada

    Corresponding Author:

    Leia C Shum, BASc

    RREACH (Robotics for Rehabilitation Exercise and Assessment in Collaborative Healthcare) Lab

    Department of Mechanical Engineering

    The University of British Columbia

    ICICS X015

    2335 Engineering Rd

    Vancouver, BC, V6T 1Z4

    Canada

    Phone: 1 (604) 822 3147

    Email: lcshum@alumni.ubc.ca


    ABSTRACT

    Background: As commercial motion tracking technology becomes more readily available, it is necessary to evaluate the accuracy of these systems before using them for biomechanical and motor rehabilitation applications.

    Objective: This study aimed to evaluate the relative position accuracy of the Oculus Touch controllers in a 2.4 x 2.4 m play-space.

    Methods: Static data samples (n=180) were acquired from the Oculus Touch controllers at step sizes ranging from 5 to 500 mm along 16 different points on the play-space floor with graph paper in the x (width), y (height), and z (depth) directions. The data were compared with reference values using measurements from digital calipers, accurate to 0.01 mm; physical blocks, for which heights were confirmed with digital calipers; and for larger step sizes (300 and 500 mm), a ruler with hatch marks to millimeter units.

    Results: It was found that the maximum position accuracy error of the system was 3.5 ± 2.5 mm at the largest step size of 500 mm along the z-axis. When normalized to step size, the largest error found was 12.7 ± 9.9% at the smallest step size in the y-axis at 6.23 mm. When the step size was <10 mm in any direction, the relative position accuracy increased considerably to above 2% (approximately 2 mm at maximum). An average noise value of 0.036 mm was determined. A comparison of these values to cited visual, goniometric, and proprioceptive resolutions concludes that this system is viable for tracking upper-limb movements for biomechanical and rehabilitation applications. The accuracy of the system was also compared with accuracy values from previous studies using other commercially available devices and a multicamera, marker-based professional motion tracking system.

    Conclusions: The study found that the linear position accuracy of the Oculus Touch controllers was within an agreeable range for measuring human kinematics in rehabilitative upper-limb exercise protocols. Further testing is required to ascertain acceptable repeatability in multiple sessions and rotational accuracy.

    JMIR Biomed Eng 2019;4(1):e12291

    doi:10.2196/12291

    KEYWORDS



    Introduction

    Current gaming and virtual reality platforms [1] that use motion-controlled interfaces offer an affordable and accessible method of tracking human kinematics. However, given that consumer-grade platforms are originally intended for playing video games and to immerse players in virtual environments, their tracking performance should be evaluated before they are employed as tools for biomechanical or clinical analysis [2]. Previously tested rehabilitation protocols using commercial gaming technology such as Wii Motes (Nintendo Co, Ltd, Kyoto, Japan) to provide positional feedback for trunk compensation [3] or a Kinect (Microsoft Corporation, Redmond, United States) to measure range and speed of motion for upper-limb exercises [4,5] have shown potential to be used as rehabilitation tools that could provide quantifiable changes in clients’ kinematic motor abilities to therapists. Other studies using accelerometers to track patterns in functional upper-limb movements were able to capture differences similar to those measured by clinical scales [6] and found benefits from objective quantitative evaluations of changes in motor ability during therapy regimens, which can be collected from in-game progress reports [7]. In addition, success has been found in translating kinematic upper-limb metrics to clinical Fugl-Meyer scoring [8] and in detecting exercise repetitions via kinematic monitoring for telerehabilitation and at-home programs [9]. Current clinical assessments for upper-limb motor function, such as the Fugl-Meyer Assessment and Wolf Motor Function Test, only provide low-resolution point-scores rated qualitatively by therapists, and kinematic analysis of upper-limb motion has been reported to be a useful addition to these clinical assessments [10]. When measuring range of motion in a clinical setting, the goniometer is considered a gold-standard clinical measurement tool used by therapists [11]. However, only static joint angles can be measured, and typically with some visual estimation and multiple testers [12].

    One of the latest (released December 2016) devices to be developed for interacting with virtual environments is the Oculus Touch (Oculus VR, LLC, Menlo Park, CA, United States) controller set. The controllers are peripheral accessories of the Oculus Rift virtual reality headset and are employed to track users’ hand movements. Their tracking system employs a proprietary algorithm that collects data from infrared sensors via constellation tracking [13] and inertial measurement units (IMUs). Given that the controllers are wireless, lightweight, low-cost devices that can be used to track a user’s hand position and orientation in 3-dimensional (3D) space, they could have the potential to be employed in rehabilitative and biomechanical motion-tracking applications. At the time of this study, there was no sufficient information about the tracking performance of the controllers provided by the manufacturer, and there is currently a lack of scientific papers employing a systematic approach to test their potential application as tools for motion-tracking data capture. As a result, in this study, we evaluated the tracking accuracy of the Oculus Touch controllers to present a preliminary evaluation that could be informative to the biomechanical and rehabilitation research community. The specific aim of the experiment was to quantify the relative positional accuracy of the Oculus Touch controllers in 3 spatial dimensions. As the controllers are intended for hand-held motion control, the evaluation setup was centered around the movement size for standing/sitting upper-limb reaching tasks.


    Methods

    Technical Setup

    An Oculus Touch controller (Figure 1), 2 Oculus Sensors, an Oculus Rift headset, and a computer running Windows 10 (Microsoft Corporation) were employed in this study.

    A custom computer application was developed in Unity 2017 (Unity Technologies, San Francisco, United States) to capture and log the controller’s position during the experiment. The data capture was performed at the headset’s native frequency of approximately 90 Hz , using the Unity OVR Plugin package to access controller data. The virtual environment was set up over a 2.4 m x 2.4 m play-space in the x-z plane to be within the recommended manufacturer play area. This space consists of 16 commercial 600 mm square force/torque plates professionally installed on a subfloor of auto-levelling epoxy and flat to within 0.5 mm (Figure 2). The y-axis was only bounded by the camera sensors’ field of view limitations.

    To ensure consistency, the Oculus Sensors were placed on the floor at 0.3 m along the front edge of the space and 1.2 m apart, equidistant from the centre line, for the entire experiment. The sensor heads were manually leveled and visually aligned to have parallel, front-facing fields of views. Both the sensors and controllers maintained an initial y-position of 0 at the floor—this would be equivalent to placing the sensors at table height and the controllers at hand height.

    All measurements were taken by securing the right-hand Oculus Touch controller to a flat L-shaped jig (Figure 2) and resting it on the floor for 5 seconds. Initial calibration of floor height and play-space size and orientation was done through the official commercial Oculus setup client.

    Figure 1. The right-side Oculus Touch controller. Left: front view. Right: top-down view.
    View this figure
    Figure 2. The experimental setup and coordinate frame. The play-space was divided into 16 squares.
    View this figure

    Experimental Procedure

    Measurements were taken along each of the 3 spatial axes (x: width, y: height, z: depth) in a single session. The x and z axes were measured in increments of 5, 10, 50, 150, 300, and 500 mm steps relative to a recorded 0 value. The estimate for a 500 mm largest step was attributed to an approximate lower arm and hand length from human anthropometric data in Huston [14].

    This length should replicate the size of a simple outward reach from the elbow. Each set of steps was taken from the zero line of each axis in both the positive and negative direction and then taken in the positive direction at +600 mm and in the negative direction at –600 mm along the same axis (Figure 3; left). Graph paper with millimeter unit markings was used to define the step sizes to the relative 0 point of each set. The graph paper step sizes were verified using a Mitutoyo 500-196 digital calipers, accurate to 0.01 mm, visually aligned to the edge of the unit markings within the third significant digit. A ruler with half-millimeter unit markings was used for steps larger than 150 mm. The bottom left corner of the jig was used as an origin for the 3 axes with respect to the controller. The x and z axes edges were aligned with the graph paper visually. The L-shaped jig was checked for orthogonality using a calibrated 90-degree ruler in all 3 directions before its use. To test the repeatability of the L-shaped jig alignment on graph paper, the controller was moved at least 300 mm away from and then toward a single point near the centre of the play-space on each axis 3 times.

    In the x-axis, 4 sets of steps were taken at 4 different depths for a total of 16 sets of steps to measure the accuracy of the controllers over the play-space area (Figure 3). The same configuration was used for the z-axis but using 4 sets of steps along 4 different x-axis values, 600 mm apart.

    Figure 3. Left: A top-down visual representation of the expected spacing of the data points in the x-axis. Right: The x-z points at which the y step sets were taken.
    View this figure
    Figure 4. Close-up of the Oculus Touch controller in the L-shape jig over an aluminium block used to measure y-axis steps. Both controller (system) and graph paper (physical world) axes are represented and were aligned visually during the experiment.
    View this figure

    The 16 y-axis step sets were taken 600 mm apart from each other in the x-z plane starting 300 mm from the 0 line in each perpendicular axis in the x-z plane. This was the approximate centre of each of the 16 tiles seen in Figure 3 (right). The y steps were measured by placing the controller and jig on level aluminum blocks of specific heights, which were measured using zeroed, calibrated digital calipers accurate to 0.01 mm. The L-shaped jig was used as a physical origin point and axes and was aligned with the top left corners of the aluminum blocks (Figure 4).

    The aluminum blocks were chosen to allow the y-axis step sizes to approximately match the same x and z step sizes, and had heights of 6.23, 12.56, 50.82, and 152.5 mm. The x and z positions were monitored and recorded but not analyzed for alignment accuracy.

    Data Analysis

    To remove motion artifacts from pressing the buttons on the controller to start and stop the data recording, the first 1.5 seconds (135 samples) were removed and the next 2 seconds (180 samples) were used as the sample data for each measured point. After those 2 seconds, the rest of the data recording (approximately 1.5 seconds or 135 samples) was also discarded, regardless of length. The data were averaged to calculate the measured value at each point.

    For each sample point, the 3D position of the controller was measured. The variation was calculated for each data point and used to determine the static precision of the system. The position error for each point was calculated by subtracting the measured displacement (Euclidean distance) by the expected step size. The Euclidean distance was used to account for any misalignment of the Oculus tracking coordinate system with respect to the physical grid. The error values were then averaged to generate values for expected displacement error in a specific area of the play-space as well as for a specific step size over the entire play-space. The percent error was calculated to normalize the error to the step size.


    Results

    Positional Accuracy

    The average and percent errors for all step sizes in the x, y, and z directions are presented in Tables 1 and 2.

    The largest absolute error was found to be 3.5 mm in the z-axis for a step size of 500 mm, which normalizes to a 0.7% error. The largest normalized error was found to be 12.7% for the smallest step of 6.23 mm in the y-direction. The largest percent errors for the x and z axes were 4.7% and 3.5%, respectively, also at the smallest step size (5 mm).

    Table 1. Position error for different step sizes at all areas of the defined play-space measured with 0.01 mm accuracy using digital calipers. The percent error was calculated using nonrounded values of error in millimeter.
    View this table
    Table 2. Additional position error for larger step sizes at all areas of the defined play-space measured with ruler and graph paper markings. The percent error was calculated using nonrounded values of error in millimeter.
    View this table

    When the average percent error was calculated for each step size across the entire play-space, it was found that the error decreased nonlinearly with increasing step size (Figure 5).

    Step sizes with values of ≤10 mm had an accuracy error greater than 2%. Normalized error averages in step sizes larger than >10 mm were fairly uniform.

    The variation was calculated for static data points at the 16 different locations in the play-space and averaged across all points to find an average noise value of ±0.036 mm (x: ±0.025 mm, y: ±0.024 mm, z: ±0.055 mm). The single-point repeatability test found that the controller was able to return to the same x-y-z point with an output measurement variation slightly above the average noise value (x: 0.080 ± 0.546 mm, y: 0.088 ± 0.063 mm, z: 0.044 ± 0.326 mm).

    To investigate how accuracy varied over the entire play-space, the displacement and percent error were averaged over each area set. These average errors over the play-space area for x, y, and z are presented in Multimedia Appendix 1 as exploratory analysis to see if there were any patterns in accuracy based on distance away from the Oculus sensors. A large error (14.0%) occurred in 1 area of the y-axis measurements. This was a result of a 2 mm error at a 6.23 mm step, resulting in a large normalized value of approximately 40% despite being a small error in absolute distance (millimeters). No distinct pattern of position accuracy based on x-z location in the play-space was observed.

    Figure 5. Percent error for different step sizes. No values were calculated for 300 mm and 500 mm in the y-axis as the largest aluminum block used matched the 150 mm step size.
    View this figure

    Discussion

    Principal Findings

    The study team found a maximum positional accuracy across measured step sizes less than 150 mm for the Oculus Touch controller of 0.76 ± 0.50 mm in the lateral x-axis, 0.72 ± 0.46 mm in the anteroposterior z-axis, and 0.93 ± 1.10 mm in the vertical y-axis direction. Larger step sizes found lower positional accuracies of 2.5 ± 1.0 mm in the x-axis and 3.5 ± 2.5 mm in the z-axis. The largest error in percent when normalized to step size (12.7 ± 9.9%) was found in the smallest step size in the y-axis set at 6.23 mm.

    The error values found are considered within an acceptable range of error for the measurement of biomechanical movement as the human perception of the just noticeable difference (JND) even for fine motor function (finger distance/position) is larger than this value (13.0% for young subjects and 16.1% for older persons [15]). In addition, in a different study [16], it was reported that the JND of the fully extended human shoulder when moved passively was found to be 0.8º. Using a 50th percentile female arm length of 702 mm [14], a 0.8º change in joint angle would cause an arc length of 9.8 mm, which is larger than both the average noise and accuracy error of the system. Therefore, the error in relative distance should not be noticeable to the user.

    In studies on the accuracy of visual assessments of angular joint positions done by physical therapists and other health professionals, it was determined that joint positions could only be determined with an error of approximately 5º within the referenced radiometry measurement [17] of wrist angle and 7.4º in shoulder abduction compared with goniometry of nonmoving subjects [18]. Measurement of glenohumeral range of motion using a goniometer was reported to have an SE between 4.4 to 9.9º [19]. Moreover, the Oculus Rift headset has a visual field of view of 100º and a resolution of 2160x1200 pixels [20]. This results in a 0.046º change in the object edge to show up as a single pixel change. With a noise level of 0.069 mm as an arclength, it is expected that visual jitter would not occur until the controller is less than 8.59 mm away from the user in the headset’s point of view.

    It was found that although normalized percent error decreased as step size increased, the absolute error (millimeter) was found to be largest when the largest step was measured. This could be the effect of an inherent scaling phenomenon found in the Oculus sensor tracking, which uses infrared image processing as 1 of its main sources of position tracking for the Oculus Touch controllers.

    We expected that the outer edges of the defined play-space would provide areas with the largest error; however, no discernable pattern was found to occur over the x-z plane. This provides evidence to support a consistent accuracy of the controllers within the documented [21] x-z field of view of the Oculus Sensors. That is, regardless of where in the play-space a user might stand, a reach of 500 mm outwards from the body would be measured sufficiently accurately.

    Comparison With Prior Work

    On the basis of the results from this study, the Oculus Touch controllers should be an adequate motion tracking alternative for biomechanics applications, as similar low-cost, commercially available systems that have been employed to measure joint movements, such as the Kinect V1 and V2, have displacement accuracies on the order of centimeters [22,23]. A previous study evaluating the tracking accuracy of the Oculus Rift head-mounted display found similar accuracy values for the larger step sizes [24]. On the other hand, more expensive and complex motion capture systems such as the Vicon-460 (Vicon Motion Systems Ltd, Oxford, United Kingdom) with submillimetric accuracy [25] allow researchers to measure movements with higher accuracy. In addition, wearable inertial sensors were able to quantify upper-limb positioning within 1 mm when custom sensor algorithms were applied [7]; however, wearable research-grade systems trade ease of use and commercial availability for higher accuracy. For kinematic tracking that does not require submillimetric accuracy, such as for monitoring changes in gross motor upper-limb movements over time in a rehabilitation program [26], for training neural networks to detect the number of repetitions during rehabilitative exercises [9], or in cases where there would otherwise be no quantitative measures [8], the Oculus Touch controllers could provide a cost-effective alternative. Comparison with the Oculus Touch controllers by using other tracking tools simultaneously during upper-limb exercises may provide better insight into which level of accuracy is optimal for different use cases, such as for automated repetition counting as opposed to for measuring joint angles for digital goniometry. In cases where the tracking technology is used to facilitate virtual environments for engaging upper-limb exercises, a higher position accuracy may provide better visual fidelity to movement in the real world, and therefore, better transfer of improved motor functions from game tasks to real-world tasks [27].

    Limitations

    Occlusion of some of the controllers’ infrared light-emitting diodes could have occurred when placing the controllers close to the floor, which might have increased the measured error. Moreover, as the system requires an initial calibration of the user’s approximate height, this also could have acted as an additional source of error. As a result, future studies should investigate the accuracy of the system away from the floor, as well as the accuracy of the controller’s 3D orientation measurement, as we only measured position error in this study. Reproducibility of the measurements made by the Oculus system should also be investigated by having multiple experimenters perform the same procedure and by comparing measurements from different tracking sessions. Standardized measurement system analysis procedures should be followed in terms of the number of repetitions used as listed in analysis of variance gage repeatability and reproducibility documentation [28]. Dynamic conditions should also be evaluated before use in clinical kinematic analysis to assess the interaction between IMU sensor drift and camera sensor correction while in motion. A limitation of the Oculus Touch controllers is that it is only capable of measuring the position and orientation of a single point as a proxy for hand position. Future studies should directly compare the Oculus Touch absolute point-position with a professional marker-based motion-tracking system to ensure the elimination of error because of the use of visual and physical measuring tools. These absolute point-position studies should also evaluate larger step size accuracies to encompass bigger movements. Additional studies could include the evaluation of inverse kinematic algorithms that employ the hand and head positions (from headset) to generate a model of the user’s arms [29,30]. This would allow direct comparison against other devices that digitally measure goniometric angles.

    Acknowledgments

    The authors would like to acknowledge the support of the University of British Columbia’s Collaborative Advanced Robotics and Intelligent Systems and Robotics for Rehabilitation Exercise and Assessment in Collaborative Healthcare Lab members. Research reported in this publication was financially supported by the Kids Brain Health Network (TotTech).

    Conflicts of Interest

    None declared.

    Multimedia Appendix 1

    Exploratory results of accuracy based on x-z location: displacement in millimeters (left) and in percent (right) accuracy error for the Oculus Touch controller in the x (top), y (centre), and z (below) directions. Error is represented by the width of the circle for each step set area, however the circle size scale was magnified for visualization purposes and it is not to scale with the rest of the chart. Standard deviation is shown in parentheses.

    PNG File, 232KB

    References

    1. Valdés BA, Shirzad N, Hung CT, Glegg S, Reeds E, Van der Loos HFM. Visualisation of two-dimensional kinematic data from bimanual control of a commercial gaming system used in post-stroke rehabilitation. 2015 Presented at: International Conference on Virtual Rehabilitation; June 9-12, 2015; Valencia, Spain p. 243-250.
    2. Bonnechère B, Jansen B, van Sint Jan S. Cost-effective (gaming) motion and balance devices for functional assessment: need or hype? J Biomech 2016 Dec 6;49(13):2561-2565. [CrossRef] [Medline]
    3. Alankus G, Kelleher C. Reducing Compensatory Motions in Video Games for Stroke Rehabilitation. 2012 Presented at: ACM SIGCHI Conference on Human Factors in Computing Systems; May 5-10, 2012; Austin, Texas p. 2049-2058.
    4. Glegg SM, Hung CT, Valdés BA, Kim BD, van der Loos HF. Kinecting the moves: the kinematic potential of rehabilitation-specific gaming to inform treatment for hemiplegia. 2014 Presented at: International Conference on Disability, Virtual Reality and Associated Technologies; September 2-4, 2014; Gothenburg, Sweden.
    5. Sookhanaphibarn K, Phukongchai W, Santad T, Choensawat W. Towards Bilateral Upper-Limb Rehabilitation after Stroke using Kinect Game. 2018 Presented at: IEEE 7th Global Conference on Consumer Electronics (GCCE); October 15-18, 2018; Osaka, Japan p. 818-819. [CrossRef]
    6. Knorr B, Hughes R, Sherrill D, Stein J, Akay M, Bonato P. Quantitative Measures of Functional Upper Limb Movement in Persons after Stroke. 2005 Presented at: 2nd International IEEE EMBS Conference on Neural Engineering; March 16-19, 2005; Arlington, Virginia p. 252-255   URL: https://doi.org/10.1109/CNE.2005.1419604 [CrossRef]
    7. Bai L, Pepper MG, Yan Y, Spurgeon SK, Sakel M, Phillips M. Quantitative assessment of upper limb motion in neurorehabilitation utilizing inertial sensors. IEEE Trans Neural Syst Rehabil Eng 2015 Mar;23(2):232-243. [CrossRef]
    8. Kim WS, Cho S, Baek D, Bang H, Paik NJ. Upper extremity functional evaluation by Fugl-Meyer assessment scoring using depth-sensing camera in hemiplegic stroke patients. PLoS One 2016;11(7):e0158640 [FREE Full text] [CrossRef] [Medline]
    9. Lee MH, Siewiorek D, Smailagic A, Bernadino A, Badia SBI. A Kinect-based Monitoring System for Stroke Rehabilitation. 2017 Presented at: 5th International Congress on Sport Sciences Research and Technology Support (icSPORTS 2017); October 30-31, 2017; Funchal, Madeira, Portugal p. 8-10.
    10. Bigoni M, Baudo S, Cimolin V, Cau N, Galli M, Pianta L, et al. Does kinematics add meaningful information to clinical assessment in post-stroke upper limb rehabilitation? A case report. J Phys Ther Sci 2016;28(8):2408-2413 [FREE Full text] [CrossRef] [Medline]
    11. Shultz SJ, Houglum PA, Perrin DH. Examination of musculoskeletal injuries. In: Examination of Physiologic Range of Motion, Fourth Edition. Champaign, Illinois: Human Kinetics; 2016.
    12. Wilk KE, Reinold MM, Macrina LC, Porterfield R, Devine KM, Suarez K, et al. Glenohumeral internal rotation measurements differ depending on stabilization techniques. Sports Health 2009 Mar;1(2):131-136 [FREE Full text] [CrossRef] [Medline]
    13. Oculus VR, LLC. 2017. PC SDK Developer Guide: Oculus Touch Controllers - Controller Data   URL: https://developer.oculus.com/documentation/pcsdk/latest/concepts/dg-input-touch/ [accessed 2018-09-19] [WebCite Cache]
    14. Huston RL. Principles of Biomechanics. Boca Raton, Florida: CRC Press; 2009.
    15. Brewer BR, Fagan M, Klatzky RL, Matsuoka Y. Perceptual limits for a robotic rehabilitation environment using visual feedback distortion. IEEE Trans Neural Syst Rehabil Eng 2005 Mar;13(1):1-11. [CrossRef] [Medline]
    16. Tan HZ, Eberman B, Srinivasan MA, Cheng B. Human factors for the design of force-reflecting haptic interfaces. In: Dynamic Systems and Control. New York City, NY: ASME; 1994.
    17. McVeigh KH, Murray PM, Heckman MG, Rawal B, Peterson JJ. Accuracy and validity of goniometer and visual assessments of angular joint positions of the hand and wrist. J Hand Surg Am 2016 Apr;41(4):e21-e35. [CrossRef] [Medline]
    18. Banskota B, Lewis J, Hossain M, Irving A, Jones MW. Estimation of the accuracy of joint mobility assessment in a group of health professionals. Eur J Orthop Surg Traumatol 2008 Apr 9;18(4):287-289. [CrossRef]
    19. Fieseler G, Laudner KG, Irlenbusch L, Meyer H, Schulze S, Delank K, et al. Inter- and intrarater reliability of goniometry and hand held dynamometry for patients with subacromial impingement syndrome. J Exerc Rehabil 2017 Dec;13(6):704-710 [FREE Full text] [CrossRef] [Medline]
    20. Orland K. Ars Technica. 2016. iFixit digs into Oculus Rift's 461 PPI OLED display, custom lenses   URL: https:/​/arstechnica.​com/​gaming/​2016/​03/​ifixit-digs-into-oculus-rifts-461ppi-oled-display-custom-lenses/​ [accessed 2018-09-19] [WebCite Cache]
    21. Oculus Blog. 2017. Oculus Roomscale - Tips for Setting Up a Killer VR Room   URL: https://www.oculus.com/blog/oculus-roomscale-tips-for-setting-up-a-killer-vr-room/ [accessed 2018-09-20] [WebCite Cache]
    22. Otte K, Kayser B, Mansow-Model S, Verrel J, Paul F, Brandt AU, et al. Accuracy and reliability of the Kinect Version 2 for clinical measurement of motor function. PLoS One 2016;11(11):e0166532 [FREE Full text] [CrossRef] [Medline]
    23. Mobini A, Behzadipour S, Foumani MS. Accuracy of Kinect's skeleton tracking for upper body rehabilitation applications. Disabil Rehabil Assist Technol 2014 Jul;9(4):344-352. [CrossRef] [Medline]
    24. Borrego A, Latorre J, Alcañiz M, Llorens R. Comparison of Oculus Rift and HTC Vive: feasibility for virtual reality-based exploration, navigation, exergaming, and rehabilitation. Games Health J 2018 Jun;7(3):151-156 [FREE Full text] [CrossRef] [Medline]
    25. Windolf M, Götzen N, Morlock M. Systematic accuracy and precision analysis of video motion capturing systems - exemplified on the Vicon-460 system. J Biomech 2008 Aug 28;41(12):2776-2780. [CrossRef] [Medline]
    26. Held JPO, Klaassen B, Eenhoorn A, van Beijnum BF, Buurke JH, Veltink PH, et al. Inertial sensor measurements of upper-limb kinematics in stroke patients in clinic and home environment. Front Bioeng Biotechnol 2018 Apr;6:27 [FREE Full text] [CrossRef] [Medline]
    27. Bertrand J, Brickler D, Babu S, Madathil K, Zelaya M, Wang T, et al. The role of dimensional symmetry on bimanual psychomotor skills education in immersive virtual environments. 2015 Presented at: IEEE Virtual Reality; March 23-27, 2015; Arles, Camargue, Provence, France p. 3-10. [CrossRef]
    28. Burdick RK, Borror CM, Montgomery DC. Design and analysis of gauge R&R studies: making decisions with confidence intervals in random and mixed ANOVA models. In: ASA-SIAM Series on Statistics and Applied Probability. Philadelphia, Pennsylvania and Alexandria, Virginia: American Statistical Association and Society for Industrial and Applied Mathematics; 2005.
    29. Fezzik T. Imperium News. 2017. VR Level Up - Inverse Kinematics   URL: https://imperium.news/vr-level-inverse-kinematics/ [accessed 2018-09-20] [WebCite Cache]
    30. Sixense Entertainment Inc. Kickstarter. 2014. SixenseVR SDK Update   URL: https:/​/www.​kickstarter.com/​projects/​89577853/​stem-system-the-best-way-to-interact-with-virtual/​posts/​823494 [accessed 2018-09-19] [WebCite Cache]


    Abbreviations

    3D: 3-dimensional
    IMU: inertial measurement unit
    JND: just noticeable difference


    Edited by G Eysenbach; submitted 21.09.18; peer-reviewed by S Berrouiguet, I Cikajlo; comments to author 02.03.19; revised version received 25.04.19; accepted 14.05.19; published 06.06.19

    ©Leia C Shum, Bulmaro A Valdés, HF Machiel Van der Loos. Originally published in JMIR Biomedical Engineering (http://biomedeng.jmir.org), 06.06.2019.

    This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Biomedical Engineering, is properly cited. The complete bibliographic information, a link to the original publication on http://biomedeng.jmir.org/, as well as this copyright and license information must be included.