AVR 2015: Augmented and Virtual Reality pp 188-196 | Cite as
IMU Drift Reduction for Augmented Reality Applications
Abstract
Pose estimation is a major task in any augmented reality (AR) application. Sensors like inertial measurement units (IMUs) and cameras are used for this purpose. IMU provides fast data, from which a pose can be determined with fewer calculations. In most cases, this data is influenced by drift. This paper investigates a drift reduction technique on positions calculated from accelerometer data and the result shows that this technique significantly reduces drift in fast dynamic movements. Furthermore, this drift reduced IMU pose can be fused with low pose updates from vision algorithms to improve the accuracy and speed suitable for AR applications.
Keywords
Augmented reality Pose estimation Inertial measurement unit Marker tracking Sensor fusionNotes
Acknowledgements
The authors wish to thank all other members of the EDUSAFE consortium. This research has been supported by a Marie Curie Initial Training Network Fellowship of the European Commission FP7 Programme under contract number PITN-GA-2012-316919-EDUSAFE.
References
- 1.Henderson, S., Feiner, S.: Exploring the benefits of augmented reality documentation for maintenance and repair. IEEE Trans. Visual Comput. Graphics 17(10), 1355–1368 (2011)CrossRefGoogle Scholar
- 2.Newman, J., Wagner, M. et al.: Ubiquitous tracking for augmented reality. In: Third IEEE and ACM International Symposium on Mixed and Augmented Reality, pp. 192–201 (2004)Google Scholar
- 3.Euston, M., Coote, P. et al.: A complementary filter for attitude estimation of a fixed-wing UAV. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (2008)Google Scholar
- 4.Hol, J.D., Schon, T.B. et al.: Sensor fusion for augmented reality. In: 9th International Conference on Information Fusion, pp. 1–6 (2006)Google Scholar
- 5.Tsai, R., Lenz, R.: A new technique for fully autonomous and efficient 3D robotics hand/eye calibration. IEEE Trans. Robot. Autom. 5(3), 345–358 (1989)CrossRefGoogle Scholar
- 6.Hol, J.D., Schon, T.B., Gustafsson, F.: Relative pose calibration of a spherical camera and an IMU. In: International Symposium on Mixed and Augmented Reality, pp. 21–24 (2008)Google Scholar
- 7.Lobo, J., Dias, J.: Relative pose calibration between visual and inertial sensors. Int. J. Robot. Res. 26(6), 561–575 (2007)CrossRefGoogle Scholar
- 8.Furgale, P., Rehder, J., Siegwart, R.: Unified Temporal and Spatial Calibration for Multi-Sensor Systems. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (2013)Google Scholar
- 9.Slifka, L.D.: An accelerometer based approach to measuring displacement of a vehicle body. Master Thesis, University of Michigan, Dearborn (2004)Google Scholar
- 10.Bishop, G. Welch, G.: An introduction to the Kalman filter. In: Proceedings of SIGGRAPH, Course 8 (2001)Google Scholar
- 11.Kato, H., Billinghurst, M.: Marker tracking and hmd calibration for a video-based augmented reality conferencing system. In: 2nd International Workshop on Augmented Reality (IWAR 1999), pp. 85–94 (1999)Google Scholar
- 12.Looser, J., Grasset, R., Seichter, H., Billinghurst, M.: OSGART - a pragmatic approach to MR. In: International Symposium of Mixed and Augmented Reality (2006)Google Scholar
- 13.Bauer, M., Bruegge, B., Klinker, G., MacWilliams, A., Reicher, T., Riss, S., Sandor, C., Wagner, M.: Design of a component-based augmented reality framework. In: International Symposium on Augmented Reality (ISAR) (2001)Google Scholar
- 14.Fiala, M.: ARTag Revision 1, A fiducial marker system using digital Techniques. NRC Technical Report (NRC 47419), National Research Council of Canada (2004)Google Scholar
- 15.Salvi, J., Armangu, X., Batlle, J.: A comparative review of camera calibrating methods with accuracy evaluation. Pattern Recogn. 35(7), 1617–1635 (2002)MATHCrossRefGoogle Scholar