3D Indoor Location on Mobile Phones Using Embedded Sensors and Close-Range Photogrammetry

  • Xiujuan Li
  • Yan Zhou
  • Hanjiang XiongEmail author
Part of the Lecture Notes in Geoinformation and Cartography book series (LNGC)


Indoor positioning on mobile phones has become more and more important to many applications. This paper presents a sensor-based 3D positioning system on mobile phones. It also proposes an efficient approach to determine the initial position which has a great influence on the position precision. By taking a photo of the pre-deployed cross board, the camera’s position can be identified immediately using the principle of single-photo resection. The proposed positioning scheme performs location estimation in three phases. First, use cross boards to precisely determine the initial position. Second, employ accelerometer, magnetometer embedded in mobile phones and INS to track the user’s position automatically. Finally, visualize the user’s location in the 3D model of the building based on GPU rendering technology. The experiment carried out in this paper has good results and indicates that the new method with the embedded sensors and close-range photogrammetry is a promising solution for indoor location.


3D indoor visualization Mobile devices 3D indoor location LBS Close-range photogrammetry Inertial navigation Sensors 



This work was supported by a grant from the National High Technology Research and Development Program of China (863 program) (No. 2012BAH35B03 and No. 2013AA12A203) and a grant from Shenzhen Science and Technology Development Program.


  1. Akenine-Moller T, Strom J (2008) Graphics processing units for handhelds. Proc IEEE 96(5):779–789CrossRefGoogle Scholar
  2. Bahl P, Padmanabhan VN (2000) RADAR: an in-building RF-based user location and tracking system. In: INFOCOM 2000. Nineteenth annual joint conference of the IEEE computer and communications societies. Proceedings IEEE vol 2, pp 775–784Google Scholar
  3. Bajaj R, Ranaweera SL, Agrawal DP (2002) GPS: location-tracking technology. Computer 35(4):92–94CrossRefGoogle Scholar
  4. Biber P, Andreasson H, Duckett T, Schilling A (2004) 3D modeling of indoor environments by a mobile robot with a laser scanner and panoramic camera. In: International conference on intelligent robots and systems, 2004. (IROS 2004). Proceedings 2004 IEEE/RSJ (vol 4, pp 3430–3435)Google Scholar
  5. Daniels J, Ha LK, Ochotta T, Silva CT (2007) Robust smooth feature extraction from point clouds. In: IEEE international conference on shape modeling and applications, 2007. SMI’07. (pp 123–136)Google Scholar
  6. González E, Prados L, Rubio AJ, Segura JC, de la Torre Á, Moya JM, Martín JL (2009) ATLINTIDA: a robust indoor ultrasound location system: design and evaluation. In: 3rd symposium of ubiquitous computing and ambient intelligence 2008 (pp 180–190). Springer, BerlinGoogle Scholar
  7. Kalman RE (1960) A new approach to linear filtering and prediction problems. J Fluids Eng 82(1):35–45Google Scholar
  8. Kuo YC, Hsiao MY, Wen CY (2013) An integrated mobile sensor platform for collaborative indoor self-positioning applications. In: TENCON spring conference, 2013 IEEE (pp 495–499)Google Scholar
  9. Prasithsangaree P, Krishnamurthy P, Chrysanthis PK (2002). On indoor position location with wireless LANs. In: The 13th IEEE international symposium on personal, indoor and mobile radio communications, 2002. (vol 2, pp 720–724)Google Scholar
  10. Smith SM, Brady JM (1997) SUSAN—a new approach to low level image processing. Int J Comput Vision 23(1):45–78CrossRefGoogle Scholar
  11. Welch G, Bishop G (2000) An introduction to the Kalman filter, from, UNC-Chapel Hill, TR95-041, November
  12. Tommaselli AMG, Reiss MLL (2005) A photogrammetric method for single image orientation and measurement. Photogram Eng Remote Sens 71(6):727–732CrossRefGoogle Scholar
  13. Winter S, Kealy A (2012) An alternative view of positioning observations from low cost sensors. Comput Environ Urban Syst 36(2):109–117CrossRefGoogle Scholar
  14. Woodman OJ (2007) An introduction to inertial navigation. University of Cambridge, Comput Lab Tech Rep UCAMCL-TR-696 14, 15Google Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Authors and Affiliations

  1. 1.State Key Laboratory of Information Engineering in Surveying, Mapping and Remote SensingWuhan UniversityWuhanChina

Personalised recommendations