Predicting Destinations from Partial Trajectories Using Recurrent Neural Network

  • Yuki EndoEmail author
  • Kyosuke Nishida
  • Hiroyuki Toda
  • Hiroshi Sawada
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 10234)


Predicting a user’s destinations from his or her partial movement trajectories is still a challenging problem. To this end, we employ recurrent neural networks (RNNs), which can consider long-term dependencies and avoid a data sparsity problem. This is because the RNNs store statistical weights for long-term transitions in location sequences unlike conventional Markov process-based methods that count the number of short-term transitions. However, how to apply the RNNs to the destination prediction is not straight-forward, and thus we propose an efficient and accurate method for this problem. Specifically, our method represents trajectories as discretized features in a grid space and feeds sequences of them to the RNN model, which estimates the transition probabilities in the next timestep. Using these one-step transition probabilities, the visiting probabilities for the destination candidates are efficiently estimated by simulating the movements of objects based on stochastic sampling with an RNN encoder-decoder framework. We evaluate the proposed method on two different real datasets, i.e., taxi and personal trajectories. The results demonstrate that our method can predict destinations more accurately than state-of-the-art methods.


Recurrent Neural Network Sampling Simulation Trajectory Data Historical Trajectory True Destination 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


  1. 1.
    de Brébisson, A., Simon, É., Auvolat, A., Vincent, P., Bengio, Y.: Artificial neural networks applied to taxi destination prediction (2015). CoRR, abs/1508.00021Google Scholar
  2. 2.
    Elkahky, A.M., Song, Y., He, X.: A multi-view deep learning approach for cross domain user modeling in recommendation systems. In: WWW, pp. 278–288 (2015)Google Scholar
  3. 3.
    Gers, F., Schmidhuber, J., Cummins, F.: Learning to forget: continual prediction with LSTM. In: ICANN (2), pp. 850–855 (1999)Google Scholar
  4. 4.
    Graves, A.: Generating sequences with recurrent neural networks (2013). CoRR, abs/1308.0850Google Scholar
  5. 5.
    Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRefGoogle Scholar
  6. 6.
    Horvitz, E., Krumm, J.: Some help on the way: opportunistic routing under uncertainty. In: UbiComp, pp. 371–380 (2012)Google Scholar
  7. 7.
    Krumm, J., Horvitz, E.: Predestination: inferring destinations from partial trajectories. In: Dourish, P., Friday, A. (eds.) UbiComp 2006. LNCS, vol. 4206, pp. 243–260. Springer, Heidelberg (2006). doi: 10.1007/11853565_15 CrossRefGoogle Scholar
  8. 8.
    Krumm, J., Horvitz, E.: Predestination: where do you want to go today? Computer 40(4), 105–107 (2007)CrossRefGoogle Scholar
  9. 9.
    Liu, Q., Wu, S., Wang, L., Tan, T.: Predicting the next location: a recurrent model with spatial and temporal contexts. In: AAAI, pp. 194–200 (2016)Google Scholar
  10. 10.
    Moreira-Matias, L., Gama, J., Ferreira, M., Mendes-Moreira, J., Damas, L.: Predicting taxi-passenger demand using streaming data. IEEE Trans. Intell. Transp. Syst. 14(3), 1393–1402 (2013)CrossRefGoogle Scholar
  11. 11.
    Pascanu, R., Mikolov, T., Bengio, Y.: On the difficulty of training recurrent neural networks. ICML 28, 1310–1318 (2013)Google Scholar
  12. 12.
    Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations by back-propagating errors. In: Neurocomputing: Foundations of Research, pp. 696–699 (1988)Google Scholar
  13. 13.
    Xue, A.Y., Qi, J., Xie, X., Zhang, R., Huang, J., Li, Y.: Solving the data sparsity problem in destination prediction. VLDB J. 24(2), 219–243 (2015)CrossRefGoogle Scholar
  14. 14.
    Xue, A.Y., Zhang, R., Zheng, Y., Xie, X., Huang, J., Xu, Z.: Destination prediction by sub-trajectory synthesis and privacy protection against such prediction. In: ICDE, pp. 254–265 (2013)Google Scholar
  15. 15.
    Zeiler, M.D.: ADADELTA: an adaptive learning rate method (2012). CoRR, abs/1212.5701Google Scholar
  16. 16.
    Zheng, Y., Li, Q., Chen, Y., Xie, X., Ma, W.: Understanding mobility based on GPS data. In: UbiComp 2008, pp. 312–321 (2008)Google Scholar
  17. 17.
    Zheng, Y., Xie, X., Ma, W.: GeoLife: a collaborative social networking service among user, location and trajectory. IEEE Data Eng. Bull. 33(2), 32–39 (2010)Google Scholar
  18. 18.
    Zheng, Y., Zhang, L., Xie, X., Ma, W.: Mining interesting locations and travel sequences from GPS trajectories. In: WWW, pp. 791–800 (2009)Google Scholar
  19. 19.
    Ziebart, B.D., Maas, A.L., Dey, A.K., Bagnell, J.A.: Navigate like a cabbie: probabilistic reasoning from observed context-aware behavior. In: UbiComp, pp. 322–331 (2008)Google Scholar
  20. 20.
    Zipser, D.: Subgrouping reduces complexity and speeds up learning in recurrent networks. In: Advances in Neural Information Processing Systems 2, pp. 638–641. Morgan Kaufmann Publishers Inc. (1990)Google Scholar

Copyright information

© Springer International Publishing AG 2017

Authors and Affiliations

  • Yuki Endo
    • 1
    Email author
  • Kyosuke Nishida
    • 1
  • Hiroyuki Toda
    • 1
  • Hiroshi Sawada
    • 1
  1. 1.NTT Service Evolution LaboratoriesNTT CorporationYokosuka-shiJapan

Personalised recommendations