Abstract
Multi-participant chat conversations are one of the most frequently employed Computer Supported Collaborative Learning tools due to their ease of use. Moreover, chats enhance knowledge sharing, sustain creativity and aid in collaborative problem solving. Nevertheless, the manual analysis of multi-participant chats is a difficult task due to the mixture of different topics and the inter-twinning of multiple discussion threads during the same conversation. Several tools that employ Natural Language Processing techniques have been developed to automatically identify links between contributions in order to facilitate the tracking of topics and of discussion threads, as well as to highlight key contributions in terms of follow-up impact. This paper proposes a novel method for detecting implicit links based on features computed using string kernels and word embeddings, combined with neural networks. This method significantly outperforms previous results on the same dataset. Due to its smaller size, our model represents an alternative to more complex deep neural networks, especially when limited training data is available as is the case of CSCL chats in a specific domain.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Stahl, G.: Group Cognition: Computer Support for Building Collaborative Knowledge. MIT Press, Cambridge, MA (2006)
Trausan-Matu, S.: Computer support for creativity in small groups using chats. Annal. Acad. Rom. Sci. Ser. Sci. Technol. Inf. 3(2), 8190 (2010)
Holmer, T., Kienle, A., Wessner, M.: Explicit referencing in learning chats: needs and acceptance. In: Nejdl, W., Tochtermann, K. (eds.) EC-TEL 2006. LNCS, vol. 4227, pp. 170–184. Springer, Heidelberg (2006). https://doi.org/10.1007/11876663_15
Manning, C.D., Schtze, H.: Foundations of Statistical Natural Language Processing. MIT Press, Cambridge, MA (1999)
Masala, M., Ruseti, S., Rebedea, T.: Sentence selection with neural networks using string kernels. Procedia Comput. Sci. 112, 1774–1782 (2017)
Gutu, G., Dascalu, M., Rebedea, T., Trausan-Matu, S.: Time and semantic similarity what is the best alternative to capture implicit links in CSCL conversations? In: 12th International Conference on Computer-Supported Collaborative Learning (CSCL 2017), pp. 223–230. ISLS (2017)
Gutu, G., Dascalu, M., Ruseti, S., Rebedea, T., Trausan-Matu, S.: Unlocking the power of word2vec for identifying implicit links. In: 17th IEEE International Conference on Advanced Learning Technologies (ICALT 2017), p. 199200. IEEE (2017)
Trausan-Matu, S., Rebedea, T.: A polyphonic model and system for inter-animation analysis in chat conversations with multiple participants. In: Gelbukh, A. (ed.) CICLing 2010. LNCS, vol. 6008, pp. 354–363. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-12116-6_29
Elsner, M., Charniak, E.: Disentangling chat. Comput. Linguist. 36(3), 389–409 (2010)
Miller, G.A.: Wordnet: a lexical database for English. Commun. ACM 38(11), 39–41 (1995)
Landauer, T.K., Dumais, S.T.: A solution to plato’s problem: the latent semantic analysis theory of acquisition, induction and representation of knowledge. Psychol. Rev. 104(2), 211240 (1997)
Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent dirichlet allocation. J. Mach. Learn. Res. 3(4–5), 9931022 (2003)
Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representation in vector space. In: Workshop at ICLR (2013)
Mehri, S., Carenini, G.: Chat disentanglement: Identifying semantic reply relationships with random forests and recurrent neural networks. In: Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers). vol. 1, pp. 615–623 (2017)
Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems, pp. 3111–3119 (2013)
Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: Empirical Methods in Natural Language Processing (EMNLP), pp. 1532–1543 (2014). http://www.aclweb.org/anthology/D14-1162
Bojanowski, P., Grave, E., Joulin, A., Mikolov, T.: Enriching word vectors with subword information. arXiv preprint arXiv:1607.04606 (2016)
Lodhi, H., Saunders, C., Shawe-Taylor, J., Cristianini, N., Watkins, C.: Text classification using string kernels. J. Mach. Learn. Res. 2(Feb), 419–444 (2002)
Ionescu, R.T., Popescu, M., Cahill, A.: Can characters reveal your native language? a language-independent approach to native language identification. In: EMNLP, pp. 1363–1373 (2014)
Ionescu, R.T., Popescu, M., Cahill, A.: String kernels for native language identification: insights from behind the curtains. Comput. Linguist. 42, 491–525 (2016)
Gönen, M., Alpaydın, E.: Multiple kernel learning algorithms. J. Mach. Learn. Res. 12(Jul), 2211–2268 (2011)
Beck, D., Cohn, T.: Learning kernels over strings using Gaussian processes. In: Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 2: Short Papers), vol. 2, pp. 67–73 (2017)
Bachrach, Y., Zukov-Gregoric, A., Coope, S., Tovell, E., Maksak, B., McMurtie, C.: An attention mechanism for answer selection using a combined global and local view. arXiv preprint arXiv:1707.01378 (2017)
dos Santos, C.N., Tan, M., Xiang, B., Zhou, B.: Attentive pooling networks. CoRR, abs/1602.03609 (2016)
Tan, M., dos Santos, C.N., Xiang, B., Zhou, B.: Improved representation learning for question answer matching. In: ACL, vol. 1 (2016)
Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM networks. In: Proceedings of 2005 IEEE International Joint Conference on Neural Networks, IJCNN 2005, vol. 4, pp. 2047–2052. IEEE (2005)
Kim, Y.: Convolutional neural networks for sentence classification. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP 2014), pp. 1746–1751, August 2014
Wang, S., Jiang, J.: A compare-aggregate model for matching text sequences. arXiv preprint arXiv:1611.01747 (2016)
Stahl, G.: Studying Virtual Math Teams. Springer, New York, NY (2009). https://doi.org/10.1007/978-1-4419-0228-3
Kingma, D.P., Adam, B.J.: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Hu, B., Lu, Z., Li, H., Chen, Q.: Convolutional neural network architectures for matching natural language sentences. In: Advances in neural information processing systems, pp. 2042–2050 (2014)
Masala, M., Ruseti, S., Gutu-Robu, G., Rebedea, T., Dascalu, M., Trausan-Matu, S.: Identifying implicit links in CSCL chats using string kernels and neural networks. In: Penstein Rosé, C., MartÃnez-Maldonado, R., Hoppe, H.U., Luckin, R., Mavrikis, M., Porayska-Pomsta, K., McLaren, B., du Boulay, B. (eds.) AIED 2018, Part II. LNCS (LNAI), vol. 10948, pp. 204–208. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-93846-2_37
Acknowledgements
This research was partially supported by the FP7 2008-212578 LTfLL, EC H2020-644187 Realising an Applied Gaming Eco-system (RAGE), and POC-2015 P39-287 IAVPLN projects.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Masala, M., Ruseti, S., Gutu-Robu, G., Rebedea, T., Dascalu, M., Trausan-Matu, S. (2018). Help Me Understand This Conversation: Methods of Identifying Implicit Links Between CSCL Contributions. In: Pammer-Schindler, V., Pérez-SanagustÃn, M., Drachsler, H., Elferink, R., Scheffel, M. (eds) Lifelong Technology-Enhanced Learning. EC-TEL 2018. Lecture Notes in Computer Science(), vol 11082. Springer, Cham. https://doi.org/10.1007/978-3-319-98572-5_37
Download citation
DOI: https://doi.org/10.1007/978-3-319-98572-5_37
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-98571-8
Online ISBN: 978-3-319-98572-5
eBook Packages: Computer ScienceComputer Science (R0)