Robust End-User-Driven Social Media Monitoring for Law Enforcement and Emergency Monitoring

Nowadays social media mining is broadly used in the security sector to support law enforcement and to increase response time in emergency situations. One approach to go beyond the manual inspection is to use text mining technologies to extract latent topics, analyze their geospatial distribution and to identify the sentiment from posts. Although widely used, this approach has proven to be technically difficult for end-users: the language used on social media platforms rapidly changes and the domain varies according to the use case. This paper presents a monitoring architecture that analyses streams from social media, combines different machine learning approaches and can be easily adapted and enriched by user knowledge without the need for complex tuning.The framework is modeled based on the requirements of two H2020-projects in the area of community policing and emergency response.


Introduction
It has been widely proven that in case of emergencies and crisis situations content published on social media platforms, such as Twitter, constitutes a valuable source of information. As stated in , Twitter has been used in various crisis scenarios to identify events, monitor event progress, summarize tweets to give a situational overview and classify them to extract relevant content. In this paper we propose a social media monitoring and analysis framework that provides support to the security sector and is based on the needs of two running H2020-projects in the area of community policing and emergency response, CITYCoP and E2MC.
CITYCoP, short for Citizen Interaction Technologies Yield Community Policing, aims at developing applications that facilitate community policing. One idea behind the project is to mine social media platforms to assess the perceived quality of life of the citizens.
The purpose of the E2MC project, short for Evolution of Emergency Copernicus services, is to extend the available Copernicus Emergency Management System (EMS) that provides an early warning and rapid mapping services in case of emergencies. By integrating an analysis platform capable of extracting crisis-related information from social media, the quality of the delivered information should be improved and the speed in which the affected area can be located and damage can be assessed should be increased. The main problem we face in both use cases is to flexibly adapt to new scenarios and provide an analysis pipeline that can be operated by a user without expert knowledge in data analytics. In order to achieve this, we propose a framework that combines different machine learning approaches and can be enriched by user knowledge about the targeted event or location.

Related Work
A variety of methods have been applied to social media in the past in order to classify crisis-related tweets and to obtain useful information. One simple method is the keyword matching approach. In a predefined keyword dictionary, relevant words are collected. According to their appearance in a message text tweets can be extracted or classified (Olteanu et al. 2014). In more sophisticated approaches machine learning techniques are applied to develop models that are capable of automatically classifying tweets. These can be separated into supervised and unsupervised methods. Supervised methods learn models based on an already labeled dataset and can then be applied to incoming tweets. Caragea et al. (2016) and Nguyen et al. (2016) use online deep learning to classify tweets as informative or not informative. Nguyen et al. (2016) additionally partition crisis-related tweets into information-specific subclasses. Most of the frameworks proposed and developed in the context of event-and crisis-analysis, such as Tweedr (Ashktorab et al. 2014) or AIDR (Imran et al. 2014), are built upon these supervised techniques which rely on labeled datasets. We decided to integrate an unsupervised learning method that is able to train models without the need for a labeled set to overcome this drawback. To the best of our knowledge, there are only few examples in literature that follow this approach.  use an unsupervised topic modeling technique, LDA, to generate candidate categories for tweet classifications. Kireyev et al. (2009) explore, how topics models can be used to analyze disaster-related Tweets. One reason why these techniques are not yet widely spread is that they uncover hidden structures that are not necessarily of interest for a user. Wang et al. (2016) present an extension of LDA, the Targeted Topic Model, that is capable of building topics around an aspect of interest. Based on this algorithm we propose an end-user driven approach to analyze tweets with respect to the location, relevant hashtags or sentiment.

System Overview
Our system contains the following components to process crawled Tweets: The text processing component applies typical text preprocessing techniques, such as sentence splitting, tokenization and stemming, to prepare the tweets for further analysis. The data is then enriched by integrating Hashtags, URL categories and handles as additional features. Furthermore, a user can define an event-profile with a georeference and the time the event started. Based on the profile, we calculate the distance of the tweets geolocation to the area of interest and the time elapsed since the event started. The sentiment analysis component identifies the polarity of texts in terms of whether the texts are emotionally positive, negative or neutral. This so called sentiment is added as an additional feature for the topic extraction component. The topic extraction component is the core step of our approach and will be described in detail in section "End-User-Driven Topic Detection". The visualization component is used to display enriched tweet data and enables the user to monitor tweets based on their location, derived sentiment and assigned topics. Tweets can be clustered based on their density and according to the assigned topics or keywords contained. Additionally, interactive filters based on keywords and topics can be applied to support specific scenarios.

End-User-Driven Topic Detection
Most text analytic models rely on a set of parameters or constraints that have to be set in advance of the model training. Since these parameters can heavily influence the model performance, this training process often requires complex tuning and expert data science knowledge. To allow a user to influence the analysis pipeline in a more natural way, we focus on an extension of the broadly used topic modeling algorithm LDA (Blei et al. 2003), the targeted topic model (Wang et al. 2016), that is capable of generating flexible models and integrating user domain knowledge about the event of interest.
LDA assumes that each document can be described as a mixture over topics, where each topic represents a distribution over words. A major problem one faces when applying LDA to analyze tweets is, that topics of interest may not be detected and the topics reflecting the dataset are to coarse. The idea of targeted topic modeling is to focus on an aspect the user is interested in by setting target keywords in advance. Based on this target a topic model is generated. This model consists of one major topic reflecting the content not related to the aspect and multiple finegrained topics related to the aspect. The topics are represented as word distributions. By defining a set of target keywords related to an event of interest, these topic models can be used to extract tweets related to an event and to uncover useful structures to cluster these tweets. Figure 4.1 shows the analysis process compared to the baseline LDA-approach and a keyword based approach. In the keywordbased approach, a keyword list is used to filter relevant tweets while in the LDA approach, keyword lists are used to identify relevant topics by comparing them to the topic-word-distributions. In both scenarios, the selection of this keywords plays a crucial rule and heavily affects the quality of the retrieved content. The advantage of the targeted topic model is, that it automatically extends the keywords assumed to be relevant based on the input keyword list and the word co-occurrences in the tweets. This allows a more intuitive keyword selection without the need for expert refinement. To cover a rich spectrum of relevant information, we extend the model and enrich the input data with additional keywords representing different forms of information, such as the sentiment, the location and other meta data. We chose keywords as the way the information is exchanged, since it is easy to inspect and understand for the end-user and can be extended to other relevant types of information.

Experiments and Results
To evaluate our methods, we perform an experimental analysis of tweet data related to three crisis-events, the flood in Alberta in 2013, the Bombings in Boston during and after the marathon in April 2013 and the explosion at the West Fertilizer Company storage and distribution facility in April 2013. Therefore, we merged two publicly available datasets, CRISISLEX26 (Olteanu et al. 2015) and CRISISLEX6 (Olteanu et al. 2014) that were labeled from crowd-source workers as "on-topic" and "off-topic" and are assigned to one of six categories, e.g., "Infrastructure and utilities" and "Donations and volunteering". The label information assigned to each tweet is only used to evaluate the performance of the trained model. The analysis itself is performed on the unlabeled dataset. After preparation, we generated three datasets with around 10,000 tweets each. To evaluate our analysis approach on these dataset, we simulated an interactive analysis workflow that consists of three steps: The identification of event-related tweets, the identification and analysis of topic structures and the visual analysis.
Identification of event-related tweets: based on an input keyword set a targeted topic model is trained and applied to classify tweets as relevant or not relevant. To examine the robustness of the model, we pick very general and simple target  Bowriver, downtown, emerg, calgari, bridg, update, medicinhat, state, local, remain, citi, power, rcmp, water, school, close, canmor, break, expect, elbow Donations: Relief, donat, love, redcross, great, text, peopl, victim, find, made, good, feel, affect, bad, provid, send, pretti, assist, custom, fan keywords for model training and the keyword matching. For the Alberta Flood the target keywords are set to all words matching the substring "flood", for the explosion in Texas we pick all words matching the substring "explo" and the substrings to identify the Boston event are "bomb" and "marathon". To assess the performance of the model, the results are compared to two different baseline-approaches: • a keyword matching based approach, in which tweets are filtered as relevant when they contain a word in the specified keyword set. • the direct application of LDA, where each tweet assigned to a topic identified as crisis-related is treated as relevant. We assumed all topics with high word probability for the words listed in the defined keyword set to be crisis-related. Table 4.1 shows the performance for each approach. In all three datasets the targeted topic model outperforms the baseline LDA regarding F1 score. In the Alberta dataset, the identification seems to be not that challenging and the use of the targeted topic model approach cannot improve the performance. One reason could be that most of the tweets contain flood-related hashtags, such as #albertaflood or #abflood.

Identification and analysis of topic structures:
In the next phase we further analyze the learned topic-word-distribution. In the experiment, we focused on the Alberta dataset and chose a model that consists of 25 topics describing the different hidden structures in the event-related tweets and one major topic assigned to all tweets identified as not event-related. In order to focus on a specific aspect of interest, we concentrate on the topics best matching manually defined keywords, in this case "blocked / closed / damaged" for damage related topics and "donate / fund / relief" for donation related topics. The results are presented in Table 4.2. The damage-focused topics not only contain words describing infrastructure, such as road, school or bridge, but are also capable of uncovering affected rivers, e.g. the Elbow River or affected towns, such as Medicine Hat, and damaged areas, such as the Bowness Park. Next, the meta features assigned to the topics are examined. Table 4.3 lists the handles, locations, hashtags, time and URLs related to damage-or donation-topics. While in the damage-related topics the fact, that a tweet position is close to the event, is very relevant, in the donation topic this is not that important. The detected URLs as well as the handles and hashtags for the donation-topic mainly focus on institutions offering help, such as the Red Cross and politicians providing support. In the damage-related topics the focus is set to news agencies, the police and the affected towns.
Visual Analysis: To obtain a better insight, we visually examine the trained topic model. Therefore, a sentiment-map is generated, showing all tweets near Alberta assigned to the damage-or the donation-topic colored with the sentiment (Fig. 4.2). In the area very close to the affected area, the sentiment mainly seems to be negative in both topics. In the surrounding areas, there are more positive tweets contained in the donation topic. Next, the density based clusters for the tweets related to damage and the tweets related to donation are generated and analyzed to uncover topic hot spots. As we infer the user's location if the tweet location is not available, this clusters are influenced by the fact that all user tweets from a city are mapped to the same spot retrieved by the OpenStreetMap API. Nonetheless, this clustering can be an indicator for topic hot spots. By examining the hot spots for damage-related topics in the affected area, we could find useful information about the situation on side regarding power supply, bridge closures and pictures of flooded areas.

Conclusions
In this paper, we propose a robust framework that monitors and analyzes streams from social media and is adaptable by the end-user without the need for complex tuning. The system is based on state-of-the-art text mining technologies, including a convolutional neural network for sentiment detection and an extension of the Latent Dirichlet Allocation functionality to identify latent topics from tweets. Identified topics or relevant keywords are visualized on a map together with their respective sentiment. We processed the text and enriched it with several meta features, such as the location to be able to extract different information types during the analysis and set the analysis focus to different aspects.
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.
The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.