Embed2Detect: temporally clustered embedded words for event detection in social media

Social media is becoming a primary medium to discuss what is happening around the world. Therefore, the data generated by social media platforms contain rich information which describes the ongoing events. Further, the timeliness associated with these data is capable of facilitating immediate insights. However, considering the dynamic nature and high volume of data production in social media data streams, it is impractical to filter the events manually and therefore, automated event detection mechanisms are invaluable to the community. Apart from a few notable exceptions, most previous research on automated event detection have focused only on statistical and syntactical features in data and lacked the involvement of underlying semantics which are important for effective information retrieval from text since they represent the connections between words and their meanings. In this paper, we propose a novel method termed Embed2Detect for event detection in social media by combining the characteristics in word embeddings and hierarchical agglomerative clustering. The adoption of word embeddings gives Embed2Detect the capability to incorporate powerful semantical features into event detection and overcome a major limitation inherent in previous approaches. We experimented our method on two recent real social media data sets which represent the sports and political domain and also compared the results to several state-of-the-art methods. The obtained results show that Embed2Detect is capable of effective and efficient event detection and it outperforms the recent event detection methods. For the sports data set, Embed2Detect achieved 27% higher F-measure than the best-performed baseline and for the political data set, it was an increase of 29%.


Introduction
Social media services such as Twitter and Facebook are becoming increasingly popular.A recent survey by Chaffey (2019) estimated the number of active social media users around the world in January 2019 as 3.484 billion; 45% of the total population.The average of the global increase in social media usage since January 2018 was found to be 9%.Another analysis was conducted on active users on social media in July 2019 to rank social media services based on popularity (Clement 2019).According to its results, the majority of the services have millions of users with Facebook leading with a user base of 2375 million.Approximately 511,200 tweets per minute were recorded in 2018 (James 2019).
The data produced on social media contain different information such as opinions, breaking news and personal updates.Also, social media facilitates fast information dispersal because of its large user base which covers a vast geographical area (Castillo et al. 2011).In some cases, social media was found to broadcast news faster than traditional news media by an analysis which compared Twitter trending topics with CNN news headlines (Kwak et al. 2010).Due to the inclusion of diverse information and real-time propagation to large groups, nowadays, there is a high tendency to consider social media as information networks which provide newsworthy contents.In 2017, the proportion of American adults who got news from social media was found to be 67% (Gottfried and Shearer 2017).Consequently, news services such as BBC and CNN also use social media actively to instantly publish news to a huge user base.Nonetheless, it is impractical to analyse the data manually to extract important or newsworthy content from social media, because of its huge volume and dynamic nature.Therefore, in order to utilise the social media data effectively, the requirement of an automated and accurate event detection method is crucial (Small and Medsker 2014).
Considering this requirement, different methods were suggested by previous research for event detection in social media as further discussed in Sect. 2. Apart from a few notable exceptions, most of the methods were based on statistical and syntactical features in data and lacked the involvement of semantical features.A language is mainly built using two phenomena, namely, syntax and semantics (Sag and Pollard 1987).Syntax defines the arrangement of words in word sequences, and semantics describes the connections between words and their meanings.Using both syntax and semantics, languages allow different word sequences to express the same idea.This impact is widely demonstrated in the social media text, due to the diversity in users.For example, consider the tweets: There are 13 million people living in poverty in the UK.13M!!! Yet some MPs will vote for the deal with NO impact assessments.That 13M could become 20M?!#VoteTheDealDown #PeoplesVoteMarch #PeoplesVote #StopBrexit Luciana Berger -Steve Barclay confirmed that no economic analysis of the #Brexit-Deal has been done... let that sink in.So how can we be expected to vote on a deal, that will affect this country for decades, today?#VoteDownTheDeal #PeoplesVote which were posted during the Brexit Super Saturday 2019.Even though both tweets describe the same idea, there are no common words between them except for a few hashtags.In addition, different word phrases such as impact assessments and economic analysis were used to mention the same subject discussed in them.In such cases, semantics are needed to understand the relationships between terms to extract valuable information.
Focusing the importance of semantics to natural language processing (NLP), word embeddings such as Word2Vec (Mikolov et al. 2013a) with high capability in preserving syntactic and semantic relationships between words were introduced.These embeddings were successfully used within many NLP related tasks such as news recommendation (Zhang et al. 2019), question classification (Yilmaz and Toklu 2020) and community detection (Škrlj et al. 2020) recently.Similarly, Word2Vec embeddings were used for an event detection approach named W E C too (Comito et al. 2019b).W E C is mainly based on pre-trained word embeddings and online document clustering.However, pre-trained models will be less effective in the context of social media due to the availability of modified or misspelt words.The inability to recognise such word can lead to a serious information loss.
Considering the lack of semantic involvement in previous methods, this research proposes a novel event detection method termed Embed2Detect which combines the characteristics of word embeddings and hierarchical agglomerative clustering.Rather than using pre-trained word embeddings, we propose to use self-learned word embeddings which can capture the characteristics specific to the targeted corpus.Also, without relying on direct clusters, we consider the temporal variations between clusters and vocabularies using a time-based sliding window model to identify event occurrences.We targeted a token-based clustering approach to successfully handle the scenarios where a single document contains multiple event details.In addition to considering the underlying syntax and semantics, event detection also requires the incorporation of statistical features in the text, to measure the qualities of events such as popularity.In our method, this requirement also fulfilled by using self-learned word embeddings and vocabulary change measures, which capture statistics.In summary, Embed2Detect is an improved method, which considers all the important features in textual data; syntax, semantics and statistics, which are needed for effective event detection.
To evaluate the proposed method, two recent social media data sets which represent two diverse domains; sports (English Premier League 19/20 on 20 October 2019 between the teams: Manchester United Football Club (FC) and Liverpool FC) and politics (Brexit Super Saturday 2019) are used.We used Twitter to collect the data because it is widely considered as an information network than social media (Adedoyin-Olowe et al. 2016;Kwak et al. 2010), and has limited restrictions with enough data coverage for this research.To measure the performance, we used the evaluation metrics of recall, precision, F-measure and keyword recall, which are widely used to evaluate the event detection methods.Further, we compared the effectiveness and efficiency of our method with three recently proposed event detection methods as baselines.We obtained promising results for the evaluation which outperformed the baseline results.
To the best of our knowledge, Embed2Detect is the only method which uses self-learned word embedding-based temporal cluster similarity for event detection in social media.In summary, we list the contributions of this paper as follows: -Proposing a novel method named Embed2Detect for event detection in social media by involving the semantical features using the self-learned word embeddings in addition to the statistical and syntactical features in the text; -Leveraging self-learned word embeddings for more effective and flexible event detection which is independent of characteristics specific to the social media service, language or domain; -The application of Embed2Detect to recent real data sets to provide an insight on effectiveness and universality of the method with a comparison over state-of-the-art methods; -The publication of recent social media data sets 1 which represent different domains (i.e.sports and politics) with ground truth event labels to support other research in the area of event detection; and -The release of method implementation as an open-source project2 to support applications and research in the area of event detection.
The rest of this paper is organised as follows.Available methods for event detection in social media and their capabilities are discussed in Sect. 2. Section 3 describes the background details including the support of word embeddings and hierarchical clustering for this research.The problem addressed by this research is stated in Sect. 4 and the proposed approach is explained under Sect. 5. Following this, a comprehensive experimental study is available under Sect.6.Finally, the paper is concluded with a discussion in Sect.7.

Related work
Considering the importance of automatic event detection in social media, different methods have been proposed by previous research with the association of different techniques and characteristics including graph theory, rule mining, clustering, burstiness and social aspect.These techniques were supported by different text representations including tokens, n-grams, vectors, etc.; and extracted keywords such as named entities, noun phrases and hashtags as further discussed below.Additionally, more comprehensive surveys done by Weiler et al. (2017) and Hasan et al. (2018) can be referred to obtain a deep understanding of qualities and discrimination of available methods.
Graph theory Following the successful application of graph theory in sociology and social network analysis, there has been a tendency to use graph-based solutions for event detection in social media.Sayyadi et al. (2009) proposed to transfer a data stream into a KeyGraph, which represents the keywords by nodes and connects the nodes if corresponding keywords co-occurred in a document so that the communities in the graph represent the events that occurred in the data stream.As keywords, noun phrases and named entities with high document frequency were considered, and for community detection, betweenness centrality score was used.Later research suggested using Structural Clustering Algorithm for Networks (SCAN) to extract the communities in the graph and social media posts as graph nodes (Schinas et al. 2015).Unlike the betweenness centrality-based cluster detection, SCAN has the ability to recognise bridges of clusters (hubs) and outliers, to allow for the sharing of hubs between clusters and recognition of outliers as noise (Xu et al. 2007).However, this approach is unable to handle the events which are unknown to the training data, as it uses a supervised learning technique to identify similar posts during the graph generation.Other recent research suggested a named entity-based method considering the high computational cost associated with graph generation (Edouard et al. 2017).After identifying the entities, only the context around them was considered to extract nodes, edges and weights.Even though keyword-based methods speed up the graph processing, they are less expandable due to the usage of language or domain-specific features for keyword extraction.
Rule mining Previous research showed a trend of applying rule mining techniques for event detection.Based on Association Rule Mining (ARM), Adedoyin-Olowe et al. (2013) proposed a method for temporal analysis of evolving concepts in Twitter which was named Transaction-based Rule Change Mining (TRCM).To generate the association rules, hashtags in tweets were considered as keywords.This event detection methodology was further evolved by showing that specific tweet change patterns, namely, unexpected and emerging, have a high impact on describing underlying events (Adedoyin-Olowe et al. 2016).Having a fixed support value for Frequent Pattern Mining (FPM) was found to be inappropriate for dynamic data streams and it was solved by the dynamic support calculation method proposed by Alkhamees and Fasli (2016).FPM considers all terms in equal utility.But, due to the short length in social media documents, frequency of a specific term related to an event could increase rapidly compared to other terms.Based on this finding, Choi and Park (2019) suggested High Utility Pattern Mining (HUPM) which finds not only the frequent but also the high in utility itemsets.In this research, the utility of terms was defined based on the growth rate in frequency.Even though the latter two approaches which are based on FPM and HUPM are not limited to the processing of special keywords such as hashtags, they are focused on only identifying the topics/events discussed during a period without recognising temporal event occurrence details.
Clustering By considering the dynamicity and unpredictability of social media data streams, there has been a tendency to use clustering for event detection.McCreadie et al. (2013) andNur'Aini et al. (2015) showed that K-means clustering can be successfully used to extract events.In order to improve efficiency and effectiveness, they clustered low dimensional document vectors, which were generated using Locality Sensitive Hashing (LSH) and Singular Value Decomposition (SVD), respectively.Considering the requirement of predefining the number of events in K-means clustering, there was a motivation for hierarchical or incremental clustering approaches (Corney et al. 2014;Li et al. 2017a;Nguyen et al. 2019;Morabia et al. 2019).Different data representations such as word n-grams (Corney et al. 2014), semantic classes (Li et al. 2017a) and segments (Morabia et al. 2019) were used with hierarchical clustering.Rather than focusing on individual textual components, Comito et al. (2019a) suggested to cluster social objects which combine the different features available within documents.This idea was further improved with the incorporation of word embeddings (Comito et al. 2019b).In addition to word embeddings, term frequency-inverse document frequency (tf-idf) vectors also used to represent documents during clustering (Nguyen et al. 2019;Hasan et al. 2019).The document clusteringbased approaches assume that a single document only belongs to the same event.But with the increased character limits by most of the social media services, multiple events can be described within a single document.Considering this possibility, individual components (e.g.n-grams, segments) are more appropriate to use with clustering.Among the abovementioned such data representations, segments are more informative, specific and easy to extract, because they are separated using semantic resources such as Wikipedia.
Burstiness In communication streams, a burst is defined as a transmission which involves a larger amount of data than usual over a short time.Van Oorschot et al. (2012) suggested that occurrences of sports events in Twitter can be recognised by analysing the bursts of tweets in the data stream.Similarly, Li et al. (2014) proposed an incremental temporal topic model which recognises the bursty events based on unusual tweet count changes.But the events which do not lead to any notable increase in the data volume would be missed if only the data at peak volumes are considered.To overcome this limitation, another research proposed to use bursts in word n-grams (Corney et al. 2014).This research argues that even when the data volume is stable, there will be an increase in word phrases specific to a particular event.However, frequency-based measures cannot differentiate the events from general topics such as car, music, food, etc., because social media contains a large proportion of data relevant to these topics.Moreover, the bursts in frequency will appear when an event becomes more popular or is trending.To overcome these issues, bursts in word acceleration was suggested by other research (Xie et al. 2016).Using the acceleration, events could be identified more accurately at their early stages.
Social aspect Recently, there was a focus on the social aspect considering the impact the user community has on events.Guille and Favre (2015) proposed an approach which focuses on the bursts in mentions to incorporate the social aspect of Twitter for event detection.Since the mentions are links added intentionally to connect a user with a discussion or dynamically during re-tweeting, the social aspect of data can be revealed using them.Proving the importance of the social aspect, this method outperformed the methods which are only based on term frequency and content similarity (Benhardus and Kalita 2013;Parikh and Karlapalem 2013).Recent research has also suggested an improved version of Twevent (Li et al. 2012) by integrating more user diversity-based measures: retweet count and follower count with segment burstiness calculation (Morabia et al. 2019).However, the measures which gauge the social aspect are mostly specific to the social media platform and the incorporation of them would require customising the main flow accordingly.
Considering the textual features used in the above-mentioned event detection approaches, it is clear to us that the majority of previous research mainly focused on statistical features (e.g. term frequency, tf-idf, or burstiness), and syntactical features (e.g.co-occurrence, or local sensitivity).As a subdomain of information retrieval from text, effective event detection in social media also requires the proper inclusion of semantical features even though we could only find a few methods which considered the underlying semantics as described in Sect.2.1.

Usage of semantics in event detection
When we closely analysed how semantics is used by previous research for event detection in social media, we found some rule-based and contextual prediction-based approaches as further discussed below.Li et al. (2017a) defined an event as a composition of answers to WH questions (i.e.who, what, when and where).Based on this definition, they considered only the terms which belong to the semantic classes: proper noun, hashtag, location, mention, common noun and verb for their event detection method.Rule-based approaches were used for the term extraction and categorisation.Likewise, another recent research (Nguyen et al. 2019) also used a rule-based approach to extract the named entities in the text in order to support their event detection method.Using the named entities, documents and clusters were represented as entity-document and entity-cluster inverted indices which were used for candidate cluster generation.Both of these methods only categorised the terms into semantical groups for the recognition of important terms related to events.Thus, none of these methods has the ability to identify the connections between words.
In contrast to the rule-based approaches, Chen et al. (2017) suggested a deep neural network-based approach for event detection.To identify event-related tweets a neural network model was used and to input the data into the network tweets were converted into fixed-length vectors using pre-trained GloVe embeddings (Pennington et al. 2014), while capturing the semantic and syntactic regularities in the text.It is not appropriate to use supervised learning techniques for real-time event detection, because they require prior knowledge of events which can vary due to the dynamic nature in data streams and eventspecific qualities.Similarly, Comito et al. (2019b) proposed to use tweet representations generated using pre-trained Skip-gram embeddings (Mikolov et al. 2013a).This method was based on incremental clustering which is more suitable for event detection in social media than supervised learning.However, both of these methods use pre-trained embeddings which unable to capture the characteristics specific to the targeted corpus such as modified or misspelt words.
In summary, based on the available literature, we could not find any event detection approach which adequately involves semantics of the underlying text considering the characteristics specific to social media.We propose our approach with the intention to fill this gap for more effective event identification.

Background
Considering the limitations in available approaches for event detection, we adopt an approach which is based on word embeddings and hierarchical clustering in this research.The background details for word embeddings and their capabilities are discussed in Sect.3.1.The basic concepts of hierarchical clustering are explained in Sect.3.2.

Word embeddings
Word embeddings are numerical representations of text in vector space.Depending on the learning method, they are categorised into two main groups as frequency-based and prediction-based embeddings.Frequency-based embeddings consider different measures of word frequencies to represent text as vectors while preserving statistical features.Unlike them, prediction-based embeddings learn representations based on contextual predictions while preserving both syntactical and semantical relationships between words.Considering these characteristics, we focus on prediction-based word embeddings in this research and will use the term 'word embeddings' to refer to them.
Different model architectures such as Neural Network Language Model (NNLM) (Bengio et al. 2003) and Recurrent Neural Network Language Model (RNNLM) (Mikolov et al. 2010) were proposed by previous research for the generation of word embeddings based on contextual predictions.However, considering the complexity associated with them, log-linear models which are known as Word2vec models (Mikolov et al. 2013a) were suggested and popularly used with NLP applications.There are two architectures proposed under Word2vec models: (1) Continuous Bag-of-Words (CBOW) and (2) Continuous Skip-gram.CBOW predicts a word based on its context.In contrast to this, Skip-gram predicts the context of a given word.According to the results obtained by model evaluations, Mikolov et al. (2013a) showed that these vectors have a high capability in preserving syntactic and semantic relationships between words.
Among the Word2Vec algorithms, we focus on Skip-gram model in this research, because it resulted in high semantic accuracy compared to CBOW (Mikolov et al. 2013a, b).Also, based on the initial experiments and analyses, Skip-gram outperformed the CBOW model.More details of the Skip-gram architecture are described in Sect.3.1.1.Following this theoretic exposure, Sect.3.1.2discusses the qualities of word embeddings obtained by training Skip-gram models on real data sets which are useful for event detection.

Skip-gram model
Skip-gram model is a log-linear classifier which is composed by a 3-layer neural network with the objective to predict context/surrounding words of a centre word given a sequence of training words w 1 , w 2 , ...w n (Mikolov et al. 2013b).More formally, it focuses on maxi- mizing the average log probability of context words w k+j | − m ≤ j ≤ m, j ≠ 0 of the centre word w k by following the objective function in Eq. 1.The length of the training context is represented by m.
The probability of a context word given the centre word; p(w k+j |w k ) is computed using the softmax function.
In Eq. 2, w o and w i represent the output and input (i.e.context and centre words) and N rep- resents the length of vocabulary.The input and output vectors of a word w is represented by v w and v ′ w .The input vectors for words are taken from input-hidden layer weight matrix M which is sized N × D where D is the number of hidden layers.Likewise, output vectors are taken from hidden-output layer weight matrix M ′ which is sized D × N .The architec- ture of Skip-gram model including weight matrices is shown in Fig. 1.
Once the model converges, it obtains an ability to predict the probability distributions of context words with good accuracy.At that point, instead of using the model for trained purpose, adjusted weights between the input and hidden layers will be extracted as word representations or embeddings.Thus, by changing the number of hidden layers of the model, the number of neurons and also the dimensionality of vectors can be changed.Following the training procedure, model weights are adjusted by learning the connections between nearby words.Provided a sufficient data corpus, learning the connections between nearby words allows capturing underlying syntax and semantics with the capability of grouping similar words more effectively.

Skip-gram vector spaces learned on event data
An event discussed in a data stream will result in a collection of documents which describe that event using a set of words related to it.Due to the learning based on contextual predictions, word embeddings has an ability to locate the vectors of contextually closer words in nearby vector space or group similar words.This characteristic allows generating nearby vectors for the event-related words when the embeddings are learned on the corresponding document corpus.
Let us consider the sample events mentioned in Table 1 which are extracted from English Premier League 19/20 on 20 October 2019 between the teams Manchester United FC and Liverpool FC relating to the players Marcus Rashford and Roberto Firmino.Both events corresponding to Firmino are about missed attempts.Rashford has two different events relating to a foul and a goal.By analysing the Twitter data posted during each minute, we could find a significant amount of tweets which discuss these events.In these tweets, foul related words were used in the context of word 'Rashford' at 16:52 and goalrelated words were used at 17:06.Likewise, missed attempt related words were used in the context of 'Firmino' at 16:40 and 17:04.
To analyse the word embedding distribution over vector space and its temporal variations relating to these events, we trained separate Skip-gram models for each time window using Twitter data.In order to provide enough data for embedding learning, 2-minute time windows were used.Using the learned embeddings, most similar words to the player names Rashford and Firmino were analysed during the time windows 16:52-16:54 and 17:06-17:08.To visualise the similar words in a two-dimensional plane, T-distributed Stochastic Neighbor Embedding (t-SNE) algorithm (Van der Maaten and Hinton 2008) was used and resulted graphs are shown in Figs. 2 and 3.
The similar word visualisation during 16:52-16:54 (Fig. 2) shows that the foul related words are located closer to the word 'Rashford' in the vector space.Also, after 12 minutes, few words related to the missed attempt at 16:40 such as 'loses' and 'destruction' can be seen closer to the word 'Firmino'.But, during 17:06-17:08, we can see more words related to the saved attempt as nearby vectors to 'Firmino', because this event occurred 2 minutes back (Fig. 3).Also, the goal scored during 17:06 can be clearly identified by the words closer to 'Rashford'.This time window has clearly separated nearby vector groups for 'Firmino' and 'Rashford' compared to the previous window 16:52-16:54 to indicate that both events are actively discussed during this time because they happened recently.These similar word analyses prove that nearby vector groups have the ability to represent the events.Thus, the events described in a document corpus can be identified using the learned embeddings.Skip-gram word embeddings locate directly as well as indirectly related words to an event in closer vector groups.For an example, the top 20 similar words to 'Rashford' at the time window; 17:06-17:08 (Fig. 3), contains the words such as 'goal', '1-0', 'mufc' and '36' which are directly related to the event goal scored at 36 minute.Also, similar words contain words such as 'huge' and 'noise' which relate indirectly to the event but describe it more.These characteristics associated with Skip-gram word embeddings can be utilised for effective event detection in social media data.

Hierarchical clustering
Even though flat clustering (e.g.K-means) is efficient compared to hierarchical clustering, flat clustering requires the number of clusters to be predefined.Considering the unpredictability associated with social media data, it is not practical to identify the number of events in advance.Therefore, hierarchical clustering is more appropriate for social media data streams.Another advantage in hierarchical clustering is the output of a hierarchy or structure of data points, which is known as dendrogram rather than just returning the flat clusters.This hierarchy can be used to identify connections between data points.Considering these advantages, hierarchical clustering is used for our event detection approach.
There are two types of hierarchical clustering algorithms, bottom-up or agglomerative and top-down or divisive (Manning et al. 2008a).In hierarchical agglomerative clustering (HAC), all data points are considered as separate clusters at the beginning and then merge them based on cluster distance using a linkage method.The commonly used linkage criteria are single, complete and average.In single linkage, the maximum similarity is considered and in complete linkage, the minimum similarity is considered.Average of all similarities are considered in the average linkage.In contrast to HAC, hierarchical divisive clustering (HDC), considers all data points as one cluster at the beginning and then divide them until each data point is in its own cluster.For data division, HDC requires a flat clustering algorithm.
HDC is more complex compared to HAC, due to the requirement of a second flat clustering algorithm.Therefore, when processing big data sets, HDC is recommended to use with some stopping rules to avoid the generation of complete dendrogram in order to reduce the complexity (Roux 2018).Since we need to process big data sets and focus on clusters as well as complete dendrograms, we decided to use HAC for this research.

Problem definition
The problem targeted by this research is automatically detecting events in (near) real-time from social media data streams.The concept behind a data stream is introduced with Definition 1.Looking at available event detection approaches, they can be mainly divided into two categories from the perspective of the input data stream, as general and focused.In the general scenario, the whole data stream is processed (McCreadie et al. 2013;Nguyen et al. 2019).In the focused scenario, a user-centred data stream filtered from the whole data stream is processed.Two types of filtering techniques as keyword-based and location-based were commonly used.In keyword-based filtering, a domain-specific data stream will be extracted using a set of keywords (Aiello et al. 2013;Alkhamees and Fasli 2016;Comito et al. 2019b).In location-based filtering, a data stream composed by a set of documents posted by users in a particular location will be extracted (Li et al. 2012;Guille and Favre 2015).Comparing the two filtering techniques, the location-based method seems to add unnatural restrictions, because events of a corresponding location can be reported by users who are located elsewhere and location details are not available with all user accounts.

Definition 1 Social media data stream
Rather than focusing on the whole data stream which is a combination of multiple domains, we address a keyword-based user-centred scenario in this research.This approach was selected considering the restrictions in other filtering techniques and real-world requirements on information extraction.In many real scenarios, people or domain experts need the quick extraction of information in an interested domain rather than extracting all the information available (Aiello et al. 2013).For examples, football fans would like to know football updates, fire brigades would like to know fire updates and BBC politics news crew would like to know political updates.
In this setup, the whole stream needs to be narrow downed initially using some keywords (seed terms) specific to the targeted domain.In the case of social media data streams, they can be easily filtered by commonly used tags in the area of interest.In the majority of cases, these tags will be known by domain experts.Also, many applications can identify trending tags in social media corresponding to different domains3 .Definition 2 introduces the concept of the keyword-based filtered data stream.
Definition 2 Filtered data stream A filtered or narrow downed data stream which consists of posts that contain at least one of the selected seed terms.
Events were described using various definitions by previous research.Sayyadi et al. (2009) defined an event as some news related thing happening at a specific place and time.Also, events were considered as occurrences which have the ability to create an observable change in a particular context (Aldhaheri and Lee 2017).Focusing on the content of events, another research described an event as a composition of answers to WH questions (i.e.who, what, when and where) (Li et al. 2017a).Considering the main idea used to describe an event, we use the Definition 3 for events.Definition 3 Event An incident or activity which happened at a certain time and discussed or reported significantly in social media.
Additionally, we use the concept of time windows to formulate the targeted problem as it is widely used in previous research (Aiello et al. 2013;Adedoyin-Olowe et al. 2016;Morabia et al. 2019).Given a filtered data stream, it will be separated into time windows so that we can assess each window to identify event occurred time windows (Definition 4).The length of time windows needs to be provided by domain experts considering the intended update rate.For highly evolving domains like football, time window length needs to be short enough to capture the quick updates and for slowly evolving domains like politics, time window length need to be large enough to capture slowly developed updates.Similarly, the control of event significance also needs to be given to the domain experts, because we cannot define a fixed significance level for different people groups and domains.
Definition 4 Event occurred time window Duration of time, where at least one event has occurred.
In summary, the aim of the system described in this paper is, given a filtered data stream, identifying event occurred time windows in (near) real-time including the corresponding event-related words.Using the hyper-parameters, users are allowed to set the significance and update rate of interested events.

Notations of terms
Providing that the proposed approach is time window-based, the notations W t and W t+1 are used to denote two consecutive time windows at time t and t + 1 .All the notations which are commonly used throughout this paper are summarised in Table 2.

Embed2Detect
As Embed2Detect, we propose an event detection approach which is based on word embeddings and hierarchical agglomerative clustering.The main novelty of this approach is the involvement of corpus oriented semantical features for event detection using self-learned word embeddings.Further, the temporal variations between clusters and vocabularies are considered to identify events without relying on clusters directly.The Embed2Detect system contains four main components: (1) stream chunker, (2) word embedding learner, (3) event window identifier and (4) event word extractor as shown in Fig. 4. Self-learned word embeddings are used during event window identification and event word extraction phases.In order to evaluate the performance of this approach, event mapper is used to map detected events with ground truth events during experiments.Each of the components is further described in Sects.5.1 -5.4.Finally, the computational complexity of Embed2Detect is discussed in Sect.5.5.

Stream chunker
Data stream mining is supported by three different time models, namely, landmark model, tilted-window model and sliding window model (Tsai 2009).In the landmark model, all the data from a specific time to present is considered equally.The tilted-window model treats recent data with high importance compared to old data.Sliding window model splits the data stream into windows based on a fixed time period or number of transactions and performs data mining tasks on the data that belong to each window.Among these models, the time-based sliding window model was widely used by previous research work in event detection (Sayyadi et al. 2009;Alkhamees and Fasli 2016;Adedoyin-Olowe et al. 2016;Choi and Park 2019).Analysing the performance of previous methods and considering the requirement of temporal event identification, Embed2Detect also uses the sliding window model with a fixed time frame for event detection in social media data streams.
Stream chunker is the component which facilitates the separation of the data stream into windows.Depending on the evolution of events which need to be identified, the length of time frames can be adjusted.Smaller time frames are preferred for highly evolving events.

Word embedding learner
In order to incorporate statistical, syntactical and semantical features in text for event detection, word embeddings are used.Without using pre-trained word embeddings, this research proposes to learn embeddings on the targeted corpus to capture its unique characteristics such as modified or misspelt words and emoticons.The word embedding learner transfers the text in social media posts in a selected time window to a vector space.For each time window, different vector spaces are learned to capture variations between them.Learned word embedding models are stored to facilitate event window identification and event word extraction.
Considering the high-quality vector representations by the Skip-gram algorithm, we used it to learn embeddings in Embed2Detect.Due to the simplicity in this model architecture and usage of small training corpora (chunks of a data stream), time complexity on learning is not considerably high to make a bad impact on real-time event detection.

Event window identifier
Given a chronological stream of time windows W 1 , W 2 , ...W t , W t+1 , ... , event window iden- tifier recognises the windows where events have occurred.Since an event is an incident or activity which happened and discussed, such occurrence should make a significant change in data in the corresponding time window compared to its previous window.Based on this assumption, our method identifies windows with higher change than a predefined threshold ( ) as event windows.From the perspective of use cases, this threshold mainly defines the significance of targeted incidents.A low value would identify less important events too.Since normalised values are used to measure the change, possible values of are ranged between 0 and 1.
Before moving into the change calculation phase, we preprocess the text in social media documents for more effective results and efficient calculations.We do not conduct any preprocessing steps before learning the word embeddings except tokenizing to preserve all valuable information, which would help the neural network model to figure things out during word embedding learning.As preprocessing in event window identification phase, redundant punctuation marks and stop words in the text are removed.Further tokens with a frequency below a predefined threshold ( ) are removed as outlier tokens (e.g.words which are misspelt, or used to describe non-event information).
An event occurrence will make changes in nearby words of a selected word or introduce new words to the vocabulary over time.For example, in a football match, if a goal is scored at W t , 'goal' will be highly mentioned in the textual context of a player's name.If that player receives a yellow card unexpectedly in W t+1 , new words; 'yellow card' will be added to the vocabulary and they will appear in the context of the player's name, except the word 'goal'.Following these findings, in Embed2Detect, we consider two measures which capture the changes in nearby words and vocabularies to calculate textual data change between two consecutive time windows W t and W t+1 .To compute the nearby word changes, we propose a measure which is based on clustered word embeddings under cluster change calculation (Sect.5.3.1).According to the Sect.3.1.2,word embeddings can be used effectively to identify nearby word changes based on both syntactical and semantical aspects.Vocabulary change calculation is described in Sect.5.3.3.The final value for the overall textual change between time windows is calculated by aggregating the two measures, cluster change and vocabulary change.As the aggregation method, we experimented maximum and average value calculations (Sect.6.3).Comparing these two methods, the best results could be obtained by using the maximum calculation.An overview for window change calculation is shown in Fig. 5 and complete flow of event window identification is summarised in Algorithm 1.

Cluster change calculation
Cluster change calculation is proposed to measure nearby word or word group changes over time.To facilitate this calculation, separate clusterings need to be generated per each time window W t .We propose to cluster tokens that include words as well as other useful symbols such as emojis.As token representations, self-learned word embeddings are used while preserving syntactical and semantical features of the underlying corpus.According to previous research, document clustering approaches were more popularly used with event detection (Nur'Aini et al. 2015;Nguyen et al. 2019;Comito et al. 2019a, b).However, with the recent increments made to character limits by social media services (e.g.Twitter increased 140 character limit to 280), there is a possibility to contain multiple event details in a single document.Therefore, the token level is more appropriate than the document level to identify events.In addition to limiting only to words, useful symbols such as emojis were incorporated as they are widely used to express ideas in social media.
As the clustering algorithm, we chose hierarchical agglomerative clustering (HAC) considering its advantages and the tendency by previous research (Sect.3.2).As the linkage method, we used the average scheme in order to involve all the elements that belong to clusters during distance calculation.In average linkage, the distance between two clusters; C i and C j is measured by following the Eq. 3 (Müllner 2011), where d(w p , w q ) represents the distance between cluster elements w p and w q which belong to the clusters C i and C j respectively.This distance is measured using cosine distance, because it proved effectiveness for measurements in textual data (Mikolov et al. 2013a, b;Antoniak and Mimno 2018).Since cosine distance calculation is independent from magnitude of vectors, it does not get biased by the frequency of words (Schakel and Wilson 2015).
Even though we use hierarchical clustering, our method does not rely on direct clusters, as the majority of available methods (Li et al. 2017a;Comito et al. 2019b).We propose to use temporal cluster changes measured over time windows using dendrograms to identify nearby word changes.In this setting, the requirement for a cluster threshold can be eliminated.This elimination is advantageous in the context of social media because it is hard to define a static threshold suitable for all event clusters considering their diversity.
After generating clusters, a matrix-based approach is used to compute cluster change.Token similarity matrices are generated for each time window W t considering its next time window W t+1 .The token similarity matrix is a square matrix of size N × N where N is the number of tokens in the vocabulary.Each cell in the matrix matrix [i, j] represents the cluster similarity between token i and token j .To calculate the cluster similarity between tokens, we propose dendrogram level (DL) similarity measure (Sect.5.3.2) which is based on hierarchical clusters of token embeddings.To compare similarity matrices between two consecutive time windows, a common vocabulary is used for matrix generation.Since we compare W t+1 against W t , preprocessed vocabulary at t + 1 vocab t+1 is used as the common vocabulary for both windows.After generating the similarity matrices at t and t + 1 using DL similarity between tokens, the absolute difference of matrices is calculated.Then the average on absolute differences is measured as the value for cluster change in W t+1 com- pared to W t .During the average calculation, we only considered the values at the upper triangular matrix except the diagonal, because the matrix is symmetric around the diagonal.

Dendrogram level (DL) similarity
A dendrogram is a tree diagram which illustrates the relationships between objects.These diagrams are typically used to visualise hierarchical clusterings.A sample dendrogram generated on a selected word set from tweets posted during the first goal of English Premier League 19/20 on 20 October 2019, between Manchester United and Liverpool is shown in Fig. 6.Each merge happens considering the distance between clusters of words and they are represented by horizontal lines.Merges between the closer groups such as the name of the player who scored the goal 'rashford' and cluster which contains the word 'goal' happen at low distances ( ≈ 0.025 ).In contrast to this, merges between distant groups such as another player name 'firmino' and cluster of 'goal' happen at high distance values ( ≈ 0.25 ).Similarly, a dendrogram built on a corpus from any domain preserves informative relationships expressed in the corpus.
Focusing on the characteristics associated with dendrograms, we suggest the dendrogram level (DL) similarity to measure token similarity based on their cluster variations.Each horizontal line or merge represents a dendrogram level.Given a dendrogram, the similarity between a word pair w i and w j is calculated as the normalised value of shared levels from root between those two words, as follows.
The numerator of Eq. 4 represents the number of shared dendrogram levels between w i and w j from the root.The denominator represents the maximum number of levels between root and leaf nodes.We added leaf node level also as a separate level during maximum level count calculation to make sure only the similarity between the same token is 1 (4) DL Similarity (w i ,w j ) = dl (w i ,w j ) max(dl r→x ∶ x ∈ L) + 1 ( DL Similarity (w i ,w i ) = 1 ).For example, the maximum number of dendrogram levels from root to leaves in the diagram in Fig. 6 is 5.By adding the leaf node level, the maximum level count becomes 6.The count of shared levels between words 'rashford' and 'goal' is 4. But, words; 'firmino' and 'goal' shares only 1 level, because they appear in distant clusters.In measures, DL similarities between these words are as follows.

Vocabulary change calculation
A vocabulary is a set of distinct words that belong to a particular language, person, corpus, etc.In this research, we consider the tokens that belong to data corpora at each time window as separate vocabularies.Vocabulary change calculation is proposed to measure new word addition into time windows over time.Also, it incorporates the statistical details in the data set.In order to have a comparable value over all time windows, we calculated normalised vocabulary change value for W t+1 compared to W t following the Eq. 5.
The numerator of Eq. 5 represents the cardinality of new tokens that appeared in the vocabulary of W t+1 compared to W t , and the denominator represents the size of the vocabulary that belongs to W t+1 .

Event word extractor
After identifying a time window as an event occurring window, event word extractor facilitates the extraction of words in that window which are related to the occurred events.Since events make changes to the textual corpus, this component marks all the words in a window W t+1 which showed cluster changes compared to its previous windows W t as event words.Since we use a common vocabulary between consecutive windows during similarity matrix generation, cluster calculation identifies the newly added words to W t+1 also as words with changes.All words belong to the word pairs with DLsimilarity change above 0 are considered as the words which have temporal cluster changes.

Computational complexity
Analysing the components of Embed2Detect architecture, word embedding learner and event window identifier are the most computationally complex components available.Compared to them, the complexity of stream chunker and event word extractor is negligible.Therefore, for time and space complexity calculations, we only consider word embedding learner and event window identifier.
The training complexity of Skip-gram architecture is proportional to C × (D + D × log N) , where C is the maximum distance of the words, D is the dimen- sionality of vectors and N is the size of vocabulary (Mikolov et al. 2013a).Under event window identifier, there are two complex sub-components, clustering and similarity matrix generation.For N sized vocabulary, the time complexity for HAC algorithm is O(N 2 log N) (Manning et al. 2008a) and for matrix generation is O(N 2 ) .By combining all these chained components, the time complexity of Embed2Detect can be calculated as O(CD log N + N 2 log N) .For the used application, both C and D values are comparatively smaller than N. Therefore, the time complexity can be further simplified to O(N 2 log N).
Following the 3-layer architecture, the space requirement of Skip-gram model is equivalent to N × D + D × N .Similarly, both HAC algorithm and similarity matrix generation have a space complexity of O(N 2 ) .Considering all cost values, the total space complexity of Embed2Detect can be summarised as O(DN + N 2 ) .Using the same assumption men- tioned above, this can be simplified to O(N 2 ).
Based on the complexities summarised above, the vocabulary size N has a high impact on the computational complexity of this approach.According to the recent reports, approximately 511,200 tweets per minute were recorded in 2019 (James 2019).For a 30-minute time window, the total tweet count would be approximately 15M.As the time window length, 30 minutes is selected as a sufficiently long period to highlight the worst-case scenario.Looking at available twitter-based word embedding models, the Word2Vec_Twitter model is trained on 400M tweets and has a vocabulary of 3M tokens (Godin et al. 2015).Another popularly used model, GloVe Twitter is trained on 2B tweets and has a vocabulary of 1.2M tokens 4 .Focusing on the worst-case scenario, if there were 3M vocabulary for 400M tweets, N can be approximated to 0.1M (100,000) for 15M tweets.Since our approach is targeted in processing a filtered data stream specific to a particular domain, N should be notably smaller than this value approximation (0.1M) in a real scenario.Further, the size of vocabulary can be controlled using the frequency threshold ( ) mentioned in Sect.5.3.Based on these findings, we can state that Embed2Detect is suitable for real-time event detection.

Experimental study
In this section, we present the main results of the experiments which are conducted on social media data sets.More details about the data sets are described in Sect.6.1.To evaluate the results, the evaluation metrics mentioned under Sect.6.2 were used.We implemented a prototype of Embed2Detect in Python 3.7 which has been made available on GitHub5 .Using this implementation, we analysed the performance of aggregation methods (Sect.6.3) and impact by text preprocessing (Sect.6.4).Furthermore, an analysis of parameter sensitivity was also conducted (Sect.6.5).To compare the performance of Embed2Detect with available methods, we considered three recent event detection methods from different competitive areas as baselines (Sect.6.6).The corresponding comparison of results is reported under Sect.6.7.A comprehensive evaluation on the efficiency of Embed2Detect is available in Sect.6.8.Finally, we conducted some experiments to suggest possible extensions to Embed2Detect using other word embedding models and the obtained results and suggestions are summarised in Sect.6.9.All of the experiments were conducted on an Ubuntu 18.04 machine which has 2.40GHz 16 core CPU processor with 16GB RAM.

Data sets and preparation
To conduct the experiments and evaluations, we used data sets collected from Twitter.Considering some major issues associated with existing and available data sets, we decided to create and release our own data sets (Sect.6.1.1).Further details on data collection methods and data cleaning methods used in this research are mentioned in Sects.6.1.2and 6.1.3respectively.

Data sets
The most recent data sets for social media event detection were released based on data in 2012 (Aiello et al. 2013;McMinn et al. 2013).The data set released by McMinn et al. (2013) (Events2012) is used the Twitter stream from 10 October 2012 to 7 November 2012.Aiello et al. (2013) released three data sets using filtered Twitter streams correspond to the domains of sports and politics.Due to the restrictions made by Twitter, both of these corpora only contained tweet IDs which can be used to download the tweets.
Downloading the tweets in Events2012 corpus, we could only retrieve 65.8% of the tweets, as the rest were removed.In addition to the issue of missing a large proportion, a few more issues were encountered with this data set as follows.Since this data set was initially designed by considering event detection as identifying event clusters, only the event descriptions were provided as ground truth without temporal details.Therefore, following a commonly used strategy, we had to separate data into 1-day time windows (Alkhamees and Fasli 2016;Morabia et al. 2019).But, the 1-day time window is too lengthy to obtain quick updates targeted by our research.After time window separation, we found that all time windows are employed with events, due to the usage of a general data stream.The event occurred time window identification cannot be properly evaluated when all-time windows hold events.
Considering the data sets released by Aiello et al. (2013), they were designed by focusing the same problem targeted by this research.But, similar to the Events2012 data set, a large proportion of data was found to be not available for downloading.For example, only 63.4% of the sports data set could be downloaded.In addition to this issue, a major change to the tweet content was made in 2017, the character limit has been increased to 280 from 140.Thus, tweets in 2012, are comparatively small in character length compared to post 2017 tweets.
Considering the above-mentioned issues in available data sets, we decided to create new data sets to evaluate our approach.Also, we believed that releasing recent data sets would helpful to the research community too.Even though the proposed method is applicable to any social media data set, considering the restrictions, support and coverage given on data collection by different services, we decided to collect data from Twitter, similar to abovementioned data sets.While generating the data sets, we focused on two different domains, namely, sports and politics to prove the domain independence of our method.Sports is known as a domain with rapid evolution and politics is known as a domain with a slow evolution (Adedoyin-Olowe et al. 2016).This domain selection is also motivated by the data sets released by Aiello et al. (2013).More details on data collection are available in Sect.6.1.2.
A similar strategy to Aiello et al. (2013) was used to ground truth (GT) preparation.We reviewed the published media reports related to the chosen topics during the targeted period and selected a set of events.Each event was supported using a set of keywords taken from news and social media to compare with the identified event words.We made these data sets including the GT labels publicly available6 .

Data collection
Data collection was done using Twitter developer Application Programming Interfaces (APIs)7 .Initially, data belonging to a particular topic was extracted using a trending hashtag.Then the hashtags found in the extracted data set were ranked based on their popularity and popular hashtags were used for further data extraction8 .
To generate the sports data set, English Premier League 19/20 match between two popular teams, specifically, Manchester United and Liverpool was selected.This match was held at Old Trafford, Manchester on 20 October 2019.During the match, each team scored a single goal.Starting from 16:30, the total duration of the match was 115 minutes including the half time break.This data set will be referred to as 'MUNLIV' in the following sections.
To generate the political data set, Brexit Super Saturday in 2019 was selected.This was a UK parliament session which occurred on Saturday, 19 October 2019.It was the first Saturday session in 37 years.Even though it was organised to have a vote on a new Brexit deal, the vote was cancelled due to an amendment passed against the deal.This event started at 09:30 and held until around 16:30.This data set will be referred to as 'Brexit-Vote' in the following sections.
For MUNLIV, we collected 118,700 tweets during the period 16:15-18:30.Among them, we used 99,995 (84.2%) tweets posted during the match for experiments, because we could extract GT events only for this period using news media.For BrexitVote, we collected 276,448 tweets during the period 08:30-18:30 but only used 174,498 (63.1%) tweets posted from the beginning of the parliament session until the vote on the amendment for experiments.Similar to the scenario with MUNLIV, the focus by news media was found to be high until the vote to extract more accurate GT events.Considering the evolution rate of each domain, for the sports data set MUNLIV, 2 minute, and for the political data set BrexitVote, 30 minute time windows are selected.After separating the data into chunks, on average there were 1,724 and 14,542 tweets per time window in sport and political data sets respectively.

Data cleaning
To learn embeddings on separate tokens, embedding models need tokenised text.Since we focused on Twitter data sets during the experiments, we used the TweetTokenizer model available with Natural Language Toolkit (NLTK)9 to tokenise the text in tweets.This tokeniser was designed to be flexible on new domains with the consideration on characteristics in social media text such as repeating characters and special tokens.It has the ability to remove characters which repeats more than 3 times to generalise the various word forms introduced by users.For example, both words 'goalll' and 'goallll' will be replaced as 'goalll'.Further, it tokenises the emotions and words specific to social media context (e.g.1-0, c'mon, #LFC, :-)) correctly.Also, we did not preserve the case sensitivity in tokenised text.
In addition to tokenising, retweet notations, links and hash symbols were removed from the text.Retweet notations and links were removed because they do not make any contribution to the idea described.Hash symbols were removed to treat hashtags and other words similarly during embedding learning.To automate these removals, text pattern matching based on regular expressions was used.

Evaluation metrics
In order to evaluate the performance of the proposed method and baselines, event words are compared with GT event keywords using the following metrics.In the equations stated below, set of all event windows in the data set, detected event windows and relevant event windows found in detected windows are represented by W, W d and W r respectively.
-Recall Fraction of the number of relevant event windows detected among the total number of event windows that exist in the data set -Precision Fraction of the number of relevant event windows detected among the total number of event windows detected -F-Measure (F1) Weighted harmonic mean of precision and recall -Keyword recall Fraction of the number of correctly identified words among the total number of keywords mentioned in the GT events (Aiello et al. 2013).To calculate a final value for a set of time windows, micro averaging (Manning et al. 2008b) is used.
T represents the event occurred time windows, w represents the words/ keywords and GT represents the set of ground truth events.While calculating the recall, precision and F-measure, a detected window is marked as a relevant event window, if all the events occurred during that time period are found in the event words identified for that window.A match between event words and a GT event is established if at least one GT keyword corresponding to that event is found from the event words.Likewise, for keyword recall calculation, if at least one word mentioned in a synonym (similar) word group in GT is found, it is considered as a match.Therefore, the total number of GT keywords is calculated as the total of synonym word groups.

Aggregation method
As mentioned in Sect.5.3, to measure the temporal data change between two consecutive time windows, Embed2Detect needs to aggregate the values computed by cluster change calculation (Sect.5.3.1) and vocabulary change calculation (Sect. 5.3.3).For this aggregation, we experimented the techniques: average and maximum considering their simplicity and common usage.The obtained results are shown in Table 3.
According to the results, for MUNLIV data set, there is a slight change in F1 between average and maximum calculations.But, we can see balanced values for both recall and precision when the maximum is used.In BrexitVote, there is a clear change in F1, with higher value using the maximum calculation.Based on the observations made on two diverse domains, we decided to use maximum calculation as the default aggregation method in Embed2Detect.

Text preprocessing
Even though preprocessing improves the effectiveness, it can strongly restrict the generality of a method.Therefore, we believe it is worth experimenting the impact of text preprocessing on Embed2Detect.To maintain the simplicity of our method, we only suggest two preprocessing techniques, removal of punctuation marks and stop words.The evaluation results obtained with different configurations of these techniques are reported in Table 4.We only used cluster change calculation in these experiments, because it has a high influence by changes in tokens.
According to the obtained results, the highest F1 for both sport and political data sets is obtained by the tokens without punctuation and stop-words.Even though there is an improvement in the performance with preprocessing, these results show that we can obtain good measures without preprocessing also.This ability will be helpful in situations where we cannot integrate direct preprocessing mechanisms.For example, identifying events in low-resource language or multilingual data sets can be mentioned.However, to conduct the following experiments, we used the tokens without punctuation and stop-words, because both data sets used in this research are mainly written in English.

Parameter sensitivity analysis
In Embed2Detect, word embedding learner and event window identifier require some hyper-parameters.Sections 6.5.1 and 6.5.2 describe the impact of different hyper-parameter settings and heuristics behind their selections.

Parameters -word embedding learning
Word embedding learning requires 3 hyper-parameters: minimum word count, context size, and vector dimension.Given a minimum word count, the learning phase ignores all tokens with less total frequency than the count.Context size defines the number of words around the word of interest to consider during the learning process.Vector dimension represents the number of neurons in the hidden layer which also will be used as the dimensionality of word embeddings.
Considering the limited amount of data available in a time window, we fixed the minimum word count to 1. Nevertheless, we analysed how the effectiveness and efficiency of event detection vary with different vector dimensions and context sizes to select optimal values.To evaluate the effectiveness, F-measure (F1) was used and results obtained for both data sets are visualised in Fig. 7. Based on the results, there was no significant change in F1 with different vector dimensions and context sizes.But, there was a gradual increase in execution time when both hyper-parameter values are increased.
Accuracy of text-similarity can be improved with the increase of both the amount of training data and vector dimensionality (Mikolov et al. 2013a).Similarly, a larger context size can result in higher accuracy in text-similarity due to the provision of more training data (Li et al. 2017b).But, these effects were not notably captured with event detection.As a major reason for this, we can mention that event detection is not that sensitive to the syntactical and semantical structure of text same as with text-similarity tasks.Also, since we train separate models for each time window, each model has comparatively small data sets to learn embedding space.Therefore, even though the vector dimensions are increased, no sufficient data will be provided for their proper adjustment.
Following the results we obtained and the findings of previous research (Mikolov et al. 2013a;Li et al. 2017b), we fixed 100 dimensions for word embeddings and length of 5 for context size.The decision on dimensionality is mainly influenced by the training data limitations and learning time.The context size is chosen considering the requirement of providing sufficient knowledge for learning and also the execution time.

Parameters -event window identification
As described in Sect.5.3, the event window identifier requires two hyper-parameters, and . is used to indicate the significance level of targeted events, and is used as a frequency threshold to remove the outlier tokens.Both of these hyper-parameters need to be set by the user or domain expert, considering the characteristics associated with the selected domain or filtered data stream.
The main idea behind event window identification is based on overall textual data change between time windows.A high overall change indicates the occurrence of a major event(s) and low change indicates minor event(s).To provide a clearer insight, we plotted the variations of temporal overall change values of data sets, MUNLIV and BrexitVote in Fig. 8 and 9 respectively.Additionally, we plotted the total tweet count of each window in these graphs to highlight that the overall change-based measure is capable of identifying events which do not make a notable change to the total tweet count too.The total tweet count changes can only capture major events which make bursts.For comparison purpose, tweet counts are scaled down using the min-max normalisation.
Focusing on Fig. 8 corresponds to MUNLIV data set, more fluctuations on overall change can be revealed, due to the rapid evolution in the sports domain.To do a deep Compared to the goal, a missed attempt is a minor event in the sports domain and overall change measure is capable of capturing that distinction successfully.Following this ability, is used to filter the events based on user preference.For example, if equals 0.2, both events missed attempt and goal will be captured by Embed2Detect.However, if the value is increased to 0.3, among those two events, only the goal will be captured.There was no high number of fluctuations for BrexitVote data set (Fig. 9) because the political domain has a comparatively slow evolution than the sports domain.Unlike with the MUNLIV data set, this effect limits the overall change values to a small range while increasing the sensitivity of .By slight variation of (e.g. from 0.15 to 0.2), capturing events can be changed.
Following these analyses, it is infeasible to define a common value for different domains as well as for a particular domain.For different domains, this value needs to be picked, mainly considering the data evolution.Within a particular domain, can be varied according to personal preferences on event importance.However, it can be simply chosen by using the domain knowledge and analysing a few past time windows.
In addition to the value, Embed2Detect uses another threshold to remove outlier tokens.Defined a value, all the tokens with less frequency that it, such as misspelt and uncommon words will be removed.The impact by different values on overall change measure is represented in Fig. 10.According to this plot, with high values, some events can be missed (e.g.time window 16:40 will not be identified with = 80 ).Also, high could unnecessarily increase the overall change of some events (e.g.overall change increase at time window 18:06 and 18:12).
The major reason behind these behaviours is the removal of event tokens with high value.Due to this, the effectiveness of event detection decreases with increasing (Fig. 11).Therefore, value only need to be sufficiently large to remove outliers.Analysing the sport and political data sets, values less than 20 is appropriate for .But similar to , also highly depend on domain-specific characteristics such as word usage and audience.Therefore, we believe that is also a hyper-parameter which needs to be controlled by domain experts.

Baseline methods
Since there is no specific data set to evaluate event detection performance, available methods cannot be compared with each other to pick the best baseline.Therefore, considering the requirements of event detection and available competitive areas, we selected three recently proposed methods as baselines.The major requirements we focused on during this selection were effectiveness, efficiency and expandability.We also covered different competitive areas which can be summarised as the incorporation of the social aspect, word acceleration over frequency, unsupervised learning (tensor decomposition and clustering) and segments over uni-grams, to make the baselines strong enough.All of these methods process the whole data stream without considering only some keywords (e.g.hashtags) to identify temporal events, similar to our approach.More details on selected baseline methods are as follows.
-MABED (Guille and Favre 2015) Anomalous user mention-based statistical method Mention anomalies were taken into consideration in this research in order to incorporate the social aspect of Twitter with event detection rather than only focusing on textual contents of tweets.User mentions are links added intentionally to connect a user with a discussion or dynamically during re-tweeting.Anomalous variations in mention creation frequency and their magnitudes were used for event detection.To extract the event words, co-occurrences of words and their temporal dynamics were used.
-TopicSketch (Xie et al. 2016) Word acceleration-based tensor decomposition method Word acceleration is suggested by this research to support event detection because it has the ability to differentiate bursty topics (events) from general topics like car, food, or music.Events have the ability to force people to discuss them intensively.This force can be expressed by acceleration and this research proposed it as a good measure over frequency for event detection.To extract the event words, a tensor decomposition method, SVD was used.
-SEDTWik (Morabia et al. 2019) Segment-based clustering method powered by Wikipedia page titles This is an extension to the Twevent system (Li et al. 2012).Text segments are focused in this research because they are more meaningful and specific than unigrams.Wikipedia page titles were used as a semantic resource during segment extraction to preserve the informativeness of identified segments.To identify the events, bursty segments were clustered using Jarvis-Patrick algorithm.Burstiness of segments is calculated using both text statistics and user diversity-based measures.

Comparison with baselines
We compared the effectiveness and efficiency of Embed2Detect with selected baseline methods: MABED, TopicSketch and SEDTWik (Sect.6.6).Effectiveness was measured using the evaluation metrics: recall, precision, F1 and keyword recall (Sect.6.2).To measure the efficiency, total time taken to execute the complete process on full data sets and the average time taken per time window by each method were used.
Similar to the hyper-parameters and in Embed2Detect, all the baseline methods have their own parameters which need to be optimised depending on the data set.Therefore, to generate comparable results, a common strategy is used to identify optimal hyper-parameters, because they make a high impact on the method's performance.For each method, we evaluated all possible hyper-parameter settings to choose the best F1 value.For MABED, we optimised the hyper-parameters: number of events (k), maximum number of words describing each event (p), weight threshold for selecting relevant words ( ) and overlap threshold ( ).For k and p, starting from a low value we kept increasing them gradually until the maximum F1, which reduces with further increasing parameter values is reached.Similarly, for and , we experimented the values around the original values reported in initial experiments (Guille and Favre 2015).For TopicSketch, we decided to optimise only the most critical hyper-parameter due to the high time complexity of this method.Thus, while keeping default values for other parameters, we tested gradually increasing values for detection threshold to obtain the best F1 value.For SEDTWik, we optimised the hyperparameters: number of subwindows (M), number of cluster neighbours (k) and newsworthiness threshold ( ).Different values for M are picked considering the time windows lengths assigned to each data sets.For k and , starting from a low value, gradually increasing values were tested to obtain highest F1.For Embed2Detect, we identified all possible values for each hyper-parameter and , and experimented with all of their combinations to get best results.Following these parameter optimisations, results obtained for MUNLIV and BrexitVote are reported in Tables 5 and 6 respectively.The corresponding parameter settings are summarised in Table 7.To measure the reported execution times, we used sequential processing for the baseline methods according to the available implementations and parallel processing with 8 workers for Embed2Detect.Both total time taken to process the whole data stream and average time taken per time window are reported.Embed2Detect outperforms the baseline methods in both data sets with F1 of 0.652 on MUNLIV and F1 of 0.889 on BrexitVote.This proves that our method has the ability to detect the events effectively in diverse domains, specifically, sports and politics than the available methods.According to the recall and precision measures, all methods tends to return high recall than precision.When preparing the GT events based on news reports, there is a possibility to miss some important events which are only discussed within the social media platform (Aiello et al. 2013;Morabia et al. 2019).Due to that, some actual events can be labelled as false positives and it will reduce the precision value.Considering recall, the majority of the methods (except TopicSketch) resulted in high values with Brex-itVote than MUNLIV data set.Comparing the GT of two data sets, BrexitVote has 72.7% of event occurred time windows while MUNLIV has only 40.4%.Due to this bias, high recall can be resulted with BrexitVote data set.Theoretically, such bias is captured because of the low evolution and less dynamicity in the political domain.
Following the execution times, for MUNLIV, MABED took 168 seconds and Embed-2Detect took 34 seconds more than MABED.But on BrexitVote, Embed2Detect completed the execution in 310 seconds -222 seconds faster than MABED.In terms of average execution time per window, Embed2Detect took 2.947 seconds to process a 2-minute window in MUNLIV data set and 28.182 seconds to process a 30-minute window in BrexitVote data set.These time measures prove that Embed2Detect is sufficiently fast for real-time event detection.More detailed analysis of intermediate processing time of Embed2Detect is reported in Appendix.

Efficiency evaluation
Processing time is a critical measure in real-time applications.For successful event detection, the huge amount of data generated in social media needs to be processed in (near) real-time.Considering the problem targeted by Embed2Detect, efficiency requirement can  be further specified as processing the data belong to a time window within a sufficiently short period.Following this requirement, we evaluated the scalability of Embed2Detect and a parallelised version of Embed2Detect by measuring their execution times for the complete process on time windows with increasing data size.The obtained results are plotted in Fig. 12.As the data size within a time window (e.g.1-minute window), 5000-25,000 tweets were considered.Focusing on a filtered data stream, the upper limit of 25,000 tweets can be mentioned as a reasonable amount to depict the real scenario.

Method
According to the results, the sequential version of Embed2Detect took nearly 10 seconds to process 5000 tweets and this increased to 41 seconds to process 25,000 tweets.The parallel version with eight workers reduced the processing time to 6 seconds for 5000 tweets and 19 seconds for 25,000 tweets.Also, we noticed that for both implementations, sequential and parallel, execution time grew linearly with data size (Fig. 12).Following these results, we can confirm that our approach is adequately efficient to facilitate real-time processing.Due to the linear growth of execution time, we can guarantee that Embed2Detect is capable of handling data bursts too.

Extension to other word embedding models
Word embedding models other than Skip-gram can also be used with Embed2Detect.Since we implemented word embedding learner as a separate module in the Embed2Detect architecture, different word embeddings can be easily connected.But, it is important to consider the learning time and associated complexities while selecting a word embedding model to satisfy the goal of real-time event detection.In this section, we discuss the appropriateness of different architectures for word embedding generation in Embed2Detect.
For this analysis, we used fastText, BERT and DistilBERT models.FastText is an updated version of the Skip-gram model which considers subword information while  learning word representations (Bojanowski et al. 2017).Both BERT and DistilBERT are transformer-based models.According to the recent advances in the domain of NLP, transformers gained success in many areas such as language generation (Devlin et al. 2019), named entity recognition (Liang et al. 2020) and question answering (Yang et al. 2019).BERT: Bidirectional Encoder Representations from Transformers (Devlin et al. 2019) is the first transformer model which gained wide attention.This model is designed to train from unlabelled text using the masked language modelling (MLM) objective and to finetune for a downstream task, as a solution for the high data requirement by deep neural networks.DistilBERT is a distilled version of BERT which is light and fast (Sanh et al. 2019).
Initially, the time taken by different architectures to learn word embeddings is measured and obtained results are summarised in Table 8.Both Skip-gram and fastText models were trained from scratch using Twitter data as suggested by this research.Following the idea presented with transformers, for both BERT and DistilBERT, we retrained available models using our data.As the pre-trained BERT model, bert-base-uncased and DistilBERT model, distilbert-base-uncased released by HuggingFace's Transformers library (Wolf et al. 2019) are selected.According to the obtained results, classic word embedding models (e.g.Skip-gram and fastText) learn the representations faster than transformer-based models (e.g.BERT and DistilBERT).
Comparing fastText and Skip-gram, fastText took more time because it processes subword information.But, incorporation of subwords allows this model to capture connections between modified words.For example, consider the goal-related words found within the top 20 words with high cluster change during a goal score: Skip-gramgoal, goalll, rashyyy, scores fastTextgoalll, goooaaalll, rashford, rashyyy, @marcusrashford, scored, scores fastText captures more modified words than Skip-gram.We could not run a complete evaluation using fastText embeddings, because it requires a manual process since GT keywords only contain the words in actual form.
Transformer-based models took more time than both Skip-gram and fastText due to their complex architecture to learn contextualised word embeddings.DistilBERT is found to be faster than BERT, however, the learning time of DistilBERT is not fast enough for real-time processing because it exceeds the tweet generation time.For example to learn from tweets posted during a 2-minute time window, it took approximately 7.2 minutes.If this model can be further distilled, there is a possibility to achieve the required efficiency to become suitable for real-time processing.However, further distillation can reduce the language understanding capability of the model as there is a 3% reduction in DistilBERT compared to BERT (Sanh et al. 2019).
According to recent literature, transformer-based models performed well on many NLPrelated tasks, because of the ability to capture the contextual sense of words.BERT is capable of generating different embeddings for the same word depending on its surrounding context.In other words, the main idea behind BERT is capturing spacial changes of words.From the perspective of processing formally written natural language, this is a very useful feature.But, in social media, language is mostly informal and for event detection using social media text, temporal changes of words need to be more focused.If we consider a particular time window of a filtered data stream, it is rarely possible to have a word with two totally different contextual meanings.Therefore, the context awareness associated with BERT is not much useful for event detection.
Further, contextualised word embeddings could incorporate an additional complexity to the event detection method.For example, during a goal scoring of a football match, the word 'goal' will be expressed by the audience of the winning team and losing team differently.Even though the surrounding contexts are varied, the meaning of the word 'goal' targeted by event detection is constant.For such a scenario, BERT will return different embeddings for 'goal' as illustrated in Fig. 13.Having multiple embeddings for monosemy words can confuse the clusters and increase the computational complexity of the method exponentially.To overcome these issues, multiple embeddings of a word can be combined using an aggregation method.But, it breaks the main objective of contextualised word embeddings.Therefore, we believe it is inaccurate to aggregate context-aware embeddings.Following these findings, we can conclude that contextualised word embedding models such as BERT are less appropriate to be used with Embed2Detect, due to their complexities which are not necessary for event detection.

Conclusions and future work
In this paper, we proposed a novel event detection method coined Embed2Detect to identify the events occurred in social media data streams.Embed2Detect mainly combines the characteristics in word embeddings and hierarchical agglomerative clustering.This method uses self-learned word embeddings to capture the features in the targeted corpus in order to facilitate domain, platform or language-independent event detection.Therefore, Embed-2Detect can be easily applied on any social media data set in any language even though the majority of available methods are limited to specific platforms (e.g.Twitter) and languages (e.g.English).Further, this approach is also applicable to multilingual data sets.The ability to process multilingual data sets can be highlighted as an important requirement to process the data in social media considering its user base which is distributed all over the world.
In contrast with prior work, Embed2Detect not only considers syntax and statistics in the underlying text but also incorporates semantics.Inclusion of semantics allows to understand the relationships between words.Due to the huge and diverse user base, social media text contains different words and word sequences which describe the same idea.Knowing the relationships between words, differently described similar ideas and their connections can be extracted.Therefore, our approach is capable to reduce the information loss experienced in previous approaches due to the lack of semantic involvement.
According to the evaluations conducted, Embed2Detect performed significantly better than the recently suggested event detection methods, namely, MABED, TopicSketch and SEDTWik on both data sets MUNLIV and BrexitVote from the domain of sports and politics.As evaluation metrics, we used recall, precision F-measure and keyword recall to conduct a comprehensive evaluation.Also, we considered data from two contrasting domains which have different word usage, audience and evolution rate to evaluate the universality of methods.In addition to focusing on effectiveness, we measured the efficiency of Embed2Detect also, because real-time event detection is a time-critical operation.Embed2Detect performed event detection in both data sets within a short time period and it could handle increasing data volume to indicate its appropriateness for real-time application.In summary, the results we obtained from the experiments conclude that Embed2Detect can detect the events in social media data effectively and efficiently without depending on domain-specific features.
As an extension to Embed2Detect, more advanced word embedding learning methods can be applied.But, considering the learning time and associated complexities, to preserve the efficiency of the method, it is more suitable to use classic word embedding models such as Skip-gram than advanced word embedding models such as BERT.Under classic word embeddings, we hope to further analyse the impact by subword and character-based models which can be used to capture the connections between informal or modified text and their formal versions.Such an approach would be useful to understand informal text which is common to the context of social media.Further, focusing on the recent improvements to the domain of NLP by available transformer-based models, their pre-trained word embeddings can be supported to generate more comprehensive event details such as summaries using the detected event words in a future phase of this research.Also, we plan to further extend our method to identify event evolution over time to facilitate both event detection and tracking together.

Intermediate processing time
Even though we reported the execution time for the full process in the paper, we did an intermediate analysis to understand the complexity of each individual step.The obtained  9 and 10.In these tables, total time reports the time taken by whole corpus and average time reports the time taken by a single time window.Comparing the two data sets, MUNLIV has 58 2-minute time windows and Brexitvote has 12 30-minute time windows.
As discussed in Sect.5.5, embedding learner and event window identifier are the complex components in Embed2Detect architecture which take a comparatively large proportion of the total execution time with the increase of data size.However, according to the obtained results, even with sequential processing, data can be processed in less time than the time taken for their generation.With parallel processing, the execution time can be further reduced to be more appropriate for real-time processing.

Fig. 4
Fig. 4 Overview of proposed method for event detection; Embed2Detect

Fig. 5
Fig. 5 Overview of window change calculation

Fig. 6
Fig. 6 Sample dendrogram (y-coordinate denotes the cosine distance and x-coordinate denotes the selected words)

Fig. 7
Fig. 7 Analysis on F1 and execution time with different values for word embedding learning parameters; vector dimension and context size (Average time taken to execute the full process on single data window is used for time values in both data sets)

Fig. 8
Fig. 8 Overall change and tweet count variations over time windows -MUNLIV

Fig. 10
Fig. 10 Analysis on impact by different values on overall temporal change Fig. 11 Analysis on F1 with different values (with =0.14)

Fig. 12
Fig. 12 Execution time on different data sizes including the effect by sequential and parallel processing

Table 1
Sample events occurred during English Premier League 19/20 on 20 October 2019 (Manchester United -Liverpool)

Table 4
Evaluation results with different preprocessing techniques

Table 5
Performance comparison of Embed2Detect with baseline methods using MUNLIV data set

Table 6
Performance comparison of Embed2Detect with baseline methods using BrexitVote data set

Table 7
Parameter settings used by each method for the best results

Table 8
Time taken to learn embeddings by different architectures