Abstract
Purpose
Computational text mining methods are proposed as a useful methodological innovation in Intimate Partner Violence (IPV) research. Text mining can offer researchers access to existing or new datasets, sourced from social media or from IPV-related organisations, that would be too large to analyse manually. This article aims to give an overview of current work applying text mining methodologies in the study of IPV, as a starting point for researchers wanting to use such methods in their own work.
Methods
This article reports the results of a systematic review of academic research using computational text mining to research IPV. A review protocol was developed according to PRISMA guidelines, and a literature search of 8 databases was conducted, identifying 22 unique studies that were included in the review.
Results
The included studies cover a wide range of methodologies and outcomes. Supervised and unsupervised approaches are represented, including rule-based classification (n = 3), traditional Machine Learning (n = 8), Deep Learning (n = 6) and topic modelling (n = 4) methods. Datasets are mostly sourced from social media (n = 15), with other data being sourced from police forces (n = 3), health or social care providers (n = 3), or litigation texts (n = 1). Evaluation methods mostly used a held-out, labelled test set, or k-fold Cross Validation, with Accuracy and F1 metrics reported. Only a few studies commented on the ethics of computational IPV research.
Conclusions
Text mining methodologies offer promising data collection and analysis techniques for IPV research. Future work in this space must consider ethical implications of computational approaches.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Existing methods for studying intimate partner violence (IPV) draw largely from the social sciences. These include primary data collection tools such as surveys (Lagdon et al., 2022; ONS, 2020), interviews or focus groups (Øverlien et al., 2020; Wood et al., 2021), as well as secondary analyses of data sourced from, for example, victim advocacy organisations (Rogers et al., 2019).
Recent developments in the field of computational social science have led to data science tools which extend and complement these established techniques (DiMaggio, 2015; Evans & Aceves, 2016). They further ease the data collection and analysis process by harnessing big data and Machine Learning (ML) (Gauthier & Wallace, 2022). The latter is a subset of artificial intelligence focused on building algorithms that ‘learn’ statistical patterns from large amounts of data.
Specifically, computational text analysis or text mining – umbrella terms for computational tools which can extract and analyse substantial quantities of text data – have been successfully utilised in fields such as social work (Victor et al., 2021), medicine (Luque et al., 2019), and education (Ferreira‐Mello et al., 2019). Indeed, a small number of studies have applied similar approaches to the study of IPV. Publications examined online support-seeking behaviours of victim-survivors (Chu et al., 2021), studied reasons given for staying and leaving abusive relationships in microblog posts (Homan et al., 2020), and identified crisis posts on social media platforms such as Facebook (S. Subramani et al., 2018a, 2018b). In addition, computational methods have offered IPV researchers access to datasets which are simply too large to evaluate manually e.g. police incident reports (J. Poelmans et al., 2009a, 2009b, 2009c, 2009d; Wilson et al., 2021), case summaries (Victor et al., 2021), and Electronic Health Records (Botelle et al., 2022).
Despite this small but growing body of work, there is yet no review addressing the application of computational text analysis methods to the study of IPV. This omission stands in the way of proposing further methodological innovation, and to opening the field to the latest transdisciplinary research approaches stemming from computer science. This article seeks to fill this gap by conducting a systematic literature review of eight online academic databases (Scopus, ProQuest, Web of Science, IEEE Explore, PsychInfo, PubMed, ArXiv.org and ACM Digital Library).
The rest of the article is structured as follows: 1) Background: a short background to both IPV and text mining is provided to give context to later discussions. 2) Research Questions (RQs): Several RQs are proposed to investigate the use of text mining methods in the IPV domain. 3) Methodology: The methodology of this review is described, including the search strategy and inclusion criteria. 4) Results: The results of the review are summarised and analysed using a 21-item checklist. 5) Discussion: The findings from the review, its limitations, and potential directions for future work are discussed. 6) Concluding remarks.
Background
Data from the World Health Organisation (2021) indicates that 27% of women worldwide aged 15–49 years who have been in a relationship have experienced some form of physical or sexual violence from an intimate partner during their lifetime. The Crime Survey for England and Wales in 2020 indicated that 4.9% of women and 2.1% of men over the age of 16 had experienced some form of non-sexual partner abuse in the last year (ONS, 2020).
Despite these figures, accurately quantifying the prevalence of IPV is difficult (Walby et al., 2017). Much abuse goes unreported due to shame, bias, and unawareness (Stark, 2009). An additional barrier to measuring IPV is a non-homogenous set of definitions for what constitutes abuse across cultures, time periods, and organisations (Alhabib et al., 2010; Barocas et al., 2016). Whilst IPV is generally understood to involve physical abuse, there are other ways in which perpetrators cause harm (psychological, sexual, coercive controlling, economic, technology-facilitated abuse etc.). These alternative abuse forms may or may not be included in definitions, resulting in skewed evaluations (Alhabib et al., 2010; Dokkedahl et al., 2019).
Much of the existing large-scale data about IPV is drawn from traditional survey- and questionnaire-based research (Australian Bureau of Statistics, 2013; European Union Agency for Fundamental Rights, 2014). Whilst such surveys are useful to understand IPV on a population level, they are also costly, infrequent, and unlikely to capture granular data (Australian Bureau of Statistics, 2013). In this context, researchers often turn to interview-based approaches (Houston-Kolnik & Vasquez, 2022; Vatnar & Bjørkly, 2008). Although valuable, one-on-one interviews may also suffer from selection-bias, sample size issues, and being time-consuming to run (Karystianis et al., 2022).
Against this backdrop, some IPV researchers are turning to secondary analysis of existing data (Australian Bureau of Statistics, 2013). Organisations that interact with victim-survivors – such as police forces or health services – collect large quantities of IPV data which they are unable to analyse manually (Botelle et al., 2022; Karystianis et al., 2022). Additionally, victim-survivors of IPV increasingly make use of online venues such as blogs and bulletin boards to express their experiences of abuse and to receive and offer support (Chu et al., 2021; S. Subramani et al., 2019). These entries generate huge amounts of text data, much of which is publicly accessible.
Computational text mining is a set of techniques which use algorithms to understand, categorise or extract information from unstructured text data (DiMaggio, 2015). These can range from simple (for example, counting the occurrences of a pair of words in a corpus (Homan et al., 2020)) to complex approaches (for example, Deep Learning classifiers which use many layered neural networks to automatically categorise texts (S. Subramani et al., 2019)). Computational text mining methodologies have been used to harness big data to research social phenomena in other domains, such as the study of online hate (Fortuna & Nunes, 2018), cyberbullying (Rosa et al., 2019) and child abuse victimisation (Shahi et al., 2021). Given the intersection between these domains, plus the existing methodological issues in IPV research, computational text mining methodologies offer a promising avenue for the study of IPV.
Research Questions
This article offers a systematic review of existing work which has applied computational text mining to the study of IPV. In doing so, it aims to provide a resource for IPV scholars who may want to use computational text methodologies in their work, providing a starting point to understand current capabilities as well as directions for future research. The article gives an introductory background to text mining methods and techniques, whilst seeking to examine the quality of current work. The authors do not assume existing knowledge of computational methodology, and all terminology will be explained within our article.
Our assessment of the academic literature is driven by three research questions: (RQ1) How have computational text analysis methods been used in IPV research?; (RQ2) What datasets are available for studying IPV using computational text analysis?; (RQ3) How have text analysis methods been evaluated in the study of IPV?
Method
A systematic review of existing academic literature was conducted according to PRISMA-P guidelines (Moher et al., 2015).
Electronic Search Strategy
Eight databases (ACM Digital Library, ArXiv.org, IEEE Xplore, ProQuest, PsychInfo, PubMed, Web of Science, Scopus) were searched, in March 2022, for all records containing both terms relating to computational text mining and terms relating to intimate partner violence, within all fields apart from the full-text (e.g. Title, abstract, keywords, publication venue), and unrestricted by date. The full search string was as followsFootnote 1:
((“artificial intelligence” OR “machine learning” OR “supervised learning” OR “unsupervised learning” OR “automatic detection” OR “automatic recognition” OR “text mining” OR “natural language processing” OR “deep learning” OR “text analysis” OR “information retrieval” OR “information extraction” OR “machine reading” OR “word embeddings” OR “feature extraction” OR “knowledge discovery” OR “data engineering” OR “knowledge engineering” OR “exploratory data analysis” OR “quantitative content analysis” OR “automatic content analysis” OR “computational methods” OR “big data” OR “predictive model”) AND (“intimate partner violence” OR “intimate partner abuse” OR “domestic violence” OR “domestic abuse” OR “family violence” OR “family abuse”))
Inclusion Criteria
Studies were included in the review if they met the following criteria:
-
Peer reviewed and pre-print academic literature;
-
The study uses computational text analysis or text mining to address an IPV-related outcome from a dataset which includes unstructured text fields;
-
The study includes results from at least one dataset (studies which discuss a purely theoretical design or prototype were excluded);
-
The main outcome of the computational model is related to the identification of types, characteristics, prevalence, behaviours and/or opinions of IPV (We excluded studies where IPV is used as an input feature rather than an outcome, for example studies measuring the impact of IPV (input) on mental health (outcome));
-
Since IPV is defined differently in different research, and sometimes is captured within other definitions of violence, we included studies with “family violence” “domestic violence” or “sexual violence” related outcomes, since these may include IPV within their definitions.
Data Extraction and Management
Records identified through database searches were imported into Rayyan (Ouzzani et al., 2016) for data management. After duplicates had been discarded, two of the authors independently performed abstract screening according to the above inclusion criteria. Cohen’s Kappa statistic was calculated at this stage to determine Inter Rater Reliability (IRR) following the procedure described by Hallgren (2012). Cohen’s Kappa was 0.69, indicating a substantial level of agreement between the two reviewers, according to guidelines from Landis and Koch (1977). Remaining disagreements were resolved following a discussion between the two reviewers.
The included papers were subsequently downloaded and a pro-forma was used to extract the information from each paper. The pro-forma was piloted with 16 initial papers and feedback was obtained from other authors, following which amendments were made. The final pro-forma consisted of the following information fields:
Authors; Name of study; Year of study; IPV-related hypothesis or outcome; Source, size and time period of dataset; Demographics of dataset (if discussed); Method and results of labelling dataset; Data pre-processing and cleaning process (if mentioned); Feature selection process (if mentioned); Model task; Types of models tested; Best performing model; Evaluation method; Evaluation metrics used; Best evaluation outcome; Summary of discussion of evaluation outcomes (if any); Summary of interpretability of the model (if discussed); Technologies mentioned; The definition of violence used by the study (if any); Summary of ethical discussion or limitations (if any); Whether any code/datasets are open source.
Quality Assessment
Existing guidelines for assessing bias, quality, and reliability of biomedical or psychological studies are difficult to apply to research using computational text-analysis methods, particularly when reviewing highly specialised systems such as those involving ML. This paper builds on existing frameworks for assessing ML and mixed methods research (Dreisbach et al., 2019; Hinds et al., 2021; Hong et al., 2018; Siebert et al., 2020) to develop a checklist of 21 ‘yes/no’ criteria which were used to assess the overall quality, reliability and potential bias of studies included in the review. A wide range of approaches are surveyed in the included studies, so some irrelevant items were excluded from the checklist depending on the study in question. For that reason, the checklist is not supposed to provide a ranking of studies but an indication of overall quality of the included works. The 21-item checklist was as follows:
-
1.
Definition of violence discussed
-
2.
Clearly described and motivated IPV-related hypothesis or outcome
-
3.
Representativeness/demographics of dataset discussed and/or analysed
-
4.
Source, size, and time period of dataset reported
-
5.
Data cleaning and sampling process reported
-
6.
Discussion of pre-processing techniques
-
7.
Appropriate model used for hypothesis
-
8.
Feature selection discussed and/or different features considered
-
9.
Different models tested and compared
-
10.
Clear and appropriate evaluation criteria
-
11.
Evaluation outcomes reported
-
12.
Evaluation outcomes discussed e.g. comparison to other work, discuss misclassifications
-
13.
Study includes discussion of model interpretability, or clearly explains model rules
-
14.
Includes ethical discussion
-
15.
Source code and/or datasets available
-
16.
Includes discussion of limitations of model and/or appropriate use
-
17.
Dataset is of an appropriate size, and balance of classes discussed
-
18.
Data labelling process is explained
-
19.
Data is labelled according to a protocol by more than one annotator and IAA reported
-
20.
Model is tested on held-out ‘test’ set
-
21.
Model is tested or deployed “in the wild”
Results
Included Studies
As can be seen in the PRISMA chart in Fig. 1, the search yielded 815 results of which 315 were duplicates, leaving 500 unique studies. Of these, 461 were excluded as irrelevant (meaning they did not mention intimate partner abuse and/or use a computational text mining methodology) during abstract screening, leaving 39 papers.
Following full text review, a further three records were excluded because: no full text was available (n = 1); the text was not written in English (n = 1); the paper discussed a purely theoretical approach which did not involve any data (n = 1). Finally, a number of papers (n = 16) were found to report on the same two broad studies, using similar datasets and models. These were the Karystianis et al. papers on the New South Wales Police Force data using a rule-based approach, n = 6 (Adily et al., 2021; Hwang et al., 2020; Karystianis et al., 2019, 2022; Wilson et al., 2021; Withall et al., 2022), and the Poelmans et al. papers on the Amsterdam-Amstelland Police Force Data using an FCA and ESOM based approach, n = 10 (Elzinga, Poelmans, Viaene, & Dedene, 2009; J. Poelmans, Elzinga, & Dedene, 2013; J. Poelmans, Elzinga, Viaene, & Dedene, 2008, 2009; Jonas Poelmans et al., 2010; J Poelmans et al., 2011a, 2011b; J. Poelmans, Elzinga, Viaene, Dedene, & Van Hulle, 2009; J. Poelmans, Elzinga, Viaene, Hulle, et al., 2009a, 2009b, 2009c, 2009d; J. Poelmans, Elzinga, Viaene, Van Hulle, & Dedene, 2009a, 2009b, 2009c, 2009d; J Poelmans et al., 2011a, 2011b)). For simplicity of reporting in this review, these records were condensed into two unique studies. This left N = 22 unique studies to be included in the following qualitative analysis. A summary of the included studies can be found in Table 1.
The N = 22 included studies cover a wide range of research questions and text mining methodologies. Outcomes include extracting topics from a corpus of social media texts (More & Francis, 2021; Rodriguez & Storer, 2020; Xue, Chen, Chen, Hu, & Zhu, 2020; Xue et al., 2019), information retrieval of abuse and injury types from police reports (Adily et al., 2021), detecting the presence or absence of mentions of domestic violence in various types of text (Allen, Davis, & Krishnamurti, 2021; Botelle et al., 2022; Victor et al., 2021), and event and entity recognition from court documents (Li, Sheng, Ge, & Luo, 2019) and victim-survivor narratives (Liu, Li, Liu, Zhang, & Si, 2019). A summary of the studies can be found in Table 1.
The quantity of this research seems to be increasing in recent years, with the majority (n = 18) of studies being published in the last 5 years, and almost a third (n = 7) being published in the last two years. This may be a reflection the increased public awareness of the ‘shadow pandemic’ of domestic abuse brought on by the COVID-19 pandemic (Xue et al., 2020). Given the interdisciplinarity of the topic, it is interesting to note that there was an equal split between studies published in computer science journals and conferencesFootnote 2 (n = 11), and those published in social science and health related venuesFootnote 3 (n = 11).
The following section reviews the included studies as follows: firstly, by giving an overview of the different text mining models and techniques used in the studies; secondly, by reviewing the characteristics of the various datasets which studies used; and finally, by discussing how studies evaluated their techniques and models and what the evaluation outcomes were. This is followed by the Discussion section which investigates the quality of the included studies, offers lessons for researchers hoping to use text mining in their own work, considers ethical concerns of using computational text mining in the study of IPV, and examines the limitations of the current review.
Models and Techniques
Supervised Techniques
Supervised techniques are those that are developed using a labelled dataset – a dataset where each instance has been annotated (labelled) with an outcome or category (for example, each Tweet in a Twitter corpus is manually labelled with either ‘about domestic abuse’ or ‘not about abuse’). These existing annotations can be used as a benchmark to evaluate automatic text mining methods, which makes supervised techniques a popular choice. The majority (n = 16) of the included studies used some kind of supervised approach. Supervised techniques are also the basis for many ML models. Supervised ML models ‘learn’ patterns from the labelled dataset to create an accurate model that can then be applied to new, unseen data (Alpaydin, 2020). This is an extremely convenient way to extend classification tasks to a dataset that is much larger than could be annotated by hand (Botelle et al., 2022).
There are two broad types of Supervised ML models: Traditional and Deep Learning models. Traditional models, such as Support Vector Machines (SVMs), K-Nearest Neighbours (KNN), LASSO Regression, and Decision Trees (DTs) iteratively try to find the best fit for the boundaries between one or more classesFootnote 4 in a high dimensional space—a process commonly referred to as model training. It is beyond the scope of this paper to explain the mechanisms behind these algorithms, but clear introductory explanations can be found in Prabakaran et al. (Prabakaran, Waylan, & Penfold, 2017). In over a third of the included studies (n = 8) a Traditional Supervised ML model was the main, or most successful, approach (Allen et al., 2021; Chu et al., 2021; Garrett & Hassan, 2019; Homan et al., 2020; Schrading, Alm, Ptucha, & Homan, 2015; S. Subramani, Vu, & Wang, 2017; S Subramani, Wang, Islam, Ulhaq, & O’Connor, 2018; Victor et al., 2021), with SVMs being the most common successful model (Garrett & Hassan, 2019; Homan et al., 2020; Schrading et al., 2015; S. Subramani et al., 2017).
Deep Learning models were used as the main approach in six studies (Botelle et al., 2022; Karystianis et al., 2021; Li et al., 2019; Liu et al., 2019; S. Subramani et al., 2019; S. Subramani et al., 2018a, 2018b), often using a traditional ML model as a comparator baseline. Deep Learning models are very large networks of decision nodes – known as neural networks – which discover extremely complex multi-dimensional relationships between input and output (Alpaydin, 2020). Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) are two broad families of Deep Learning models (S. Subramani et al., 2019). Long Short-Term Memory (LSTM) models are an extension of RNNs often used for text classification tasks (S. Subramani et al., 2019).
Transformer based models, such as BERT (Devlin, Chang, Lee, & Toutanova, 2018), are very large deep models that have already learnt a statistical representation of a language (most commonly, English) from huge amount of data. For instance, the original BERT model was trained on a corpus of books and Wikipedia entries of over 3 billion words (Devlin et al., 2018). Since these pre-trained models have a wide ‘understanding’ of language already, they are very adaptable to new tasks, even those where there is little data available. One included study used BioBERT (Lee et al., 2020), an adaptation of the original BERT model specifically suited for biomedical text mining tasks, to identify instances of IPV in Electronic Health Records (Botelle et al., 2022).
Deep Learning models often achieve better results than Traditional ML in complex tasks (Botelle et al., 2022; S. Subramani et al., 2018a, 2018b). However, their drawback is their high level of opacity, which explains why they are frequently referred to as ‘black boxes’. Processes like feature ablationFootnote 5 (Karystianis et al., 2021) and dimensionality reductionFootnote 6 (S. Subramani et al., 2019) can help to visualise and understand the most important factors in the decision of a model. Additionally, recent advances in the domain of explainable machine learning have resulted in tools such as Local Interpretable Model-Agnostic Explanations (LIME) (Ribeiro, Singh, & Guestrin, 2016) which can be used to provide insight into the decision-making mechanisms of Deep Learning models. Nonetheless, their results can still prove difficult to interpret (Karystianis et al., 2021).
The remaining two studies which used a supervised approach used rule-based models to automatically classify data, using existing labels to test the accuracy of their rules (Karystianis et al., 2022; J Poelmans, Van Hulle, et al., 2011). Hand-crafted rule-based models have the advantage of being very transparent and efficient in comparison to ML models. It is probably not a coincidence that the two studies which used this approach were both actively working with police forces, who are likely to value transparency highly. Rule-based models performed very well in both studies (0.89 F1-score for abuse types (Karystianis et al., 2022); Accuracy > 0.89 for identifying domestic violence in police reports (J Poelmans, Van Hulle, et al., 2011)). This suggests that they should not be overlooked in favour of more modern but complex tools such as Deep Learning models.
Unsupervised Techniques
Six studies used unsupervised topic modelling or exploration as their primary approach (More & Francis, 2021; Rodriguez & Storer, 2020; Sanchez-Moya, 2017; Xu, Zeng, Tai, & Hao, 2022; Xue et al., 2020; Xue et al., 2019). Here we use ‘unsupervised’ to mean that a dataset has no labels or annotations—it is simply a collection of instances of raw text data (for example, a collection of Tweets without any categories or labels assigned to each Tweet).
Unsupervised Clustering
Four of the six studies used Unsupervised Machine Learning (Unsupervised ML) models, which analyse the latent structure of a text corpus to identify related clusters, or topics, in a process called topic modelling. The most common topic modelling approach was Latent Dirichlet Allocation (LDA), used in three studies (More & Francis, 2021; Xue et al., 2019, 2020), whilst the other study used Structural Topic Modelling (STM) (Rodriguez & Storer, 2020).
Unsupervised Exploratory Approaches
Two studies used forms of exploratory data analysis as their primary method of investigating text data. Xu et al. (Xu et al., 2022) deployed a custom rule-based approach to sentiment analysis. The latter describes the practice of analysing texts according to their positive or negative emotional tone.
Sanchez-Moya (2017) used Linguistic Inquiry and Word Count (LIWC) (Pennebaker, Francis, & Booth, 2001), a computational tool for linguistic analysis. This technique was also used in four other studies as an addition, or an input into, more complex models (Allen et al., 2021; Rodriguez & Storer, 2020; S Subramani et al., 2018a, 2018b). LIWC is a dictionary-based method, in that it counts the number of words in a text which belong to a series of dictionaries of words from particular linguistic categories (e.g. positive affect, negative affect, biological processes, analytical thinking, emotional tone) (Sanchez-Moya, 2017). Dictionary-based methods are a simple but powerful instrument than can be very efficient, and used across multiple studies, once the hurdle of creating the initial dictionary has been passed. Other included studies created their own dictionaries of IPV-related terms (Adily et al., 2021; Li et al., 2019; J. Poelmans, Elzinga, Viaene, & Dedene, 2009).
Technologies
Matlab, R and Python were mentioned most often as technologies used in the studies, reflecting their popularity for data science applications. At least seven studies mentioned using Python (Chu et al., 2021; Garrett & Hassan, 2019; Homan et al., 2020; More & Francis, 2021; Schrading et al., 2015; Xu et al., 2022; Xue et al., 2019), although many studies did not report any specific technology or programming language used.
Datasets
Source
Most of the datasets used in the included studies were sourced from social media (n = 15) with the remainder coming from police forces (n = 3), health services (n = 1), litigation proceedings (n = 1), children’s social workers (n = 1), and a single study which directly recruited participants (n = 1). A summary of the datasets can be found in Table 2.
As expected from a search conducted in English, most datasets (n = 18) are in English, with the others being in Chinese (n = 3) and Dutch (n = 1). Of those datasets sourced from a particular locality (e.g. police data), the US, UK, Australia, China and the Netherlands are represented. Datasets are notably missing from other countries where English is widely spoken, such as Canada, India, Pakistan, South Africa or Nigeria. Around a quarter of the datasets (n = 6) describe abuse from the perspective of a 3rd party reporting on the abuse (e.g. a police officer or healthcare professional). Conversely, a small number (n = 2) describe abuse from the perspective of the victim-survivor narrating their own experience(s). The remaining datasets (n = 14) contain a mix of perspectives (e.g. social media groups where some posts are from the victim-survivor perspective and some are from 3rd parties describing abuse which happened to someone else or offering support). No datasets explore either text written from the perspective of a perpetrator, or direct evidence of abuse in text (e.g. abusive text messages).
Size
The size of the datasets varies considerably, from 309 diary entries (Allen et al., 2021) to over 1 million unique Tweets (Xue et al., 2020). The size of each text within a dataset also varies, from a single Tweet (Homan et al., 2020) to entire litigation texts (Li et al., 2019) or case summaries (Victor et al., 2021). Of the datasets used for supervised ML tasks, the average size was 73,847 instances.
Labelling Process
Data labelling is often a time consuming and costly part of computational text mining, which can discourage research from taking place in new areas. In addition, data labelling has a direct impact on the outcome of classification models, since any bias or inaccuracies in the labelling process are likely to be picked up and replicated by the model (Bechmann & Zevenbergen, 2019; Dignum, 2017). For this reason, accurate and transparent labelling is of paramount importance, especially in sensitive research.
Most datasets were labelled by supervised student reviewers. However, some datasets took advantage of existing properties of the data to create labels – for example, by using hashtags applied to tweets (Homan et al., 2020), participant surveys administered alongside the collection of text data (Allen et al., 2021), or police assigned labels collected during the incident reporting process (J Poelmans, Van Hulle, et al., 2011). Such techniques can significantly reduce the time and cost burden for researchers and show the benefit of trying to find label-type properties within existing data.
Evaluation
Test and Train Set
A test set is a portion of the dataset that is set aside during model development, and subsequently used to evaluate the algorithm’s final performance on held-out data. Leaving part of the data out during model development helps avoid overfitting, where models learn the statistical characteristics of a dataset “too well”, in a way that means their results don’t generalise to other data (Arango, Pérez, & Poblete, 2019). For small datasets, a mechanism called k-fold Cross Validation (k-fold CV) is often used to evaluate a model’s performance, in combination with or instead of a separate test set. This involves separating the data into k different segments. The model is then allowed to see all but one of these segments when it is training, and after training has finished, the left-out segment is used to test the model. The process is then repeated k times, each time leaving out a different segment. The results of these k times are then averaged to give an overall evaluation metric.
Evaluation Metrics
All studies using supervised techniques were evaluated using a test set or k-fold CV. Accuracy and F1 score were the most common metrics used to report how well the model performed at correctly categorising the texts. Accuracy refers to the overall percentage of instances which were correctly classified. The F1 score is an alternative metric which balances Precision (also known as specificity, or true negative rate) and Recall (also known as sensitivity, or true positive rate). The F1 score is useful in situations where one class is much larger than another – in this case, Accuracy scores can be unhelpfully biased towards the dominant class (Rosa et al., 2019).
However, comparison of models across different datasets using reported metrics should be done cautiously, since much of the performance of a model depends on the data it was trained on. Some datasets simply have too much overlap between the characteristics of different classes, making it difficult for a model to distinguish between them.
Taking into account these comments on the limitations of metrics, there is a very wide range of accuracies in the studies, from 0.69 (which would usually be considered too low to be used in any practical application) (Karystianis et al., 2021) to 0.97 (as good of a performance as can reasonably be expected from most models) (Botelle et al., 2022). There was no single type of model or technique which performed well across the studies. This reflects the variability of model tasks within the studies and demonstrates the importance of choosing the right model for the task in question.
Unsupervised Evaluation
Evaluation of the studies which used unsupervised approaches was much more variable, reflecting the difficulties in evaluating unsupervised methods more broadly (Zhao et al., 2015). Some unsupervised studies did not include any explicit evaluation of their technique (Xu et al., 2022) or were using tools developed and tested in previous research (such as LIWC (Sanchez-Moya, 2017)). Other studies which used unsupervised topic modelling attempted to evaluate the optimal number of topics, using methods such as Rate of Perplexity Change (RPC) (Xue et al., 2019).
Discussion
Overall, the N = 22 studies showcase different models and techniques which can be used for IPV research, as well as a variety of datasets and evaluation mechanisms. The quality of studies varied considerably across the included works—full results from the Quality Assessment (i.e. 21 ‘yes/no’ criteria) are reported in Table 3. This variation in quality reflects the innovative nature of this new, interdisciplinary area. There are not yet clear guidelines about how to use text mining methodologies in social science research. In addition, challenges arise when attempting to assess quality across such a heterogeneous set of studies. For example, some papers did not report any pre-processing steps (Criteria 6) since this is not useful in Deep Learning architectures (S. Subramani et al., 2018a, 2018b). Other studies did not report demographic characteristics of their dataset (Criteria 3) due to ethical concerns about collecting personal identifiers (Rodriguez & Storer, 2020; Xue et al., 2019).
The following section provides a more detailed discussion of the reviewed studies, focusing on lessons learned for future research, and issues of ethics and bias raised by using computational methods to research IPV.
Lessons for Future Research
Examining aspects of the included studies offers lessons for future research, particularly regarding the definition of violence, open source code, and overall study design. These issues are discussed in more depth below.
Definition of Violence
The definition of violence is mentioned in just over half the studies (n = 13), but many do not define IPV at all, or very briefly reference a definition from another entity, such as the WHO (Chu et al., 2021). Studies tend to discuss the definition of violence in most detail when examining the dataset labelling process for supervised techniques. Labelling data often highlights conflicting definitions between annotators and necessitates a more in-depth description of what constitutes violence (Botelle et al., 2022; J. Poelmans, Elzinga, Viaene, & Dedene, 2009). Considering wider difficulties defining IPV within research (Alhabib et al., 2010; Barocas et al., 2016), future researchers should ensure they carefully describe and motivate the specific definition of IPV used in their work.
Open Source
Unfortunately, no projects in the study reported that their code was open source. The latter describes a trend in computer science to make code and data available freely online, to facilitate collaborators wishing to build similar applications.Footnote 7 Only two projects mentioned that their dataset would be made available upon request (Botelle et al., 2022; Xu et al., 2022). This is perhaps unsurprising when it comes to datasets, given the sensitive nature of the data involved. However, future work could consider making source code available for other researchers, to encourage knowledge-sharing within this field.
Study Design
In general, future projects could consider a number of factors in study design. Firstly, researchers may reflect where novel data can be sourced, and whether data from multiple sources can be joined-up for additional insight (Karystianis et al., 2021). Secondly, once a model has been developed, researchers could consider deploying or testing it in an active service-provision environment. For example, research projects from Poelmans et al. (2013) and Karystianis et al. (2022) successfully worked with police forces to implement knowledge-discovery techniques within their day-to-day operations, and models revealed edge cases of abuse that the police had previously missed (Hwang et al., 2020; J. Poelmans, Elzinga, Viaene, & Dedene, 2009). A project to detect sexual and physical domestic violence in Electronic Health Records is now live on systems of an NHS trust in the UK (Botelle et al., 2022).
Moreover, when designing methodologies, researchers must consider more than just the choice of model. Rule-based, Traditional ML, Deep Learning and Unsupervised approaches all performed well in different included studies, demonstrating that the context and appropriateness of a model is more important than its type. The importance of initial data exploration and feature selection should not be ignored, as these processes (referred to as feature engineeringFootnote 8) significantly increase the quality of outcomes. For example, Subramani et al. (2017) did not use the raw text, but instead the outcome of LIWC (see Unsupervised Exploratory Approaches, above), as the input to their ML model (S. Subramani et al., 2017). Finally, several studies highlighted the importance of mixed methods in their research, and the significance of pairing quantitative methods with qualitative insights (Rodriguez & Storer, 2020; Victor et al., 2021).
Ethical Concerns and Bias
Ethics and Context
In general, little attention was paid to ethics across the studies, with only six publications including an explicit ethical discussion. However, a large number (n = 14) of studies do mention limitations of their work or discuss appropriate contexts for model use. For example, Victor et al. indicate that whilst their model performs well enough to be used for generating accurate descriptive statistics about domestic violence in a dataset of child welfare case summaries, it would be inappropriate for use in decision making about individual cases (Victor et al., 2021). They highlight the importance of qualitative analysis when using ML methods in an interdisciplinary context, giving three examples of how qualitative analysis can enrich ML research in this domain: understanding the data-generating mechanism, its context, content and what inferences can reasonably be made; understanding outliers and misclassifications in order to improve the model; and applying insights from the model to help standardize the assessment or documentation of abuse (Victor et al., 2021).
Bias
Allen et al. comment on the lack of diversity in their sample, which contained mostly white participants (Allen et al., 2021). Since non-white groups may be more likely to experience IPV (Breiding, Chen & Black, 2014), this lack of diversity is especially troubling. However, very few studies commented on the demographic representativeness of their dataset and whether downstream applications built on their models risked bias towards certain groups.
Future Work
Given the recent emphasis within ML communities on ethical principles of accountability, responsibility and transparency (Dignum, 2017; Floridi et al., 2018), future work must take more of a focus on discussing the foundational ethical questions raised by this kind of research. Researchers might consider following ethical guidelines for ML such as those proposed by the Association of Internet Researchers (Bechmann & Zevenbergen, 2019). The consequences of ignoring such ethical discussions are significant: At their worst, ML models could contribute to the invalidation and minimisation of different experiences of abuse, for example by classifying an instance as ‘not abuse’ and leading to a victim-survivor not receiving services or justice after having experienced great harm (Blackwell, Dimond, Schoenebeck, & Lampe, 2017). Victim-survivors of IPV have experienced situations in which they have had their opinions and experiences repeatedly invalidated, belittled, denied and manipulated (Stark, 2009). Researchers must be aware of the potential mis-use of their research to extend this denial of the victim-survivor’s reality. Models are representations of reality, but they are not reality themselves, and the way text mining research is conducted and presented should reflect this understanding.
Limitations
The current work is subject to several limitations. Firstly, since the search strategy only included academic literature, it is possible that important grey literature may have been missed. Secondly, the search terms included other types of violence such as “family violence” and “sexual violence”, aiming to capture all definitions of violence that may include IPV. Some of the reviewed studies may therefore have included incidents of non-partner abuse in their data. Finally, the Quality Assessment criteria used in the review were developed by combining multiple existing methods and were not thoroughly evaluated on different types of studies outside this review. They should therefore not be used as a ranking mechanism or to draw concrete conclusions about the quality of individual studies.
Conclusion
Twenty-two studies which used computational text mining to investigate IPV were identified through a systematic literature review of eight academic databases. The studies included datasets from social media, police forces, a healthcare provider, and social work and legal settings. A variety of supervised and unsupervised text mining techniques were used on these datasets for tasks which included detecting the presence or absence of IPV as well as identifying abuse types, extracting entities and events, or understanding themes. Some studies commented on the ethics or real-world deployment of their findings, but future research could include more in-depth discussion of these. Additionally, potential areas for future work may include sourcing datasets from other geographies and types of organisations, explorations into sub-types of abuse, plus the application of emerging text-mining methods in the IPV field as they develop.
Notes
NB The search string was adapted to fit the search functions of different databases. In ArXiv.org, only the IPV-related part of the search string was used since all research on ArXiv.org was assumed to have a computational element and the search function did not allow for so many search terms.
e.g. NAACL, IEEE Transactions, Databases Theory and Applications.
e.g. Violence Against Women, Journal of Interpersonal Violence, Journal of Medical Internet Research.
In ML literature, class refers to an outcome, category, or label that a model is trying to optimise for. For example, if a model was being built to automatically categorise (or classify) Electronic Health Records as to whether or not they contained a mention of domestic abuse, the two classes would be “abuse present” and “abuse absent”.
Feature ablation refers to removing one or more inputs (features) and observing the change in model performance in order to understand how different features affect the decision of a model.
Dimensionality Reduction refers to the mathematical process of transforming a very high-dimensional space into a space with fewer dimensions, whilst preserving important characteristics of the data. This can help to visualise clusters or decision boundaries within a complex model.
This can be done using free services such as GitHub (https://github.com), GitLab (https://about.gitlab.com), Zenodo (https://zenodo.org) or Dataverse (https://dataverse.org).
Feature engineering is the set of steps that transform raw data into numeric values that are usable by ML models. These numbers are features that represent each instance/sample (e.g. an abusive sentence or a non-abusive one) and are used as inputs by the ML models.
References
*Indicate studies that were included in the Systematic Review
*Adily, A., Karystianis, G., & Butler, T. (2021). Text mining police narratives to identify types of abuse and victim injuries in family and domestic violence events. Trends And Issues In Crime And Criminal Justice (630), 1-12.
Alhabib, S., Nur, U., & Jones, R. (2010). Domestic violence against women: Systematic review of prevalence studies. Journal of Family Violence, 25(4), 369–382.
*Allen, K., Davis, A. L., & Krishnamurti, T. (2021). Indirect Identification of Perinatal Psychosocial Risks from Natural Language. IEEE Transactions on Affective Computing.
Alpaydin, E. (2020). Introduction to machine learning: MIT press.
Arango, A., Pérez, J., & Poblete, B. (2019). Hate Speech Detection is Not as Easy as You May Think: A Closer Look at Model Validation. Paper presented at the Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France. . https://doi.org/10.1145/3331184.3331262
Australian Bureau of Statistics. (2013). Bridging the data gaps for family, domestic and sexual violence. Retrieved from https://www.abs.gov.au/statistics/people/crime-and-justice/bridging-data-gaps-family-domestic-and-sexual-violence/latest-release#improving-the-evidence-base. Accessed July 2022.
Barocas, B., Emery, D., & Mills, L. G. (2016). Changing the domestic violence narrative: Aligning definitions and standards. Journal of Family Violence, 31(8), 941–947.
Bechmann, A., & Zevenbergen, B. (2019). AI and machine learning: Internet research ethics guidelines (IRE 3.0 6.1). Internet Research: Ethical Guidelines 3.0 Association of Internet Researchers Unanimously Approved by the AoIR Membership October 6, 33.
Blackwell, L., Dimond, J., Schoenebeck, S., & Lampe, C. (2017). Classification and its consequences for online harassment: Design insights from heartmob. Proceedings of the ACM on Human-Computer Interaction, 1(CSCW), 1–19.
*Botelle, R., Bhavsar, V., Kadra-Scalzo, G., Mascio, A., Williams, M. V., Roberts, A., . . . Stewart, R. (2022). Can natural language processing models extract and classify instances of interpersonal violence in mental healthcare electronic records: an applied evaluative study. BMJ Open, 12(2), e052911. https://doi.org/10.1136/bmjopen-2021-052911
Breiding, M. J., Chen, J., & Black, M. C. (2014). Intimate partner violence in the United States–201. National Center for Injury Prevention and Control, Centers for Disease Control and Prevention.
*Chu, T., Su, Y., Kong, H., Shi, J., & Wang, X. (2021). Online Social Support for Intimate Partner Violence Victims in China: Quantitative and Automatic Content Analysis. Violence against women, 27(3), 339-358.
Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
Dignum, V. (2017). Responsible artificial intelligence: designing AI for human values.
DiMaggio, P. (2015). Adapting computational text analysis to social science (and vice versa). Big Data & Society, 2(2), 2053951715602908.
Dokkedahl, S., Kok, R. N., Murphy, S., Kristensen, T. R., Bech-Hansen, D., & Elklit, A. (2019). The psychological subtype of intimate partner violence and its effect on mental health: Protocol for a systematic review and meta-analysis. Systematic Reviews, 8(1), 1–10.
Dreisbach, C., Koleck, T. A., Bourne, P. E., & Bakken, S. (2019). A systematic review of natural language processing and text mining of symptoms from electronic patient-authored text data. International Journal of Medical Informatics, 125, 37–46.
*Elzinga, P., Poelmans, J., Viaene, S., & Dedene, G. (2009). Detecting Domestic Violence. In Proceedings of the 11th International Conference on Enterprise Information Systems, Volume AIDSS (pp. 11–18). Springer.
European Union Agency for Fundamental Rights. (2014). Violence against women: An EU-wide survey. Publications Office of the European Union.
Evans, J. A., & Aceves, P. (2016). Machine translation: Mining text for social theory. Annual Review of Sociology, 42, 21–50.
Ferreira-Mello, R., André, M., Pinheiro, A., Costa, E., & Romero, C. (2019). Text mining in education. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 9(6), e1332.
Floridi, L., Cowls, J., Beltrametti, M., Chatila, R., Chazerand, P., Dignum, V., & Rossi, F. (2018). AI4People-an ethical framework for a good AI society: opportunities, risks, principles, and recommendations. Minds and Machines, 28(4), 689–707.
Fortuna, P., & Nunes, S. (2018). A Survey on Automatic Detection of Hate Speech in Text. ACM Computing Surveys, 51(4), 85. https://doi.org/10.1145/3232676
Gauthier, R. P., & Wallace, J. R. (2022). The computational thematic analysis toolkit. Proceedings of the ACM on Human-Computer Interaction, 6(GROUP), 1–15.
*Garrett, A., & Hassan, N. (2019, August). Understanding the silence of sexual harassment victims through the# whyididntreport movement. Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (pp. 649–652).
Hallgren, K. A. (2012). Computing inter-rater reliability for observational data: An overview and tutorial. Tutorials in Quantitative Methods for Psychology, 8(1), 23.
Hinds, J., Parkhouse, T., & Hotchin, V. (2021). Assessing the quality of studies using machine learning for personality assessment: A systematic review.
*Homan, C., Schrading, J., Ptucha, R., Cerulli, C., & Alm, C. (2020). Quantitative methods for analyzing intimate partner violence in microblogs: Observational study. Journal of Medical Internet Research, 22(11), e15347.
Hong, Q. N., Fàbregues, S., Bartlett, G., Boardman, F., Cargo, M., Dagenais, P., & O’Cathain, A. (2018). The Mixed Methods Appraisal Tool (MMAT) version 2018 for information professionals and researchers. Education for information, 34(4), 285–291.
Houston-Kolnik, J. D., & Vasquez, A. L. (2022). Cognitive Interviewing: Lessons Learned and Recommendations for Structured Interviews with Survivors of Crime. Journal of Family Violence, 37(2), 325–335.
*Hwang, Y. I., Zheng, L., Karystianis, G., Gibbs, V., Sharp, K., & Butler, T. (2020). Domestic violence events involving autism: a text mining study of police records in New South Wales, 2005–2016. Research in Autism Spectrum Disorders, 78, 101634.
*Karystianis, G., Adily, A., Schofield, P. W., Greenberg, D., Jorm, L., Nenadic, G., & Butler, T. (2019). Automated analysis of domestic violence police reports to explore abuse types and victim injuries: Text mining study. Journal of Medical Internet Research, 21(3), e13067.
*Karystianis, G., Adily, A., Schofield, P. W., Wand, H., Lukmanjaya, W., Buchan, I., Nenadic, G., & Butler, T. (2022). Surveillance of domestic violence using text mining outputs from Australian police records. Frontiers in Psychiatry, 12, 2412.
*Karystianis, G., Cabral, R. C., Han, S. C., Poon, J., & Butler, T. (2021). Utilizing text mining, data linkage and deep learning in police and health records to predict future offenses in family and domestic violence. Frontiers in Digital Health, 3, 602683.
Lagdon, S., Jordan, J.-A., Devine, P., Tully, M. A., Armour, C., & Shannon, C. (2022). Public understanding of coercive control in Northern Ireland. Journal of Family Violence, 38(1), 39–50.
Landis, J. R., & Koch, G. G. (1977). An application of hierarchical kappa-type statistics in the assessment of majority agreement among multiple observers. Biometrics, 33, 363–374.
Lee, J., Yoon, W., Kim, S., Kim, D., Kim, S., So, C. H., & Kang, J. (2020). BioBERT: A pre-trained biomedical language representation model for biomedical text mining. Bioinformatics, 36(4), 1234–1240.
*Li, C., Sheng, Y., Ge, J., & Luo, B. (2019). Apply event extraction techniques to the judicial field. In Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing (pp. 492–497).
*Liu, Y., Li, Q., Liu, X., Zhang, Q., & Si, L. (2019). Sexual Harassment Story Classification and Key Information Identification. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management (pp. 2385–2388).
Luque, C., Luna, J. M., Luque, M., & Ventura, S. (2019). An advanced review on text mining in medicine. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 9(3), e1302.
Moher, D., Shamseer, L., Clarke, M., Ghersi, D., Liberati, A., Petticrew, M., & Stewart, L. A. (2015). Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Systematic Reviews, 4(1), 1–9.
*More, K., & Francis, F. (2021). Analyzing the Impact of Domestic Violence on Social Media using Natural Language Processing. In 2021 IEEE Pune Section International Conference (PuneCon) (pp. 1–5). IEEE.
ONS. (2020). Domestic abuse victim characteristics, England and Wales: year ending March 2020.
Ouzzani, M., Hammady, H., Fedorowicz, Z., & Elmagarmid, A. (2016). Rayyan—a web and mobile app for systematic reviews. Systematic Reviews, 5(1), 1–10.
Øverlien, C., Hellevik, P. M., & Korkmaz, S. (2020). Young women’s experiences of intimate partner violence–narratives of control, terror, and resistance. Journal of Family Violence, 35(8), 803–814.
Pennebaker, J. W., Francis, M. E., & Booth, R. J. (2001). Linguistic inquiry and word count: LIWC 2001. Mahway: Lawrence Erlbaum Associates, 71(2001), 2001.
*Poelmans, J., Elzinga, P., Viaene, S., & Dedene, G. (2008). An exploration into the power of formal concept analysis for domestic violence analysis. In Industrial Conference on Data Mining (pp. 404–416). Springer.
*Poelmans, J., Elzinga, P., Viaene, S., & Dedene, G. (2009a). A case of using formal concept analysis in combination with emergent self organizing maps for detecting domestic violence. In Industrial conference on data mining (pp. 247–260). Springer.
*Poelmans, J., Elzinga, P., Viaene, S., Dedene, G., & Van Hulle, M. M. (2009b). Analyzing domestic violence with topographic maps: A comparative study. In International Workshop on Self-Organizing Maps (pp. 246–254). Springer.
*Poelmans, J., Elzinga, P., Viaene, S., Hulle, M. M. V., & Dedene, G. (2009c). How Emergent Self Organizing Maps Can Help Counter Domestic Violence. In 2009c WRI World Congress on Computer Science and Information Engineering (Vol. 4, pp. 126–136). IEEE.
*Poelmans, J., Elzinga, P., Viaene, S., Van Hulle, M. M., & Dedene, G. (2009d). Gaining insight in domestic violence with Emergent Self Organizing Maps. Expert systems with applications, 36(9), 11864-11874.
*Poelmans, J., Elzinga, P., Viaene, S., & Dedene, G. (2010). Curbing domestic violence: instantiating C-K theory with formal concept analysis and emergent self-organizing maps. Intelligent Systems in Accounting, Finance and Management, 17(3), 167.
*Poelmans, J., Elzinga, P., Viaene, S., & Dedene, G. (2011a). Formally analysing the concepts of domestic violence. Expert systems with applications, 38(4), 3116-3130.
*Poelmans, J., Van Hulle, M., Viaene, S., Elzinga, P., & Dedene, G. (2011b). Text mining with emergent self organizing maps and multi-dimensional scaling: A comparative study on domestic violence. Applied soft computing, 11(4), 3870-3876.
*Poelmans, J., Elzinga, P., & Dedene, G. (2013). Retrieval of criminal trajectories with an FCA-based approach. In Proceedings of the FCAIR 2013 Formal Concept Analysis meets Information Retrieval workshop (Vol. 977, pp. 83–94).
Prabakaran, S., Waylan, M., & Penfold, C. (2017). An Introduction to Machine Learning. Retrieved from https://bioinformatics-training.github.io/intro-machine-learning-2017/. Accessed 9th December 2021.
Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). Model-agnostic interpretability of machine learning. arXiv preprint arXiv:1606.05386.
*Rodriguez, M. Y., & Storer, H. (2020). A computational social science perspective on qualitative data exploration: Using topic models for the descriptive analysis of social media data. Journal of Technology in Human Services, 38(1), 54-86.
Rogers, M., Rumley, T., & Lovatt, G. (2019). The change up project: Using social norming theory with young people to address domestic abuse and promote healthy relationships. Journal of Family Violence, 34(6), 507–519.
Rosa, H., Pereira, N., Ribeiro, R., Ferreira, P. C., Carvalho, J. P., Oliveira, S., & Trancoso, I. (2019). Automatic cyberbullying detection: A systematic review. Computers in Human Behavior, 93, 333–345.
*Sanchez-Moya, A. (2017). Corpus-driven insights into the discourse of women survivors of Intimate Partner Violence. Quaderns De Filologia-Estudis Linguistics, 22, 215-243.
*Schrading, N., Alm, C. O., Ptucha, R., & Homan, C. M. (2015). An analysis of domestic abuse discourse on reddit. In Proceedings of the 2015 conference on empirical methods in natural language processing (pp. 2577–2583).
Shahi, N., Shahi, A. K., Phillips, R., Shirek, G., Lindberg, D. M., & Moulton, S. L. (2021). Using deep learning and natural language processing models to detect child physical abuse. Journal of Pediatric Surgery, 56(12), 2326–2332.
Siebert, J., Joeckel, L., Heidrich, J., Nakamichi, K., Ohashi, K., Namba, I., & Aoyama, M. (2020). Towards guidelines for assessing qualities of machine learning systems. Paper presented at the International Conference on the Quality of Information and Communications Technology.
Stark, E. (2009). Coercive control: The entrapment of women in personal life. Oxford University Press.
*Subramani, S., Vu, H. Q., & Wang, H. (2017). Intent Classification Using Feature Sets for Domestic Violence Discourse on Social Media. In 4th Asia-Pacific World Congress on Computer Science and Engineering (pp. 129–136). IEEE.
*Subramani, S., Wang, H., Islam, M., Ulhaq, A., & O’Connor, M. (2018a). Child Abuse and Domestic Abuse: Content and Feature Analysis from Social Media Disclosures. In Australasian Database Conference (pp. 174–185). Springer, Cham.
*Subramani, S., Wang, H., Vu, H. Q., & Li, G. (2018b). Domestic Violence Crisis Identification From Facebook Posts Based on Deep Learning. IEEE Access, 6, 54075-54085.
*Subramani, S., Michalska, S., Wang, H., Du, J., Zhang, Y., & Shakeel, H. (2019). Deep Learning for Multi-Class Identification From Domestic Violence Online Posts. IEEE Access, 7, 46210-46224.
Vatnar, S. K. B., & Bjørkly, S. (2008). An interactional perspective of intimate partner violence: An in-depth semi-structured interview of a representative sample of help-seeking women. Journal of Family Violence, 23(4), 265–279.
*Victor, B. G., Perron, B. E., Sokol, R. L., Fedina, L., & Ryan, J. P. (2021). Automated identification of domestic violence in written child welfare records: Leveraging text mining and machine learning to enhance social work research and evaluation. Journal of the Society for Social Work and Research, 12(4), 631–655.
Walby, S., Towers, J., Balderston, S., Corradi, C., Francis, B., Heiskanen, M., & Palmer, E. (2017). The concept and measurement of violence against women and men. Policy Press.
*Wilson, M., Spike, E., Karystianis, G., & Butler, T. (2022). Nonfatal strangulation during domestic violence events in new south wales: Prevalence and characteristics using text mining study of police narratives. Violence Against Women, 28(10), 2259–2285.
*Withall, A., Karystianis, G., Duncan, D., Hwang, Y. I., Kidane, A. H., & Butler, T. (2022). Domestic Violence in Residential Care Facilities in New South Wales, Australia: A Text Mining Study. Gerontologist, 62(2), 223-231. doi:https://doi.org/10.1093/geront/gnab068
Wood, L., Backes, B., Baumler, E., & McGiffert, M. (2021). Examining the impact of duration, connection, and dosage of domestic violence services on survivor well-being. Journal of Family Violence, 37(2), 221–233.
World Health Organisation. (2021). Violence against women: Key facts. Retrieved from https://www.who.int/news-room/fact-sheets/detail/violence-against-women. Accessed 9 Dec 2021.
*Xu, H., Zeng, J., Tai, Z., & Hao, H. (2022). Public attention and sentiment toward intimate partner violence based on Weibo in China: A text mining approach. Healthcare (Basel), 10(2), 198. https://doi.org/10.3390/healthcare10020198
*Xue, J., Chen, J., & Gelles, R. (2019). Using Data Mining Techniques to Examine Domestic Violence Topics on Twitter. Violence and gender, 6(2), 105-114.
*Xue, J., Chen, J., Chen, C., Hu, R., & Zhu, T. (2020). The hidden pandemic of family violence during COVID-19: Unsupervised learning of tweets. Journal of Medical Internet Research, 22(11), e24361.
Zhao, W., Chen, J. J., Perkins, R., Liu, Z., Ge, W., Ding, Y., & Zou, W. (2015). A heuristic approach to determine an appropriate number of topics in topic modeling. Paper presented at the BMC bioinformatics.
Funding
Whilst conducting and at the time of submitting this research, Lilly Neubauer was funded by an EPSRC studentship in Cybersecurity, grant number EP/S022503/1. Dr Leonie Maria Tanczer’s salary was supported by a variety of research grants, including EP/S035869/1 (SPRITE+), EP/V011189/1 (REPHRAIN), ES/S004424/1 (VAMHN), MR-VO49879/1 (VISION), and MR/W009692/1 (UKRI FLF). Isabel Straw was funded by a scholarship from UK Research and Innovation, grant number EP/S021612/.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interest
The authors declare that they have no conflict of interest.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Neubauer, L., Straw, I., Mariconti, E. et al. A Systematic Literature Review of the Use of Computational Text Analysis Methods in Intimate Partner Violence Research. J Fam Viol 38, 1205–1224 (2023). https://doi.org/10.1007/s10896-023-00517-7
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10896-023-00517-7