Keywords

1 Introduction

The evolution of the information society has brought great changes to the structure of crime, as traditional contact crimes keep declining, the new paradigm crime represented by telecom network fraud increases sharply in China. The average growing rate of telecom network fraud has reached 34% over the years [1], which has become the most prominent crime reported by the citizens. SMS is an important first-contact channel in many telecom network fraud cases, especially in the Loan Fraud, Part-time Fraud and Investing Fraud, etc. As shown in Fig. 1, if victims perform corresponding operations with fraud SMS, such as replying SMS or adding QQ, WeChat (the top 2 popular social media software in China) mentioned in the texts, the criminals would make further contact with victims in other communication channels where the real identities are difficult to be traced. Therefore, the identification and prevention of fraud SMS has become a crucial issue.

Fig. 1.
figure 1

Communication channel under different fraud procedures.

Despite of some studies that have investigated detecting spam text already, these methods can not be applied in Chinese fraud SMS classification problem ideally due to the following reasons:

  • Unlike the giant scale of normal SMS, the number of fraud SMS is relatively small. According to [2] and [3], the number of Smishing (SMS Phishing) for telecom network fraud only accounts for 0.05% of the whole SMS in China during 2020. It's important to design an effective method to filter out numerous normal SMS before the fraud SMS detection. Such scenarios are quite common and important for telecom operators, which are required to curb the fraud SMS by regulatory departments.

  • Compared with English which has a relatively small alphabet, Chinese is logographic with a large set of characters. Besides, the battle between criminals and polices is extremely fierce, and the attack methods in fraud SMS, which has huge difference with English could be updated quickly in several hours. Typical attack methods contain Word Split (split a character into several characters), Homophony (different words or characters have the same PINYIN [4]) and Glyph Similar (obfuscated characters have similar structures with the original characters), etc.

  • Differing from the traditional spam text (such as Gambling, Porn, Abuse, etc.) and Smishing (which mainly aims at collecting sensitive data from people), the domain-knowledge plays an important role in the identification of the specific type of fraud SMS. For example, there are lots of zeros in the Loan fraud SMS, but the SMS with many zeros could be the normal SMS of banking notice as well. The fraud domain-knowledge related studies are very limited in the previous work.

To tackle these challenging problems, we firstly conduct a measurement study to obtain a better understanding about the fraud SMS, then we design a novel two-stage algorithm called TFC to classify the Chinese fraud SMS. In the first stage, we utilize the key insights found in our measurement to quickly filter out the normal SMS. In the second stage, we exploit BERT-wwm [5] to obtain the semantic features, and thoughtfully designed 61-dim features base on the fraud domain-knowledge to capture the various attack methods and different kinds of abnormal SMS characteristics. By conducting extensive experiments with real-world SMS dataset, we prove that TFC is able to achieve the best performance compared with five alternative algorithms, and the generated features are robust to many kinds of classification algorithms in the meanwhile.

The main contributions of this paper can be summarized in three aspects:

  • We conduct a comprehensive analysis on the labeled telecom network fraud SMS and discover some interesting insights. To mitigate the problem of lack of Chinese fraud SMS datasets, we release part of our labeled dataset online [6] for research convenience.

  • We propose a novel two-stage algorithm to address the fraud SMS classification problem. The algorithm can rapidly filter out normal SMS in Stage 1. After composing deep features and domain-knowledge related features in stage 2, five types of SMS (namely, normal SMS, Loan fraud SMS, Part-time fraud SMS, Investing fraud SMS and Gambling SMS) can be effectively classified with simple machine learning algorithm.

  • We demonstrate our algorithm's performance through real-world data driven experiments. Compared with various algorithms, our algorithm is able to obtain the highest F1-score and accuracy. Impacts of the Stage 1 and domain-knowledge keywords, and the performance of different ML classifiers with TFC are also included.

The rest of the paper is organized as follows. In Sect. 2, we review the previous work of SMS fraud and spam text related problems. In Sect. 3, we conduct a comprehensive measurement analysis with fraud SMS and obtain several interesting insights. We introduce our two-stage algorithm and detailed feature construction procedures in Sect. 4. Section 5 presents the results of experiments using real-world data. Section 6 concludes the paper.

2 Related Work

SMS fraud related problems have gained lots of attention for many years. Joo et al. [7] proposed an enhanced security model for detecting Smishing attack in smart devices. Goel et al. [8] designed a algorithm framework consist of three phases for Smishing attack detection. Mishra et al. [9] investigated Smishing detection problem through SMS Content, URL and Source Code, then a prototype of the proposed system with 96.29% overall accuracy is developed. Pervaiz et al. [10] investigate the scope and scale of the problem of SMS fraud in Pakistan, Delany et al. [11] and Abdulhamid et al. [12] reviewed and summarized the SMS spam filtering method. As for the fraud detection in the telecom area, previous works mainly use Call Detail Record (CDR) to identity the fraudulent user. For example, Olszewski [13] proposed an approach based on user profiling with Latent Dirichled Allocation (LDA), and detecting fraudulent behavior on the basis of threshold type classification with use of the KL-divergence. Farvaresh et al. [14] employed a hybrid approach consisting of preprocessing, clustering and classification phases to identifying customers’ subscription fraud. The studies of SMS fraud from telecom operator's perspective are very limited.

There are several attack methods in spam text, such as adversarial text, wrong-spell words etc. Another thread of research studied the related problem from different aspects. Wang et al. [15] proposed an adversary-generation algorithm for increasing the variance of adversarial training data; Ebrahimi et al. [16] investigated the robustness of models with adversarial examples trained by the proposed two translation-specific types of attacks. Li et al. [17] presented an adversarial defense framework designed for Chinese based deep learning model, all the results of experiments show that these methods could improve model performance to defend against adversarial text attacks. Yeh et al. [18] presented a novel spelling error detection and correction method based on N-gram ranked inverted index; Xiong et al. [19] designed a unified framework HANSpeller for Chinese text spelling error detection and correction. Karan et al. [20] and Nobata et al. [21] investigated the abusive language detection problem with Cross-Domain method and NLP features based classification methodology separately. All these works focused on the characteristics of the spam text, and the differences between spam text and fraud SMS are not considered.

Differing from the previous work, our work focuses on two aspects. Firstly, we aim at addressing the fraud SMS classification problem from telecom operator’s perspective, rather than identifying the fraud with CDR [13, 14] or detecting Smishing from the view of mobile devices [7,8,9]. Secondly, while previous works focus on English [15, 16] and Chinese [17,18,19] spam text related studies, our work studied the Chinese fraud SMS classification problem which can not be applied by many proposed methods in other languages due to the huge Chinese grammatical difference and the unique characteristics between fraud SMS and spam text.

3 Measurement Analysis

In this section, we mainly conduct the measurement analysis with fraud SMS dataset, which are reported by massive citizens and collected by the Guangzhou Anti-Fraud Center from 11/02/2020 to 11/11/2020, named RD1. Figure 2(a) plots the fraud SMS daily count over a week, and we can observe that the average fraud SMS count is larger in weekdays compared to the weekends. As shown in Fig. 2(b) (We use ‘#’ as the abbreviation of word ‘number’ in this paper), in terms of the fraud SMS types, Part-time fraud, Loan fraud and Investing fraud accounted for 85% of the whole RD1, and the Part-time fraud SMS has the highest proportion, nearly 50%.

Fig. 2.
figure 2

(a) Daily fraud SMS count of RD1. (b) # of fraud samples percentage.

Table 1. Examples of attack methods and corresponding # of Uncommon Chinese Characters and Other Characters

Before stepping into further analysis, we define five types of characters in the SMS as follows:

  • Common Chinese Characters. Define the common words and sub-common words in the List of Commonly Used Characters in Modern Chinese [22] as Common Chinese Characters, and the dict size is 3500.

  • Uncommon Chinese Characters. Define Chinese characters that not belong to the Common Chinese Characters as Uncommon Chinese Characters.

  • Numeric Characters. Define the number characters from 0 to 9 as Numeric Characters.

  • English Characters. Define the alphabet characters from a to z and their upper case A to Z as English Characters.

  • Other Characters. Define the characters in SMS that do not falling into 4 types mentioned above are Other Characters, such as symbolic characters like $ ! #, etc..

After we dive into the RD1, we obtained two interesting insights. One is that Uncommon Chinese Characters and Other Characters often appear in the fraud SMS. As shown in Fig. 3, to pass the keywords based spam detection methods and systems, the criminals tend to modify the SMS with several adversarial attacks, such as Word Split, Word Homophony, Glyph Similar, or combination of them. Therefore Uncommon Chinese Characters and Other Characters would have a considerable amount. Some examples of attack methods are presented in Table 1. Figure 4(a) plots the CDF (Cumulative Distribution Function) of the sum of Uncommon Chinese Characters and Other Characters in the fraud SMS. We can observe that only 22% of them have less than 10 Uncommon Chinese Characters and Other Characters. It even exceeds 30 in some fraud SMS samples, which proves that Uncommon Chinese Characters and Other Characters are frequently adopted in fraud SMS.

Fig. 3.
figure 3

Fraud SMS and Gambling SMS sample.

Fig. 4.
figure 4

(a) The CDF of sum of Uncommon Chinese Characters and Other Characters. (b) The CDF of SMS length in RD1.

Another insight is that there are often contained some contact information in the fraud SMS, which increases the count of Numeric Characters and English Characters. Table 2 shows that 99.41% of the fraud SMS have Numeric Characters and 69.54% of them have English Characters, it's reasonable that the proportion with Numeric Characters is much higher than the English Characters, because the fraud SMS often contains QQ or Wechat which represented by 5–11 digit numbers directly. From Fig. 4(b), we can observe that the minimal length of the fraud SMS in RD1 is 17, and around 60% of the samples are longer than 50 characters. We argue that the length of the fraud SMS are commonly longer than normal one because of the expressions of contact information and the adversarial attacks (e.g., Word Split) in the fraud SMS.

Table 2. Statistics of Numeric Characters and English Characters in RD1.

Given the above measurement results, the most intuitive idea is that we can design features based on the appearance of Uncommon Chinese Characters or Other Characters, Numeric Characters or English Characters to classify the fraud SMS. However, the authors found out that some other abnormal SMS have the same characteristics. Taking the Gambling SMS sample in Fig. 3 for example, the # of Uncommon Chinese Characters is 12, and the # of Numeric Characters and English Characters are also greater than 0. If only these two types of features are considered, too much Non-fraud SMS will be recalled in practice. What's more, we plan to classify fraud SMS in a fine-grained manner, i.e., classifying the Part-time fraud, Loan fraud and Investment fraud SMS separately. For this purpose, a much more accurate and robust algorithm needs to be developed.

4 Algorithm Design

4.1 Model Overview

The goal of fraud SMS classification problem is to design a classification algorithm \(cls(\cdot )\) for any input text \(x \in X, x=\{{x}_{1},{x}_{2},\ldots ,{x}_{n}\}\) and output its fraud category. In this paper, we propose an algorithm composed by two stages, which can be defined as:

$$cls\left(x\right)={cls}_{ml}\left(V \left({cls}_{s1}\left(x\right)\right);\theta \right)=\widehat{y}$$
(1)

where Stage 1 \({cls}_{s1}(\cdot )\) is used to quickly filter out normal SMS, i.e. SMS not belongs to any fraud category. \({cls}_{ml}(V (\cdot );\theta )\) would be noted as stage 2, where \(V (\cdot )\) is a feature extraction function that convert SMS into continuous or discrete features, \({cls}_{ml}(\cdot ;\theta )\) is a machine learning classifier aims at obtaining the specific category of the SMS, and \(\theta \) are the parameters of the classifier.

The procedures of our method are shown in Fig. 5. Concretely, Stage 1 is introduced Subsect. 4.2, and we presents the b) of Fig. 5 in the Deep Model of Subsect. 4.3. Figure 5's part c) and d) are explained in the PINYIN and BISHUN and Hard Matching of Subsect. 4.3 separately.

Fig. 5.
figure 5

Algorithm procedures.

4.2 Stage 1 - Normal SMS Filter

In Stage 1, we plan to use simple method to filter out massive normal SMS quickly while keeping the algorithm's recall at a high level. Based on the measurement analysis in Sect. 3, two indicators are considered as follows:

$${I}_{nco}\left(x\right)= \mathbf{1}\left({c}_{uc}\left(x\right)+{c}_{other}\left(x\right)>{T}_{1}\right)$$
(2)
$${I}_{nen}\left(x\right)= \mathbf{1}\left({c}_{num}\left(x\right)+{c}_{en}\left(x\right)>{T}_{2}\right)$$
(3)

where \({c}_{uc}(x)\) is the count of Uncommon Chinese Characters in SMS \(x\), \({c}_{other}(x)\) denoted the count of Other Characters in \(x\),\({c}_{num}(x)\) and \({c}_{en}(x)\) stand for the count of Numeric Characters and English Characters in \(x\) respectively, T1 and T2 are the threshold parameters. As mentioned in Sect. 3, differing from the normal SMS, the fraud SMS often uses obfuscated characters to avoid detection. Many of these obfuscated characters are homophony or glyph similar compared with original characters, and majority of them are Uncommon Chinese Characters and Other Characters. Given this context, we argue that when the sum of Uncommon Chinese Characters and Other Characters is bigger than 0, i.e. T1 is 0, \(x\) could be a fraud SMS in a certain extent. On the other hand, criminals often need more contact channels to deliver information to victims other than SMS, and the information of network communicating channels is frequently appeared in the SMS. So \({I}_{nen}\left(x\right)\) is introduced to catch these information indirectly, which has already been explained in Sect. 3. The final output of Stage 1 is given by Eq. 4

$${cls}_{s1}\left(x\right)={\varvec{x}}\odot ({I}_{nco}\left(x\right) \wedge { I}_{nen}\left(x\right))$$
(4)

i.e. only the SMS with Uncommon Chinese Characters or Other Characters appearance, and Numeric Characters or English Characters appearance simultaneously will be sent to the stage 2. For simplicity, we set T1 and T2 equal to 0, and the detailed analysis of threshold setting will be discussed in the ablation study.

4.3 Stage 2 - Fraud SMS Classification

In stage 2, we aim to classify the SMS provided by previous stage in a find-grained manner. Due to the impressive advantage of a huge corpus in text classification, we exploited a deep neural model trained from Chinese corpus to extract semantic information. Besides, for the reason that there are lots of attacked texts in the fraud SMS, which are able to fool the deep model trained by the normal SMS, we design an artificial extractor, which can match specific keywords defined by anti-fraud professionals. Finally, features from deep model and artificial extractor were combined and fed into a machine learning model, e.g. SVM (support vector machine) to obtain the final classification results. The entire progress of stage 2 can be expressed as:

$$ V(x) = f_{{deep}} (x) \oplus f_{{pb}} (x,k_{1} ) \oplus f_{{hm}} (x,k_{2} ) $$
(5)
$${cls}_{s2}\left(x\right)={cls}_{ml}\left(V\left(x\right);\theta \right)$$
(6)

where \({cls}_{ml}\) is a machine learning classifier, \({f}_{deep}\) is deep model, \({f}_{pb}\) is PINYIN and BISHUN similarity extractor, and \({f}_{hm}\) is hard matching extractor. \(\oplus\) represents feature combination. We define \({f}_{pb}\left(x,{k}_{1}\right) \oplus{ f}_{hm}(x,{k}_{2})\) as artificial feature because it depends on pre-defined keywords group \({k}_{1}\) and \({k}_{2}\).

Deep Model.

In the field of natural language processing, pre-trained deep language models have became a very important basic technology. We use the BERT-wwm, which released by Harbin Institute of Technology and IFlytek, for deep feature extraction [5]. BERT-wwm is one of the best open source deep models trained on Chinese wikipedia (including simplified Chinese and traditional Chinese) data and have better performance in formal text modeling. Specifically, we directly use the pre-trained model for deep feature extraction, i.e. taking the SMS as input, and use the corresponding output vector of the first symbol (‘[CLS]’) as the semantic representation.

PINYIN and BISHUN.

Fraud SMS commonly use attacks against detection, and these transformed words could disturb the semantic meaning of sentences, affecting the accuracy of semantic features. Inspired by [23], we propose a strategy based on PINYIN [4] and BISHUN (stroke order) to recognize attack methods of some specific fraud keywords. Detailed procedures of our similarity judgement strategy are shown in Fig. 6. Firstly, the algorithm will search the Word Split attack in the SMS, and make corresponding corrections if such obfuscated characters exists. Then the keywords generated by anti-fraud specialists will be searched in the SMS, and output a boolean vectors whether each keyword exists. We argue that two words can be regarded as same only when they are same or highly similar in some aspects. Concretely, two perspectives for judgement are considered in our strategy:

  • SAME PINYIN. For homophony attack, one character would be transformed to others that has the same pronunciation. Since that Chinese characters can be pronounced base on their PINYIN, two words with the same PINYIN (without tones) could be regarded as the same word.

  • BISHUN similarity. A Chinese character has five basic writing strokes, i.e. 横 (horizontal), 竖 (vertical), 撇 (left-falling), 捺 (right-falling) and 折 (turning). By citing each basic stroke as number 1–5 and arranging them in writing order, BISHUN (stroke order) can uniquely identify a number sequence of a Chinese character. For glyph similar attack, transformed character usually shares some similar structure with the original one. For this reason, two words with high BISHUN similarity would be regarded as the same word in our method. The BISHUN similarity is calculated by adding the length of left common substring and the length of right common substring, then dividing by the length of keyword character composition string.

To explain our strategy clearly, we construct an example to further illustrate the process. Given an attacked SMS ‘徽亻言’ and a keyword ‘微信’ (wechat). Our method will try to combine each two adjacent characters first. In this case ‘亻言’ would be corrected to ‘信’. Though the PINYIN of ‘徽’ (hui) is different to ‘微’ (wei), their BISHUN ‘徽’ (33225215542343134) and ‘微’ (3322521353134) maintain long left common substring (3322521) and right common substring (3134). We calculate the BISHUN similarity as Eq. 7:

Fig. 6.
figure 6

PINYIN and BISHUN similarity judgement procedures.

$$\left(len\left(left\right)+len\left(right\right)\right)/len(keyword)\approx 0.846$$
(7)

Due to the reason of the high BISHUN similarity, the strategy would determine ‘徽亻言’ as an attack form of ‘微信’. Designed by anti-fraud specialists, we obtain 54 keywords in total from different fraud categories, therefore a 54-dim boolean features will be generated finally.

Hard Matching.

Although PINYIN and BISHUN are able to catch the majority attack methods of words, there are still some attack methods that could bypass the detection, such as non-Chinese special characters. Due to the limited non-Chinese special characters which could be used for attack, we adopt a hard matching method to determine whether these keywords exist in the SMS. As shown in Table 3, we design 6 hard matching expressions to search pre-defined keywords. We also use the SMS length as an additional feature since the number of special characters is also related to the SMS length (which has been partially investigated in Fig. 4(b)), and thus obtain a 7-dim hard matching feature. Finally, these hard matching features are combined with PINYIN and BISHUN features, resulting in a 61-dim artificial feature.

Table 3. Hard matching expressions.

ML Classifier.

After the deep features and artificial features extraction, each SMS is represented as continuous or discrete features. Any multi-class classifier can be deployed to learn feature distribution and give its prediction. In the experiments, we commonly use SVM in one-versus-all manner to train on SMS features with all hyper-parameter of the SVM set to default. The detailed analysis of ML classifier selection is also discussed in the ablation study.

5 Experiments

In this section, we conduct real-world data experiments to estimate the performance of our algorithm.

5.1 Dataset and Experiments Setting

Since we only have the fraud samples provided from citizens, the normal SMS and the gambling SMS samples need to be collected in another channel. We randomly sampled some normal SMS from [26] and [27] datasets and further gathered some gambling SMS from volunteers. Table 4 summarizes statistics of the two datasets explored in this paper, and we made DATASET2 available [6] for research usage.

Table 4. Sample statistics of two datasets.

As for the experiments with DATASET1, the fraud and gambling samples in training and testing set are separated based on the collected date, the normal samples are randomly distributed in the training and testing set. Due to the limited sample size of DATASET2, we choose the entire DATASET1 as training set, and the whole DATASET2 as testing set for DATASET2 related experiments.

5.2 Comparison of Different Algorithms

For comparison with our proposed algorithm TFC, five strategies are considered as follows:

  • JWE [24], in which exploiting stroke-level information for improving the learning of Chinese word embeddings. We get the sentence embedding vectors by averaging the word embedding vector obtained from the algorithm and Chinese word-split tools \cite{Jieba}.

  • cw2vec [25], in which jointly embed Chinese words as well as their characters and fine-grained subcharacter components. The sentence embedding vectors are obtained with the same procedures in JWE.

  • BERT-wwm, we feed the SMS into BERT-wwm model directly to get the deep features.

  • Stage 1 + BERT-wwm, we utilize Stage 1 to filter some normal SMS, and feed the rest SMS into BERT-wwm model to get the deep features.

  • PyCor + BERT-wwm, we use an open source Chinese text error correction tool Pycorrector [29] to correct the obfuscated characters in the SMS, and exploit BERT-wwm model to get deep features of the corrected SMS.

All the features calculated from TFC and the above five strategies will be sent to SVM with the same parameters to obtain the final classification result.

Table 5 presents the performance of different algorithms with DATASET1 and DATASET2, among them our proposed algorithms (TFC without Stage 1 and TFC) achieve best performance on both datasets. Particularly, we can observe that strategy with BERT-wwm has better performance compared with JWE and cw2vec. From the comparison of TFC without Stage 1 and PyCor + BERT-wwm, we conclude that our artificial features outperform than PyCor, which is able to deal with the obfuscated characters more effectively. From the comparison of TFC without Stage 1 and TFC, we can find that Stage 1 has negative effects on the overall performance. We argue that the result is reasonable, because the main purpose of Stage 1 is quickly filtering out the normal SMS, and it's inevitable that some fraud SMS samples may have 0 Uncommon Chinese Characters and Other Characters, so these samples will be misclassified as normal in Stage 1. Meanwhile, it's quite easy to identify the filtered normal SMS in stage 2, which further enhances the performance of TFC without Stage 1.

Table 5. Performance of different algorithm.
Table 6. Performance comparison of different domain-knowledge features.
Table 7. Classification examples under different algorithms.

Table 6 demonstrates the effectiveness of domain-knowledge related features. TFC uses three types of features, i.e. deep features from BERT-wwm, PB features from PINYIN and BISHUN extractor and HM features from Hard Matching principle. Deep features extracts semantic information of SMS and provides the basic classification performance. HM features and PB features further improve accuracy on DATASET1 and DATASET2, respectively. After combining all three types of features (i.e. TFC without Stage 1), our algorithm achieves better performance on both datasets.

Here we give a few examples to illustrate the effectiveness of TFC. As shown in Table 7, for example (a) and (b), only a few of characters in the sentence are changed, and the algorithm based on large corpus (like TFC and pycorrector + BERT) can effectively extract the semantic information of the sentence and obtain good performance. In example (c) and (d), almost each character in the sentence has been changed to its attacked form. It's challenging to correct them by the algorithm based on transition probability, which further effects the performance of deep features. The artificial features added by TFC can effectively identify the keywords in various attack methods, which could guide the classifier to make the correct choices.

5.3 Ablation Experiment

Recall that the main purpose of Stage 1 is filtering out normal SMS precisely and quickly, so we relabeled the Gambling and Fraud SMS as abnormal SMS. Therefore the output of Stage 1 can be viewed as a binary classification problem, i.e., classify normal and abnormal SMS. We studied the effects of T1 (the right hand side of Eq. 2) and T2 (the right hand side of Eq. 3) with DATASET1 train and test set in Fig. 7 and Fig. 8. From Fig. 7 we can find out when T2 is 0, changing T1 from 0 to 6 has trivial effects on the F1-score and Accuracy, but the results dropped significantly if T1 is greater than 7. Figure 8 shows the similar results when T1 is 0, and we can conclude that the reasonable value range for T2 is [0, 8]. Figure 9 plots the Stage 1's effect on total running time of TFC and # of samples. We can conclude that the total running time could be decreased 56.5% by simply setting T1 and T2 to be 0, which is quite important when the algorithm runs in the huge datasets. And the # of samples decreased 61.5% as well. The above results prove that the Stage 1 is able to effectively reduce the size of dataset while classifying the normal and abnormal SMS with satisfactory performance.

Given lots of domain-knowledge keywords are adopted in Stage 2, we also investigate their corresponding effects with DATASET1. We can change the problem into a binary classification problem for each fraud type, for example, Non Loan fraud SMS v.s Loan fraud SMS. Table 7 presents the results of binary classification with the impacts of the domain-knowledge keywords, we can observe that three types of fraud SMS have better performance when they exploit the related keywords. To explore the reason that both F1-score and accuracy are lower after utilizing the keywords for Part- time Fraud, the authors plot the percentage of samples that each fraud type's keywords appeared in the test set in Fig. 10. We can observe that the Part-time fraud keywords are appeared in 98.6% Part-time fraud SMS, but it also shown in 92.4% Investing fraud SMS and 75.5% Gambling SMS respectively, which means the uniqueness of the keywords is limited. As for the Investing fraud keywords, since the keywords appearance percentage in the Investing fraud samples is relatively higher, which helps the algorithm achieve bigger improving in both F1-score and accuracy (shown in Table 8).

Fig. 7.
figure 7

Accuracy and F1-score under different T1 (T2 = 0).

Fig. 8.
figure 8

Accuracy and F1-score under different T2 (T1 = 0).

Fig. 9.
figure 9

Accuracy and F1-score under different T2 (T1 = 0).

Table 8. Impacts of the domain-knowledge keywords.

We modify the final classification algorithm from SVM to RF (Random Forest), MLP (Multi-Layer Perceptron) and KNN (K-Nearest Neighbor) to study the different types algorithm's performance, parameters of all the algorithms are used in default setting with scikit-learn [30], and Table 9 demonstrates that the features generated by TFC are robust to many algorithms, and the performances are quite close compared with each other. Researchers can apply advanced or parallel algorithms to further improve the performance or speed up the running time.

Fig. 10.
figure 10

The percentage of samples for each fraud type keywords.

Table 9. Performance of TFC with different ML algorithm.

6 Conclusion

This paper mainly investigates the fraud SMS classification problem. We conduct a comprehensive measurement study of the fraud SMS, and observe some interesting phenomena. In addition, we propose a two-stage algorithm called TFC, which is able to quickly filter out normal SMS in the first stage, and classifying fraud SMS precisely in the second stage with the semantic deep features and the domain-knowledge based artificial features. By comparing with other algorithms using real-world dataset, we demonstrate that our algorithm TFC achieves better performance on both F1-score and accuracy. In the future, we plan to study how to extract the domain-knowledge keywords precisely and automatically.