1 Introduction

The number of complex documents and texts that need a thorough comprehension of data mining techniques to effectively categorize documents in numerous applications has significantly increased recently. Text categorization is a method that may effectively manage and organize texts while also making it easier for users to find essential information fast, making it a significant research topic in the field of information processing.

The process of feature selection is crucial to both text classification and data mining (Peng and Fan 2017; Shang et al. 2013). Text data are unstructured data made up of words. In order to enhance computer processing, textual data are frequently represented using the vector space model (VSM). It is used to transform unstructured data into structured data and treats the document as a collection of words (features) (Zhang and Duan 2019). Actually, not all feature is useful in creating the document classifier. These features could include those that are unnecessary or irrelevant. Additionally, it negatively impacts the classification outcomes when irrelevant features outnumber relevant ones. In this situation, choosing a subset of the original features frequently enhances classification performance (Sangodiah et al. 2014). As a result, feature selection in the context of text classification is a procedure that seeks to identify a minimal number of significant text features in order to minimize text classification error (Sebastiani 2002).

The current feature selection algorithms have significant difficulties; thus, we urgently require some feature selection techniques that can adapt to large amounts of data and have high accuracy and operational efficiency. Researchers in the field of machine learning are currently paying close attention to feature selection. The following two points summarize the primary reasons: (1) Irrelevant and duplicated features have an impact on the performance of some learning algorithms. According to certain studies, the amount of training data grows exponentially as the number of irrelevant characteristics rises (Langley 1994a; Jain and Zongker 1997). As a result, feature selection not only decreases computing complexity and increases accuracy of classification, but also aids in the discovery of simpler algorithmic models; (2) issues with high-dimension features in huge data processing are ongoing. The growth of data mining has created a pressing need for large-scale data processing, including gene analysis and information retrieval. High-dimensional feature spaces are inherently unsuitable for machine learning, according to empirical evidence. For some learning algorithms, "dimension disaster" or "combination explosion" is lethal. Therefore, in the case of massive data, feature selection is necessary for dimension reduction.

Feature selection is an important step in many pattern recognition and machine learning tasks, especially in the field of data analysis. The goal of feature selection is to identify the most relevant and informative features in a dataset that can be used to build a model.

One of the main challenges of feature selection is dealing with high-dimensional datasets. With the increasing amount of data being generated, it is becoming more important to find ways to efficiently select the most relevant features to improve the performance of models. In addition, many datasets contain redundant or irrelevant features that can negatively impact the performance of models and increase the risk of overfitting. There are several techniques for feature selection that can be summarized as follows:

  • Filter methods: Evaluate each feature individually based on a pre-defined criterion, such as information gain or correlation with the target variable, to select a subset of features.

  • Wrapper methods: Evaluate the performance of a machine learning model with a given set of features, and use this information to guide the selection of a new subset of features.

  • Embedded methods: Incorporate feature selection as part of the model training process, and use regularization or penalization to select a subset of features.

  • Deep learning-based feature selection: In recent years, deep learning techniques have been applied to feature selection to leverage the capacity of neural networks to learn complex representations of data.

  • Feature selection for interpretability and fairness: With the increasing importance of transparency and ethical considerations in machine learning, feature selection techniques that prioritize interpretability and fairness have become popular.

  • Ensemble feature selection: Ensemble methods have gained popularity in feature selection as they can combine the strengths of multiple feature selection algorithms to produce better results.

  • Feature selection in high-dimensional data: As the size of datasets continues to increase, feature selection techniques that can effectively handle high-dimensional data have become increasingly important.

  • Feature selection with dimensionality reduction: Combining feature selection with dimensionality reduction techniques such as PCA or t-SNE has become a popular approach to overcome the challenge of high-dimensional data.

The choice of feature selection method depends on the nature of the data, the task being performed, and the resources available. In many cases, a combination of methods may be used to obtain the best results.

In conclusion, feature selection is an essential step in many data analysis tasks and requires careful consideration to identify the most relevant and informative features in a dataset. The selection of the right feature selection technique is critical to building effective models and improving the performance of machine learning algorithms. As a result, we introduce a new feature selection method for text classification in this study in order to decrease the size of the subset of the selected features and increase the classifier's effectiveness without compromising its accuracy. The suggested approach relies on association analysis using the apriori algorithm (Agrawal and Srikant 1994) to reduce the high-dimensional feature space.

2 Research contributions and motivations

The following provides an overview of the study's main contributions:

Using frequent and Correlated items, the proposed feature selection approach extracts significant features from textual data that

  • Can identify the frequent and correlated features that are correlated with each another and highly related to the target variable.

  • Select features using an association analysis method.

  • Can significantly reduce the number of features that are selected.

  • Can eliminate features that are both unnecessary and redundant.

The main advantages of using frequent and correlated items to extracts significant features are as follows:

  • Improved model accuracy: By selecting only the most important and relevant features, frequent and correlated items can help improve the accuracy of machine learning models. This is because models trained with fewer features are less likely to overfit the data and can be more robust to noisy or irrelevant features.

  • Reduced computational cost: Feature selection can also reduce the computational cost of training and deploying machine learning models. This is because models trained with fewer features require less memory and processing power, making it possible to train and deploy them on resource-constrained devices.

  • Improved interpretability: By selecting only the most important features, frequent and correlated items can make machine learning models more interpretable and easier to understand. This can be especially important in applications where transparency and accountability are important, such as in healthcare or finance.

  • Improved fairness: By removing irrelevant or biased features, frequent and correlated items can help improve the fairness of machine learning models. This can be especially important in applications where fairness is a concern, such as in criminal justice or hiring.

3 Problem statement

The feature space with high dimensions is one of the most significant concerns in the problems of text categorization. As a result, choosing distinguishing features is crucial for text categorization. There are two main factors for choosing some features over others, according to Forman (2007). Due to scalability, using a smaller subset of features takes less time to compute since employing a large number of features consumes a lot of resources like memory, processing power, storage, network bandwidth, etc. The second factor has to do with how well the algorithm performs. For instance, algorithms work better when features that only add noise rather than additional information are ignored.

Problems associated with using large number of feature sets during the creation of the model can be summarized as follows:

  • Overfitting: A large number of features can lead to overfitting, where the model becomes too complex and is unable to generalize to new data. This can result in poor performance on unseen data.

  • Computational complexity: Training and deploying machine learning models with a large number of features can be computationally expensive and may require significant amounts of memory and processing power.

  • Increased risk of bias: Large feature sets may contain irrelevant or biased features, which can negatively impact the performance and fairness of machine learning models.

  • Decreased interpretability: Models with a large number of features can be difficult to understand and interpret, making it challenging to understand the factors that are driving predictions.

  • Feature redundancy: With a large number of features, it is possible that some features may be highly correlated with each other, leading to feature redundancy. This can increase the risk of overfitting and can make it challenging to interpret the results of the model.

In order to select significant features, a novel feature selection technique based on frequent and correlated items for text categorization is introduced. The proposed approach integrates association analysis theory with data mining. Interesting relationships between data items can be found via association analysis (Zhao and Liu 2009). By mining frequent and correlated items from the training dataset, the proposed technique aims to find features that are both correlated with each other and strongly related to the target attribute. Also, eliminate features that are unnecessary and redundant in an efficient manner.

4 Related work

Since the 1970s, when a significant amount of research was published, feature selection has been an active subject of research. This section reviews a number of recent studies that are related to the feature selection approaches for identifying significant and distinguishing features.

To identify heterogeneous features, Pawening et al. (2016) suggested a method of selecting key features based on mutual information (MI). The suggested method employed a joint mutual information maximization (JMIM) approach that considers the class label while selecting features. Further, to transform non-numerical attributes into numerical ones, it also employed the unsupervised feature transformation (UFT) technique.

An enhanced approach for the chi-square (Chi2) test that incorporates interclass concentration and frequency was presented by the authors (Sun et al. 2017). Three factors, in-class dispersion, frequency, and interclass concentration, were used to improve the Chi2 test.

The authors of Kaoungku et al. (2017) introduced an effective method for data classification combined with feature selection based on association rule mining in order to select features having a substantial impact on the target attribute.

In Qu et al. (2019), association rule-based feature selection (ARFS) was proposed. The frequent 2-item set of the target and feature variables was extracted from the dataset using association rules. The sequential forward selection strategy was coupled to search for feature subsets, and the performance of the decision trees algorithm was then used as the evaluation criterion for the selected feature subsets.

The authors of Larasati et al. (2019) merged the support vector machine (SVM) classifier with feature weighting and feature selection techniques. By taking a total of K = 500 of the top-ranked variables, the Chi2 was utilized to considerably reduce the number of variables. The weight of each variable that was chosen was determined using the feature weighting method.

In Zhou et al. (2020), a feature selection technique called WCFR (Weight Composition of Feature Relevancy) was proposed. The proposed algorithm utilizes standard deviation to assign weights to the relationships between features and feature sets.

The authors of Zhou et al. (2022) present a feature selection approach based on the combination of mutual information and correlation coefficient (CCMI) to evaluate the connections between various features.

The authors of Wang and Zhou (2021) present a combined feature selection method utilizing the chi-square test and the minimum redundancy approach. Through the chi-square test, the features most closely associated with the classes are selected, and then, a subset of these features with low redundancy is further chosen.

In Pathan et al. (2022), the authors utilize a filter-based feature selection method to identify the most pertinent medical features for predicting heart disease. Additionally, the correlation and interdependence among the various features were investigated.

The experiments mentioned above show that the majority of feature selection methods may effectively detect irrelevant features by utilizing various evaluation functions. However, it mostly focused on either removing the redundant features or taking the interaction of the features into account. In contrast, our approach uses association analysis as a technique for feature selection that seeks to remove the redundant and unnecessary features while also taking into account the interactions between the features. Association rule mining is a well-known technique in data mining. It is the process of deducing correlations between events or items, and these relationships can be expressed as rules for ease of comprehension and convenience when using the rules to forecast the presence of an event or item in the future.

5 Methodology

Since poor-quality data can decrease the effectiveness of model creation, data quality is crucial for categorization. Numerous irrelevant features are considered during the model-building process, which is the cause of the problem with the low performance. In this paper, we thus suggested a new feature subset selection method that looks for important features and also considers feature interaction. In addition, the proposed method use association as a metric rather than more conventional metrics such distance (Kononenko 1994; Liu and Zhang 2016; Anggraeny et al. 2018), dependency (Barraza et al. 2019; Sinayobye et al. 2019), and consistency (Zhao and Liu 2009; Dash and Liu 1997) to assess the relevance between the target attribute and feature(s). The proposed method for classifying texts goes through five stages: text preprocessing, feature extraction, feature selection using the proposed methodology, classification, and finally performance evaluation. Figures 1 and 2 depict the essential steps of the algorithm and the overall structure of the proposed system.

Fig. 1
figure 1

Main steps of the proposed text categorization system

Fig. 2
figure 2

The framework of the proposed text categorization system

5.1 Text preprocessing

In text mining, preprocessing is a vital step and crucial activity that is used to minimize the number of features in a dataset and enhance the performance of the classification strategy in terms of classification accuracy and resource needs. Therefore, most text collections and documents contain irrelevant terms, including stop words, misspellings, and Lang, which need to be removed. Unnecessary features and noise can negatively affect system performance in many algorithms, particularly probabilistic and statistical learning algorithms. For text categorization, we took into account three standard preprocessing techniques: tokenization (Verma et al. 2014), stop-word removal (Saif et al. 2014), and lemmatization (Samir and Lahbib 2018).

5.2 Feature extraction

In information retrieval and text mining, feature extraction is a crucial step. It transforms the text from an unstructured original into structured information that a computer can recognize and process by quantifying the distinctive words retrieved from the text that indicate the meaning of the text. Liu et al. (2018). The most used statistics for feature weighting, the TF-IDF approach, is used in this study to select and weight unique words features from the dataset (Soucy and Mineau 2005). The following equation may be used to determine the weight of any word in any document:

$$ W_{ji} = tf_{ji} {* }\log \frac{M}{{df_{j} }} $$
(1)

where \( {W}_{ji}\) is denotes the importance of the word j in the document i, M denotes the number of documents, \({tf}_{ji}\) is the frequency of the word j in document i, and \({df}_{j}\) denotes the number of documents includes the word j (Ahuja et al. 2019).

5.3 Proposed feature selection method

A new feature selection method is proposed in this section. The proposed method was developed especially for text classification problems. It looks for the important features while taking into account how the features interact with the class feature and with each another. It comprises three steps as shown in Figs. 3 and 4.

  1. 1.

    Identifying frequent items for each class label

  2. 2.

    Pruning frequent items

  3. 3.

    Identifying feature subset.

Fig. 3
figure 3

The proposed method for feature selection

Fig. 4
figure 4

The steps of the proposed method for feature selection

  1. 1.

    Identifying frequent items for each class label

In this stage, the frequently used items from the preprocessed training dataset are mined. The goal of mining frequently occurring items is to find the undiscovered, interesting relationships between patterns. The frequent set of items are those whose occurrence frequency exceeds a predetermined minimum support (Min_supp) criteria. The minimum support of the item sets is a parameter that is used to determine the minimum frequency or occurrence of an item set in a dataset. the minimum support of an item set can be calculated using the following equation:

$$ {\text{Min}}\_{\text{supp }}\left( X \right)\, = \,\left( {{\text{number }}\,{\text{of}}\,{\text{ transactions }}\,{\text{in}}\,{\text{ dataset }}\,{\text{containing }}\,X} \right)/\left( {{\text{total }}\,{\text{number}}\,{\text{ of}}\,{\text{ transactions}}\,{\text{ in}}\,{\text{ the }}\,{\text{dataset}}} \right) $$
(2)

where X is an item set, and Min_supp (X) is the minimum support of the item set. For example, if there are 100 transactions in a dataset and an item set {A, B} appears in 30 of them, then the support of {A, B} is 30/100 = 0.3 or 30%.

At this step, the apriori algorithm is being used on a text document. An apriori algorithm initially determines all frequent sets consists of single item (also known as 1-item sets) that meet a particular Min _supp criteria. You can continue performing this until you can no longer produce new frequent item sets (Agarwal and Srikant 1994).

In this study, we concentrate on frequently occurring features sets with two or more items. The next step is to prune these sets of items based on specific constraints after collecting the frequent items for each class label.

  1. 2.

    Pruning Frequent set of items

The main goals of the pruning procedure in machine learning are as follows:

  • Reducing overfitting: Pruning helps to remove irrelevant or redundant features that may be contributing to overfitting in the model.

  • Improving interpretability: Pruning can result in a simpler and more interpretable model, which is easier to understand and explain.

  • Reducing computation time: Pruning can reduce the size of the model, which can lead to faster training and inference times.

  • Improving generalization performance: By removing irrelevant or noisy features, pruning can improve the ability of the model to generalize to unseen data.

  • Reducing the risk of over-fitting: Pruning can help to reduce the risk of overfitting, especially when dealing with high-dimensional data.

  • Balancing between accuracy and complexity: Pruning helps to balance between accuracy and complexity by finding the optimal trade-off between model size and performance.

  • Reducing the number of features: Pruning helps to reduce the number of features, which can improve the performance of the model and reduce the risk of overfitting.

Reduce the number of sets of items that were created throughout the frequent item mining process with the pruning approach. This is because some sets of items might not be able to distinguish between classes, which could result in inaccurate classification. As a result, we must prune things to remove redundant and irrelevant data. Figures 5 and 6 show how our suggested algorithm prunes items by implementing the techniques described as follows:

  • First, find the associated items:

Fig. 5
figure 5

Pruning frequent items algorithm

Fig. 6
figure 6

Steps of computing all-confidence for each item set

To calculate the level of mutual association in the set of items, all-confidence is used. All-confidence (Klemettinen et al. 1994) of the set of items Y = (k1,…, ki), denoted as all_conf (k), is defined as follows:

$$ {\text{All - confidence }}\left( Y \right) \, = \, s\left( Y \right)/\max \, \left( {{\text{support}}\left( {k_{1} } \right),..,{\text{support}}\left( {k_{i} } \right)} \right) $$
(3)

According to Eq. 2, all-confidence for each frequent item set is calculated, and then, each item set having an all-confidence value less than or equal to the minimum all-confidence (Min_allconf) threshold is eliminated.

  • Secondly, pruning using the item set redundancy approach:

Using the item set redundancy approach, the set of items is also pruned if it contains another item set.

The correlated and frequent items, whose support and all-confidence are higher than the threshold, respectively, are mined at the end of the pruning process. The effectiveness of the item sets for categorization increases with closer associations between the items. Additionally, unnecessary item sets are removed in accordance with the idea that an item set should contain other item sets.

  1. 3.

    Identifying the Subset of Features

Features are chosen at this phase from the collection of frequent and correlated item sets. As a result, we ought to keep a set of items that are efficient in predicting class labels. For this reason, features that have a proportion of frequency occurrences in the collection of frequent and correlated item sets below the specified minimum frequency threshold are eliminated. The method then returns the features that are significant for predicting the class attribute after examining their distribution across the item sets taken from the training dataset. Our method for determining the frequent features subset is shown in Fig. 7.

Fig. 7
figure 7

Steps of selecting a frequent feature subset

The final feature subset, which retained frequent and significant features while removing redundant and unnecessary features as well as taking feature interaction into account, is determined at the end of the feature selection step.

5.4 Classification process

To properly classify unseen texts, classification algorithms are utilized. Several well-known classifiers are used in this study to evaluate how the selected features by the proposed approach affect classification accuracy, including Naive Bayes (NB), decision trees (DT), and logistic regression (LR).

  1. 4.

    Naïve Bayes (NB)

NB (Langley 1994b; Gopal 2019) is a popular classification method that is especially useful for high-dimensional datasets. Although it is a straightforward approach, NB can sometimes achieve better results than more sophisticated classification techniques. This is because NB estimates the likelihood of each input feature or attribute for a particular outcome, and then employs the Bayes rule to calculate the posterior probability for each class ci in the dataset.

$$P\left({C}_{i}|y\right)=P\left(y|{C}_{i}\right)P\left({C}_{i}\right)/P\left(y\right).$$
(4)

where

$$P\left(y\right)=\sum_{j}P\left(y|{C}_{j}\right)P\left({C}_{j}\right).$$
(5)

where P (Ci): the apriori probability of class Ci. P(y): the probability density function of feature y. P (y|Ci): the likelihood of feature y given that it belongs to class Ci. P (Ci|y): the posterior probability of class Ci given the observation of y.

5.5 Decision trees (DT)

DT are hierarchical structures commonly used for prediction and classification purposes, and they are considered effective and popular tools. At the top of the hierarchy is the root node, which represents the attribute that is chosen. From there, a branch is created for each available attribute value, and the structure is constructed through a series of if-else conditions. The leaves of the tree hold the result, which in the case of text classification, is a class label (Sohrabi and Karimi 2018).

To determine the homogeneity of a sample, the DT algorithm employs entropy (Zhang et al. 2019). The entropy is calculated using the following mathematical formula:

$$ {\text{Entropy}} = - \sum p\,\log p $$
(6)

where p is the frequency of the words on the left and right sides of the word w.

  1. 5.

    Logistic Regression (LR)

The LR (Cessie and Houwelingen 1992) is a linear model that is commonly utilized in classification problems. It assesses the connection between the response (dependent) variable and one or more explanatory (independent) variables for a given dataset to determine the importance and strength of the explanatory variables on the response variable. The LR model typically generates probabilities by employing the logistic function, which is also known as the sigmoid function which is given by Shu et al. (2018):

$$f\left(y\right)=\frac{L}{1+{e}^{-k\left(y-{y}_{0}\right)}}.$$
(7)

where e is the natural logarithm base, L is the curve's maximum value, y0 is the y-value of the sigmoid's midpoint, and k is the logistic growth rate or steepness of the curve.

5.6 Performance evaluation

The classification accuracy, precision, recall, F-measure and accuracy under ROC curve (AUC) are well-known assessment measures that may be used to assess the effectiveness of the proposed technique (Sokolova et al. 2006).

$$\mathrm{Accuracy}=\frac{\mathrm{TP }+\mathrm{ TN}}{\mathrm{TP }+\mathrm{ FP }+\mathrm{ TN }+\mathrm{ FN}}$$
(8)
$$\mathrm{Precision}=\frac{\mathrm{TP }}{\mathrm{TP }+\mathrm{ FP}}$$
(9)
$$\mathrm{Recall}=\frac{\mathrm{TP}}{\mathrm{TP }+\mathrm{ FN}}$$
(10)
$$\mathrm{F}1=2*\frac{(\mathrm{Precision }\times \mathrm{ Recall})}{(\mathrm{Precision}+\mathrm{ Recall})}$$
(11)

6 Experimental results and analysis

We have performed experiments to evaluate the effectiveness of the proposed feature selection approach. Three different classification techniques, DT, NB, and LR, as well as two different feature selection techniques, analysis of variance (ANOVA) and Chi2 are evaluated using the proposed algorithm. On a PC running 64-bit Windows 7 at a speed of 3 GHz with 4 GB of main RAM, we carried out our testing. Python programming is used to carry out the experiment.

ANOVA (Analysis of Variance) and chi-squared (χ2) test are both statistical methods used to compare groups or categories, but they are used in different situations and make different assumptions about the data.

ANOVA has the benefit of enabling for the comparison of means across many classes, which enables the analysis of more complicated datasets. This statistical technique is highly flexible and can handle both continuous and categorical data, whereas Chi2 is a straightforward and intuitive test that involves counting the number of observations in each category. It is a widely accepted and well-established method with a rich history of research and it is easy to interpret the results.

The main difference between ANOVA and Chi2test is that ANOVA can be used to test for interactions between variables and can be extended to include other factors such as repeated measures or random effects, while Chi2 test is mainly used to test for independence between two categorical variables.

In summary, ANOVA is used to test for differences in means of continuous data between multiple groups, while Chi2test is used to test for differences in frequencies or proportions of categorical data between multiple groups.

6.1 Preparing data for evaluation

In this study, experiments are conducted on the SMS spam collection dataset from the UCI machine learning repository (http://archive.ics.uci.edu/ml) to evaluate the classification system. SMS spam collection dataset is one of the most popular text classification datasets. That contains a set of 5196 unique words and 5572 SMS messages that belong to two different classes (ham, and spam).

Lemmatization was applied to convert words from the word vectors generated in the preprocessing stage, which involved removal of irrelevant characters, symbols, and words. 70% of the texts were utilized for training the classifier, by constructing a frequent subset of features. The performance of the classifier was evaluated using the remaining 30% of the texts. Table 1 displays the description of the SMS spam collection dataset.

Table 1 The description of the SMS spam collection dataset description

6.2 Results analysis and evaluation

The evaluation of the proposed methodology is presented in this section. The first part focuses on the generation of the frequent feature subset. The second part utilizes a series of tests to gauge the effectiveness of the selected features.

6.3 Frequent feature subset generation

The objective of the experiment is to apply the suggested feature selection technique to determine the most suitable features from a list of dataset features. The experiment involves testing various Min_supp values while keeping the Min_allconf threshold and the minimal frequency threshold constant, in an effort to attain the best results for the proposed method. The Min_supp threshold values were set to 0.002, 0.003, 0.004, and 0.0045 to extract frequent item sets from the preprocessed SMS spam collection dataset using the apriori algorithm. The minimal frequency threshold was established at 0.048 to identify frequent and correlated features from the pruned item sets, and the Min_allconf threshold was fixed at 0.13 to reduce the frequent item sets. The Min_allconf threshold and Minimum Frequency threshold values were determined as the optimal values through a trial-and-error process.

Table 2 shows the number and percentage of frequent features selected using our proposed feature selection method form the preprocessed SMS spam collection training dataset that contains 543 words/features.

Table 2 Percentage and number of frequent features using the proposed feature selection method for preprocessed SMS spam collection training dataset

The performance of the proposed feature selection method is compared with the two well-known feature selection methods that are ANOVA and Chi2 in the term of the number of selected features. The percentage of choosing the highest scoring features using ANOVA and Chi2 methods is set at 5%, 6%, 8%, 10%, 20%, 30%, and 40%. The number of features selected using ANOVA and Chi2 methods from the preprocessed SMS spam collection training dataset is listed in Table 3.

Table 3 Percentage and number of selected features using the ANOVA and Chi2 feature selection method for the preprocessed SMS spam collection training dataset

The number of features chosen by the three feature selection techniques is compared in Tables 2 and 3. It is clear that all feature selection techniques might considerably reduce the number of features chosen. However, our proposed feature selection approach does more than simply try to locate the frequent and correlated features that are related to each other and the target variable. It may also be used to eliminate redundant and unnecessary features. In contrast to the ANOVA and Chi2 procedures, the ranking criteria are not considered in the variable selection process. The Chi2 approach is used to determine the independence of two variables, whereas the ANOVA method is used to determine the variance between two groups (i.e. features and the target).

The following part will assess how well the features identified using the two methods performed.

6.4 Results evaluation

Several studies were carried out to determine the usefulness of our hypotheses and the quality of the selected features. The experiments were carried out using the DT, NB, and LR well-known classification techniques for prediction in three scenarios: (1) without sing feature selection method, (2) with using the well-known feature selection technique that are ANOVA and Chi2 and (3) with using the proposed feature selection method. The performance of different classifiers was then evaluated. For each classifier, the default parameters were applied.

According to Table 4, which compares the performance of several classification approaches using the first and second scenarios in terms of F-measure, the accuracy of classification, precision, recall and AUC, the best results are denoted in bold format.

Table 4 Comparison of different classifiers when using the first and second scenarios in terms of F-measure, accuracy, precision, recall and AUC

From Table 4, when applying the classification technique on the SMS spam collection dataset using the first scenario, we noticed that DT achieves higher performance 94.34% accuracy.

The purpose of this experiment is to compare the first scenario with the second scenario and assess the efficacy of applying classification algorithms using the ANOVA and Chi2 feature selection approach. The training dataset's features were chosen using various percentages of the highest-scoring features.

In Table 4, we can see that when trying to apply the classification using the second scenario, it is typically not always successful, especially when using low percentages for selecting features. This might be because when using low percentages, some essential features are filtered out, which has an impact on the effectiveness of the classifiers.

For the SMS spam collection dataset, in most cases when using the features selected by two feature selection methods improve the performance of classifiers. Whereas when using feature selection percentage 20% and 30%, the selected features contribute to improving the performance of all classifiers most often. In the case of LR, the performance is enhanced when Chi2 is used with a percentage of 20%, resulting in 94.61% accuracy, and 30%, resulting in 95.27% accuracy. On the other hand, using ANOVA with a percentage of 20% and 30% improves the accuracy to 94.52% and 95.31%, respectively.

These results also demonstrated that, even though the improvement percentage is modest, using the features chosen by two feature selection methods with a percentage of 20% and 30% can improve the performance of all the classification methods employed. However, the number of features used in this case is significantly lower than that used in the first scenario.

The findings from Table 4 are summarized in Table 5, which also includes the number of features used to train the classifier and the best classification accuracy and F-measure obtained for each classifier using the second scenario.

Table 5 The best results achieved for each classification technique using the second scenario

As shown in Table 5, the results show that the best performance for all classifiers is achieved when using the second scenario with Chi2 and ANOVA feature selection methods.

For the SMS spam collection dataset, the NB obtained the best accuracy of 95.49% using 45 features selected by Chi2 method. The DT obtained the best accuracy of 95.38% using 56 features selected by ANOVA method. The LR obtained the best accuracy of 94.497% using 221 features selected by ANOVA method.

Figure 8 shows comparative results of classifiers using first and second scenarios. It can be seen that when applying classifiers using the second scenario, the results are improved compared to the first scenario in terms of classification accuracy and F-measure. It is observed that, for the SMS spam collection dataset the NB with Chi2 gives the best performance with 95.49% accuracy using only 8% of features.

Fig. 8
figure 8

Comparative results of classifiers using first and second scenarios using SMS spam collection dataset

The classifiers performance using the third scenario with the proposed feature selection method for SMS spam collection datasets is shown in Table 6.

Table 6 Performance comparison of different classifiers using the proposed method with different Min_supp threshold values for SMS spam collection dataset

The purpose of this experiment is to compare the third situation to the first and second scenarios in order to evaluate how well classification techniques applied using the proposed feature selection approach perform. In the proposed feature selection method, features were selected from the SMS spam collection dataset under different values for the Min_supp that are 0.002, 0.003, 0.004, and 0.0045. For all experiments, the minimum frequency threshold was fixed at 0.048 and the Min_allconf threshold was fixed at 0.13.

From Table 6, we noticed that when applying classifiers with the proposed approach to select features, their performance is better.

When comparing the outcomes of the third scenario with those of the first scenario, we discovered that the classification approach works significantly better when using the third scenario than when using the algorithms with the first scenario, as shown in Fig. 9. In addition, for the SMS spam collection dataset, we can be observed that the three classifiers using the third scenario have the best performance using only 7% of features.

Fig. 9
figure 9

Comparison of the best results for classifiers using first and third scenarios in the term of accuracy for the SMS spam collection dataset

The second situation (as described in Table 5) and the third scenario given in Table 6 are used to compare the best results obtained when using the classifiers in Fig. 10. When using the third scenario to apply the classifiers, it performs significantly better than when using the second situation to apply the algorithms. But in the case of LR the SMS spam collection dataset, we observed that when applied with ANOVA with 30% of features, it performs better than when applied with the proposed method in terms of classification accuracy. We also noticed that the difference in accuracy between the two methods is not significant, but the percentage of features used in the proposed method is 7%, which is much lower than that used in the ANOVA method.

Fig.10
figure 10

Comparative results of classifiers using second and third scenarios for SMS spam collection dataset

It can be concluded that for the SMS spam collection dataset, the NB and DT classifiers using the third scenario have the best performance with accuracy (95.64%) and (95.38%), respectively, using only 7% of features.

7 Conclusion

This research introduces a new technique for text classification feature selection that focuses on identifying frequent and correlated items. The aim of the proposed method is to select features that are both frequent and highly correlated with each other and the target attribute. Additionally, it can eliminate unnecessary and redundant features from the feature space.

The results of the experiments conducted in this study are as follows:

  • The use of ANOVA and Chi2 procedures sometimes improved the effectiveness of classifiers in terms of accuracy, precision, recall, F-measure and AUC, especially when using fewer features in the second scenario compared to the first scenario.

  • When comparing the results of the third scenario with other scenarios using classification algorithms, it was observed that:

  • The third scenario achieved the best results in terms of accuracy, precision, recall, F-measure and AUC when using the features obtained from the proposed feature selection approach.

  • The proposed approach significantly reduced the number of selected features while still maintaining high quality and efficiency, resulting in lower processing costs and better classification performance.

  • For the SMS spam collection datasets, the NB and DT classifiers using the proposed feature selection method outperformed other approaches, achieving an accuracy of (95.64%) and (95.38%), respectively, using only 39 out of 543 features, which is only 7% of the total features.

Finally, this research successfully achieved its objectives and the results demonstrate the efficacy of the proposed feature selection technique for text classification.