Introduction

The evolution of research topics can be distinguished in three main stages: (i) embryonic stage, (ii) early stage, and (iii) recognized stage (Salatino, 2019). Prediction of future research trends has been a long-standing need for many stakeholders: including research evaluation, funding or awarding research policy making, and even commercial investment. Traditionally, expert interviews, workshops with experts and first-line stakeholders, and Delphi studies were used to identify the current research trends (Rotolo et al., 2015). However, a vast and rapidly growing literature makes it more and more difficult to identify the scientific research trends than ever before (Gibney, 2014). Manually identify emerging topics that will have long-term scientific impact is becoming time-consuming and resource-intensive, especially in their early stage of post-publication. Because the evaluation of whether an idea is novel or surprising depends crucially on already-existing knowledge, however, we are facing the literature explosion challenge (Bornmann & Mutz, 2015).

To overcome this problem, computer-aided methods and systems to detect emerging research trends are in increasing demand. Previous efforts have focused on bibliometric analysis. They devoted to the detection of what is emerging, citation counts, amounts of keywords or controlled terms, and the number of authors or publications are used as indicators of emerging research topics (Xu et al., 2021a). However, these citation indicators are not equal to research quality and the accumulation of publications will cause time delay of trends detection (Parraguez et al., 2020). Network analysis is also widely used in trend detection. Co-citation network and bibliograhic coupling are the main stream methods (Porter et al., 2019). Machine learning approaches are also becoming popular (Xu et al., 2019). Another branch of literature applies outliner detection and time series prediction, such as Non-negative Matrix Factorization (NMF) topic modeling method and time series analysis models (Klavans et al., 2020; Weismayer & Pezenka, 2017).

However, the citation and publication delay make the prediction is less timely. To solve the time delay problem, other scientific research-related data sources were introduced. Patent analysis and proposed technical keyword-based analysis of patents were used to monitor emerging technologies (Lee et al., 2018). Patent-paper citation network was used to represent importance indices (Wustmans et al., 2021). Another two biomedical informatics research applied the clinical guideline citation as a gold standard of recognition (Bian et al., 2017, 2019). Inspired by these novel data, we treated the cover papers of journals as important indices for research trend recognition.

Cover image and cover story of an academic journal, to some extent, represent the most important research of a journal. Although the researchers prefer to read online, one article chosen to be prominently placed on the journal's website drew more attention (Wang et al., 2015). The reason why research selected by editors as cover papers is complex: originality, importance, interdisciplinary interest, timeliness, accessibility, elegance, surprising conclusions, or any other factors (Costa & Salvidio, 2020). Some journals, such as Nature Chemistry, claimed that the choice to feature a particular article on the cover of the journal did not imply that the editors think it was better than the other papers in the issue (“Cover story”, 2010). However, recently published empirical study demonstrated that both citations and altmetric scores of cover papers were significantly higher than those of non-cover papers (Kong & Wang, 2020). Novel, important and high value research will more likely to lead the research trends (Xu et al., 2021a). For this reason, cover papers of top field journals can be treated as a relative gold standard of research with potential of becoming frontiers.

The widely accepted definition of emerging trends are as follows: (a) radical novelty, (b) relatively fast growth, (c) coherence, (d) prominent impact, and (e) uncertainty and ambiguity (Rotolo et al., 2015). Some researchers found that the greater the extent to which scientific knowledge (a paper) contains emerging ideas, the bigger its scientific impact is (Kwon et al., 2019). These factors are also considered in the decision process of editors and reviewers of the top-ranked journals to publish articles on a particular topic (Uzzi et al., 2013). It is suggested that there are many overlaps between the emerging researches and cover papers published on the top journals. For this reason, we translated the “emerging topics detection” mission to a “cover paper selection” problem. The top journal editors are thought as reliable “gatekeepers” with good judgement of scientific research (Primack et al., 2019). We applied the machine learning technologies to imitate human intelligence of the top journal (chief) editors, to select the novel, important, and potentially high impact research, and these research have more possibility to become the leader of the research trends. In other words, cover papers of the top journals is a reliable source for emerging topic detection.

Most current research trends detection methods are retrospective, such as citation-based analysis or publication trends analysis. However, even “bursty research topics” (Katsurai & Ono, 2019) detection has time delay, for the reason that emerging research topics formation requires a certain number of publication counts (Fang & Costas, 2020), not to mention citation counts. To overcome this problem, our applied the “real-time” prediction method. Newly-published papers can quickly evaluate whether they have the potential to be selected a cover paper, meaning that they will lead the research trends. Our work is prospective, and it will facilitate the process of emerging trends detection.

Our research focused on the prediction of the potential of being selected as cover paper by top journal (chief) editors. It will, to some extent, to overcome the time delay of the accumulation of publications and citations. And what’s more important, our model have the ability to identify the emerging topics potential by learning the research evaluation of top journal editors.

Related work

Text mining

Text mining technology is becoming popular in tackling the literature overwhelming problems (Antons et al., 2020). Compared with the traditional bibliometrics-based method, text mining digs into the paper content (Chen, 2005). Topic model algorithm was widely used, and has been applied to detect the emerging research trends (Kawamae & Higashinaka, 2010). Some advanced methods were used to improve the Latent Dirichlet Allocation (LDA) model’s performance, such as the Generalized Dirichlet multinomial regression (g-DMR) topic model and structural topic model (STM) (Bai et al., 2020).

By using combination of document influence model (DIM) and citation influence model (CIM), the growth, coherence, and influence indicators were extracted from papers and then were used to train the machine learning (ML) model to predict emerging topics (Xu et al., 2021b). In their following work, DIM was replaced by the topical n-grams (TNG) model to exploit the potential topic models (Xu et al., 2021a). Combined with the topic model, author and citation, author-link topic model, and dynamic author citation topic model were proposed to improve prediction accuracy (Anderson et al., 2012). An improved research grouped the documents by time in each topic to assess the topics’ popularity (Li et al., 2018).

Recent years have witnessed the exciting technology “word embedding” in NLP field (Katsurai, 2020). It is often used to identify emergent word meanings, usage, and semantical meaning (Di Carlo et al., 2019; Hamilton et al., 2016; Mihalcea & Nastase, 2012). Recently published Leap2trend (Dridi et al., 2019), applied temporal embeddings and tracked the dynamics of keywords over time, to detect trending scientific topics instantly. And it proved that semantic characterization of research topics yielded better results than keywords for tracing evolving research trends.

Machine learning

Using the keyword vectors, the k-nearest neighbors (KNN) algorithm was used to predict the forward citations of the patents (Woo et al., 2019). Traditional unsupervised ML technologies, such as PCA, also have good performance in clustering frequently-occurring-together patent records in emerging topics (Wang et al., 2019). Xu et al. (2019) applied the hybrid approaches that combine citation-link and lexical techniques to predict emerging research trends by using several ML models. They also used SVM and random forest to predict important citations (An et al., 2021). Feed-forward multilayer neural network was applied to identify the emerging technologies at early stages by using 18 real-time patent indicators (Lee et al., 2018).

Advanced neural deep learning models became popular for detecting emerging trends providing better performance than existing ML technologies. After building a semantic network on quantum physics, a convolutional neural network (CNN) using states of the semantic network of the past was used to predict future developments (Krenn & Zeilinger, 2020). Using a multimodal approach (including the patent abstract, claims, and indices), researchers built a framework combining of CNN and bidirectional long short-term memory (BiLSTM) in the early detection of valuable patents (Chung & Sohn, 2020). Considering the use of a full citation network, Graph Neural Network (GNN) was used to predict the citation counts, and then identify a set of top papers for predicting research trends (Cummings & Nassar, 2020). Using a manually curated corpus collected from arXiv.org, reinforcement learning and GAN were applied to detect short- and mid-term research trends (Eger et al., 2019). It should be pointed out that the prediction performance of machine learning model has not been well defined.

Link prediction is another widely-used approach in machine learning, which applies the network topology as feature. Network topological features, nodes similarity, and other measures can be utilized to predict the missing links of citation network (Foulds et al., 2015; Takeda & Kajikawa, 2010). Treating the snapshot of keywords network over time as dynamic network, the node centrality measures and node community information were induced to predict the future citation link (Behrouzi et al., 2020). The network topological indicators and dynamics information were used to construct a classification dataset. ML methods were then applied to classify the links and predict future associations among keywords. Autoregressive integrated moving average model (ARIMA), a time series analysis model, was used to predict the values of topological evolution (Choudhury & Uddin, 2016). Main path analysis (MPA) of citation network is another very promising method for understanding the evolution of a scientific domain (Jiang et al., 2020; Xu et al., 2020).

One of the key steps in the process of text classifier model construction is the gold-standard selection. In the process of semi-automated screening of biomedical citations for systematic reviews, research included in the systematic reviews were thought as gold standard of discriminate “relevant” from “irrelevant” citations (Wallace et al., 2010). In the task of relevant references identification, papers cited by the clinical guidelines were thought as the gold standard of research relevance evaluation (Bui et al., 2015). It means the citation of clinical guideline is the most relevant high-quality research in the specific research field (van Dinter et al., 2021).

Data and methods

Dataset

Essential Science Indicator (ESI), as an analytical tool in the Web of Science (WoS), is essential for researchers to identify the journals in a certain research area. To better predict the research trends and ignore the discipline factors that may cause confusion to the predicted result, we analyzed papers published in the top-tier journals of the material science discipline. The reason why we chose the material science as our data sample is that it has seen remarkable growth in material science recent years, which not only allows us to collect a larger dataset but also makes it possible for a more precise detection of the research trends. By using the ESI, 23 journals with high qualityFootnote 1 were selected, 13 journals with few papers and non-cover papers were excluded from the next analytical step.

The remaining 10 journals, including Nature Materials, Nature Nanotechnology, Advanced Materials, Advanced Energy Materials, ACS Energy Letters, Advanced Functional Materials, ACS Nano, Small, Journal of Materials Chemistry A, and Nano Letters were used for in-depth analysis. We retrieved bibliometric information of papers published in these 10 journals from the WoS database during the past ten years (2011–2020), manually encoding a dummy variable “paper type” as 1 if it is a cover paper. The cover papers were collected by browsing the journals’ website and identifying the research from every issue. In final, 84,447 papers were obtained, of which 2502 were cover papers and 81,945 were non-cover papers. The cover papers included in this study are collected manually by browsing the journals’ websites and reading every issue’s cover. All the included cover papers were cross checked by the two authors independently.

Feature for research trends prediction

To formally predict the research trends, a set of features that are related to the research trend of the paper are included. As is shown in Table 1, these features were divided into three categories, two of which were paper-related and author-related features. The paper-related features include: (1) novelty of research was calculated by using previous approach based on co-citation network of WoS data (Uzzi et al., 2013; Wang et al., 2017), that primarily grounded in exceptionally conventional combinations of prior work yet simultaneously features an intrusion of atypical combinations (Bornmann et al., 2019); (2) international collaborations, which are defined in this study as occurring when authors of paper are from two or more nations; (3) if the paper is highly cited (which is tagged in WoS as “highly cited”); (4) if the paper is a hot paper (which is tagged in WoS as “hot paper”); (5) number of references cited by the paper; (6) title length of the paper; (7) number of keywords of the paper; (8) number of words in the abstract of the paper, and (9) first year citations. The author-related features include: (1) number of authors of the paper; (2) number of addresses of the authors; (3) if the authors are funded; (4) if the fund is a national one, and (5) productivity. The data of these two types of feature categories can be obtained from the WoS database.

Table 1 Definition of features.

Besides the bibliometric information of papers that can be applied to trace the research trends, some online forms such as social media, research blogs, news articles, and policy documents are useful for research trends. These online discussions reflect the research frontier more timely than the citation count, thus, the research trend can be more accurately predicted by taking into account these features. Considering some recently published articles, especially those published in the year 2020, have a lower citation rate in spite of their significance, we calculated how many times these articles were cited by different online sources.

Altmetrics is a better way to understand all potential impacts of scientific research from the online forms (Costas et al., 2015; Priem et al., 2010). The count that papers mentioned by online forms, such as posts, Tweet, MSM, accounts, feeds, patent, and policy documents, can be gathered by altmetrics methods. Altmetrics indicators reduce the time window for predicting the number of citations that need to be accumulated, and the usage of the online altmetrics indicators is easy-going, which can be downloaded and obtained by a simple calculation. Restricted by the data access, we can only collect the summarized altmetrics data from altmetrics.org. To overcome the time accumulation effect, we normalized the altmetrics data by the age of the publication. Because the accumulate velocity of altmetrics indicators are heterogeneous, we normalized the slow source indicators (Feeds, patents and policy) using the mean value calculated by the publication years. And the fast source indicators (Posts, tweets, MSN and accounts) were included directly in the model (Fang & Costas, 2020). Specifically, the R package “rAltmetrics” was used to collect the data (Karthik, 2017). We categorized these 7 features as “altmetric feature” (Table 1).

The affiliations of author, to some extent, represent the team size, which has increasing almost in all areas. Works from larger teams build on popular developments and get the attention immediately (Wu et al., 2019). There is robust correlation that exists between citation impact and team size (Wuchty et al., 2007). The number of keywords, pages and references not only reflect the research productions but also reflect the authors’ energies for the publication. It is also found that more references and longer titles will receive more attention (Tan et al., 2016).

Method for prediction

ML technologies with access to a large dataset could discover and predict research trends using feasible features. We use the supervised ML methods to effectively detect the research trends. To be specific, we used seven ML technologies (Gaussian naive Bayes, AdaBoost, random forest, logistic regression, k-nearest neighbors, Calibrated ClassifierCV, and decision tree), which are more widely used method developed specifically for the prediction task. The prediction is built from the scientific papers we collected, using two-thirds of the sample as a training data set to find the best parameters for each of the ML algorithm. Then, the well-studied algorithm was applied to the 33% test data set to evaluate the rate of prediction.

To comprehensively evaluate the performance of predicting the research trend of these ML methods, miss errors in which the cover papers were not detected are adopted. Typically, we used recall score, the gold standard of approach for evaluating miss errors, to detect the proportion of cover papers.

The recall score is a criterion for evaluating true positive rate, a higher recall rate means more cover papers can be correctly predicted. As the error matrix of ML in Fig. 1, the recall score rate and precision rate are calculated by the formulas:

$$ \begin{aligned} & {\text{Recall}}\;{\text{rate}} = \frac{TP}{{TP + FN}}, \\ & {\text{Precision}}\;{\text{rate}} = \frac{TP}{{TP + FP}} \\ \end{aligned} $$
Fig. 1
figure 1

Error matrix for machine learning

The aim of this study is to evaluate the seven selected ML models, finding the best model with a higher recall rate which is beneficial to provide a user-friendly approach for the detection of research trends.

To compare the accuracy performance, the area under the receiver-operating-characteristic (ROC) curve (AUC) was used (Ling et al., 2003). The ROC curve was also adopted to compare the ML models’ performance across the entire range of error rates in each of the models (Bradley, 1997).

To ensure the best baseline performance of the ML algorithm, the grid search method was used to find the optimal parameters. Tenfold cross-validation was conducted to training the data, which provide the optimal parameter combination. By implementing the “GridSearchCV” in the sklearn package (Pedregosa et al., 2011), we systematically searched the optimal parameter of each ML technology. To be consistent with the purpose of finding the best ML model with a higher recall score, the optimal parameter was determined by the best recall score on the test dataset. The optimal parameters of the seven ML models are presented in Table 2.

Table 2 Optimal parameters for the seven machine learning models

To evaluate the generalization ability of our well-trained models, we further collected data from three most high-impact journals published in 2019, including Cell, Nature, and Science. All the bibliometrics data were also collected and cover papers were manually identified. The CNS sample contained 1622 papers, 118 of which was cover paper. We tested the recall score of the sample to check the externality of the trained ML models, checking if the models in terms of detecting research trends performed well in these journals as well.

Results

Descriptive analysis

Table 3 presents the descriptive statistical analysis of 17 variables, and five dummy variables were excluded in this analysis.

Table 3 Description of features and outcome variable.

Topic distribution result

We formed the categories of topics of the papers we collected by applying computer-assist techniques–topic model to check the topic distribution of the cover papers. Specifically, a probabilistic topic model called LDA was employed to automatically identify main topics (the optimal number of topics is 19) from each abstract of the papers (Blei et al., 2003). The underlying assumption of LDA is that the abstract of a paper consists of multiple topics, each topic is a discrete probability distribution over words (Steyvers & Griffiths, 2007). Thus, the topic name (which are manually labeled) could be inferred based on algorithmically found top words.

Before the text mining analysis, preliminary data processing process (including words stemmization, words lowercase, punctuation, numbers and stop words removing) were conducted to ensure the clusters of topics will have significant meanings. To determine the number of topics in our analysis, four metrics in R package “ldatuning” were also used to examine the model performance and to estimate a reasonable number of topics (Arun et al., 2010; Cao et al., 2009; Deveaud et al., 2014; Griffiths & Steyvers, 2004). We found that 19 topics best average performance across the four metrics. Thus, we clustered the 84,447 abstracts of papers into the optimal 19 topics.

We then classified each paper into the topic with the highest probability distribution. The descriptive statistics of the cover paper were presented in Table 4. It can be seen that the proportion of cover paper in each topic is around 3%, which is consistent with the proportion of those in the sample (2.96%). The result suggests that the topic distribution of the cover papers is equal, cover papers normally distribute in the research topic in the material science discipline.

Table 4 Summary statistics of 19 topics

We also conducted a sub-group prediction analysis by using the 19 topic models to produce clusters separately. We adapted the model including overall the features, and found that the performance was consistent in all the sub-clusters. The average recall rate was 0.68, precision rate was 0.56, and the AUC is 0.71, some prediction results (4/19) even exceeded the best performance of the prediction model. This could be explained partly by the topic specific characters of the prediction model.

Evaluation result of machine learning models

Considering the data is unbalanced, the proportion of cover paper is small, it will introduce bias during the training process. As a result, the training rest will bring about the true positive issue, most of the cover paper will not be predicted correctly. To solve this issue and minimize the impact of the sample bias when we randomly split the data into training and test data sets, an approach for constructing the classifiers from unbalanced data-synthetic Minority Oversampling Technique (SMOTE) is used (Chawla et al., 2002). SMOTE creates a “synthetic” minority class by using the over-sampling approach, it generates synthetic cover paper samples so that a balanced training data set is created. We used the papers published in 2016 as training sets, in which the number of cover papers is only 248, while the number of non-cover papers is 7,074. When using SMOTE, the number of cover papers in the training data set is equal to that of non-cover paper, with the number of 7074. The other publications published in 2011–2015 and 2017–2020 were used to evaluate the prediction performance.

Figure 2 displays the ROC curve of the seven ML models. We first excluded the “novelty”, “Altmetrics” and “citation in the first year” from all the features, and then added them one by one (Fig. 2a–c). It could be found that the Altmetric features indeed slightly improved the performance of prediction models, while the citation feature lessened the performance. It is hinted that citation based indicators could not make sense. Additionally, it is worth noting that the Logistic regression and Linear SVM model with the best performance in the most tests. The AUC of the prediction model including overall features is 0.7, as is shown in Table 5 with the recall rate and the precision rate. We also tested the prediction performance by using datasets 2011–2015 and 2017–2020, separately. We found that the AUCs of 2017–2020 are higher than datasets of 2011–2015 (Appendix Fig. 4).

Fig. 2
figure 2

ROC of the seven machine learning models

Table 5 Prediction performance of 7 machine learning models with overall features

In this research, we preferred the high performance of the recall rate. The false negative rate (non-cover paper predicted as cover paper) was considered with lower priority. It means that we would include some “false positive” papers, which is the trade-off between the high recall rate and precision rate.

When excluding international collaborations and productive features, we found that the performance decreased (Recall = 0.63, Precision = 0.37, AUC = 0.61), which means the two author-related features are important. Although a research investigated Scopus data of 2005 found that international collaboration tends to produce conventional research (Wagner et al., 2019), a recently study focused on COVID-19 field reached the opposite conclusion (Liu et al., 2021). Our study of materials science supported, to some extent, the later opinion, for the reason that this filed is labor intensive and device induced.

Externality of the machine learning models

The ROC curve of the CNS papers is present in Fig. 3. It can be seen that the performance of the logistic regression (recall = 0.71, precision = 0.51) and the linearSVC (recall = 0.73, precision = 0.49) are better. The generalization of the prediction model was consistent with that of the samples in the material science discipline. It indicates that our ML method in terms of identification of cover paper could be applied in a wide range of samples. Moreover, the prediction performance of external data was better than that in material science, which could be explained by the three journals are all high-quality.

Fig. 3
figure 3

The ROC curve of the CNS papers

Validation of field expert review

The prediction of results was validated by 10 experts from materials science. We mapped the prediction results to previous topic model clustered groups. Experts reviewed the results and concluded that the papers we presented included almost all the emerging topics in recent years. It is also highlight that our prediction results in 2019 covered all the emerging topics, which means the framework we proposed overcome the time delay problem of traditional emerging topics detection.

Ten filed experts were invited to evaluate the model performance by choosing the cover papers in two pre-designed tests. The experts are currently working in the material science department and physical science department of two universities. Their research interests are mental materials, material chemistry, electronic materials, optical material, computational materials, nanomaterial and nanotechnology, and theoretical physics. All the experts (3 professors, 3 associated professors and 4 Ph.D. Candidates) published more than at least ten papers in the materials field. Before the formal test, every expert selected 3 topics, and was assigned 20 cover papers and 80 non-cover papers to train their ability of cover paper identification. It was ensured that papers belong to every topic has two “reviewers” to evaluate whether it could be selected as cover paper. After that, we conducted a test accuracy experiment to compare the performance of the prediction model and the experts. Cohen's к statistic was used to evaluate the experts’ inter-rater reliability.Footnote 2

Ten experts were assigned to identify the 10 cover papers from 100 papers published in 10 materials journals in 2021. Innovation and importance were considered firstly in the process of evaluation. The cover papers and non-cover papers are all from the topics they selected. The tests were conducted 4 times and every expert evaluated more than 400 papers, which ensured that every paper was reviewed by two experts independently.

The average recall rate of the prediction model was 0.734 and the precision rate was 0.690, which was higher than the average recall rate of experts at 0.595 and 0.673, respectively. The agreement of the two experts was slight consistency (к = 0.58). Considering the experts have priori knowledge, such as could have read some papers before, prediction model would still perform on par with the experts. In addition to a little higher performance, the prediction model has also been consistency of producing results.

Discussion and conclusions

This study provides a compressive ML framework to predict the emerging research trends in an innovative way. It is useful to overcome the time delay problem of traditional publication counts based methods. We have developed a principled and extensible approach for identifying leader papers of emerging topics by combining the topic model and prediction models. By consulting with 10 field experts, the cover papers we selected are proved that they included all the emerging topics. This result ensured the effectiveness of our approach to detect emerging papers instantly. What’s more, our model applied some easily collected bibliometrics data and this will helpful to simplify a great deal of the pre-processing works. The comparison test has shown that prediction model has comparable performance with materials science experts. Additionally, the non-fatigue characteristic of prediction model enables constant training and learning until achieving satisfactory accuracy. It could reduce the cognitive burden on researchers, and will helpful to increase the efficiency of understanding the research domains.

The strategy presented here is based on the hypothesis that paper featured on the cover of a top journal has more possibility to become a leader paper of an emerging topic. Cover papers are closely related to leader papers of emerging topics in terms of relatively fast growth of publications and citations, radical novelty and prominent impact. The process of cover paper selection is, to some extent, a kind of “knowledge distillation” in essence (Gou et al., 2020). The ML models we proposed can be thought as a top editor, whose has a good scientific taste and judgement of identifying important and novel papers. It is also critical to understand that the cover paper of top journal is not quite equal to the leader paper of emerging topics. However, when facing with the challenge of literature flood, our approach can reduce the time of citation accumulation. And this might seem to be an acceptable solution of information consolidation.

Although the idea of the prediction model we constructed stemmed from the cover paper selection, the readers of a newly-published paper always passively get some related information, such as citation counts or discussions on the twitter in the real world. The model, to some extent, is a post-publication peer review, rather than editors’ review. Additionally, it is suggested that over half of altmetric events happened within 1 month after the research published (Fang & Costas, 2020).This makes the post-publication “review” become more efficient. It is noted that, in this research, we preferred the high performance of the recall rate. The false negative rate (non-cover paper predicted as cover paper) was considered with lower priority. It means that we would include some “false positive” papers, which is the trade-off between the high recall rate and precision rate.

Our analysis also showed some important results. First, Altmetrics has effectively replace the citation counts in the prediction models. The precision performance of Altmetrics based model was significantly higher than citation counts based model. Considering one aim of this study is focused on reducing the time delay of data collection, Altmetrics is a good indicator, especially in those research trends prediction mission. Thus, our results suggest that feature engineering of bibliometrics data would probably be an impractical endeavor. It is suggested that over half of altmetric events happened within 1 month after the research published (Fang & Costas, 2020). This makes the post-publication “review” become more efficient. More deep learning model should be introduced to avoid feature selection.

This study inevitably has some limitations. As pointed out previously, our approach is incomplete despite the promise inherent in the circumstantial prediction. The expert reviewed our results found that our model predicts many unrelated papers. This is owing that we pursue the highest precision rate of our model. Secondly, although the approach has been applied in biological journals and top interdisciplinary journals, more work should be conducted to ensure the fitness of our prediction models. Some cases, such as sleeping beauty literature should be used to validate our model to evaluate the ability of weak signal detection (Ke et al., 2015).

Future research will focus on the identifying the causal relationship between cover letter and emerging research trends. Cover papers have some advantages of high visibility, and this will make readers have more interests to conduct follow-up research. Is there a bidirectional casualty between them? And how the evaluation dynamics of the relationship between them? Indeed, our model should be extended to consider the dynamics of topics, and this will useful in sub-topic level emerging trends detection. Another interesting challenge for a long-term research topic would be the comparison of our methods with traditional prediction models in handling weak signals of emerging topics. To ensure the fairness of assessment, the performance evaluation framework is ugly needed.