Skip to main content
Log in

Formulation of rules for the scientific community using deep learning

  • Published:
Scientometrics Aims and scope Submit manuscript

Abstract

In a deluge of scientific literature, it is important to build scientific quantitative rules (SQR) that can be applied to researchers' quantitative data in order to produce a uniform format for making decisions regarding the nomination of outstanding researchers. Google Scholar and other search engines track scholars’ papers, citations, etc. However, the scientific community hasn't agreed on standards a researcher must meet to be regarded as important. In this paper, we suggest rules for the scientific community based on the top five quantitative scientific parameters. The significance of the parameters is measured based on two factors: (i) parameters’ impact on the model’s performance while classifying awardees and non-awardees, and (ii) the number of award-winning researchers elevated in the ranking of researchers through each respective parameter. The experimental dataset includes information from researchers in the civil engineering, mathematics, and neuroscience domains. There are 250 awardees and 250 non-awardees from each field. The SQR for each discipline has attained an accuracy of 70% or more for their respective award-winning researchers. In addition to this, the top ranked parameters from each discipline have elevated more than 50% of the award-winning researchers into their respective ranked lists of the top 100 researchers. These findings can guide individual researchers to be on the list of prestigious scientists, and scientific societies can use the SQR to filter the list of researchers for subjective evaluation in order to reward prolific researchers in the domain.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15

Similar content being viewed by others

Notes

  1. https://figshare.com/articles/dataset/ScientometricData/21206231.

References

  • Ain, Q. U., Riaz, H., & Afzal, M. T. (2019). Evaluation of h-index and its citation intensity-based variants in the field of mathematics. Scientometrics, 119(1), 187–211.

    Article  Google Scholar 

  • Alonso, S., Cabrerizo, F. J., Herrera-Viedma, E., & Herrera, F. (2009). h-Index: A review focused on its variants, computation, and standardization for different scientific fields. Journal of Informetrics, 3(4), 273–289.

    Article  Google Scholar 

  • Alonso, S., Cabrerizo, F., Herrera-Viedma, E., & Herrera, F. (2010). hg-index: A new index to characterize the scientific output of researchers based on the h-and g-indices. Scientometrics, 82(2), 391–400.

    Article  Google Scholar 

  • Ameer, M., & Afzal, M. T. (2019). Evaluation of h-index and its qualitative and quantitative variants in Neuroscience. Scientometrics, 121(2), 653–673.

    Article  Google Scholar 

  • Ayaz, S., & Afzal, M. T. (2016). Identification of conversion factor for completing-h index for the field of mathematics. Scientometrics, 109(3), 1511–1524.

    Article  Google Scholar 

  • Brigato, L., & Iocchi, L. (2021, January). A close look at deep learning with small data. In 2020 25th International Conference on Pattern Recognition (ICPR) (pp. 2490–2497). IEEE.

  • Burrell, Q. (2007). Hirsch index or Hirsch rate? Some thoughts arising from Liang’s data. Scientometrics, 73(1), 19–28.

    Article  Google Scholar 

  • Cabrerizo, F. J., Alonso, S., Herrera-Viedma, E., & Herrera, F. (2010). q2-Index: Quantitative and qualitative evaluation based on the number and impact of papers in the Hirsch core. Journal of Informetrics, 4(1), 23–28.

    Article  Google Scholar 

  • Crowder, R., Hughes, G., & Hall, W. (2002). December). An agent-based approach to finding expertise. In D. Karagiannis & U. Reimer (Eds.), International Conference on Practical Aspects of Knowledge Management (pp. 179–188). Springer.

    Chapter  Google Scholar 

  • Egghe, L. (2006). Theory and practice of the g-index. Scientometrics, 69(1), 131–152.

    Article  Google Scholar 

  • Falagas, M. E., Pitsouni, E. I., Malietzis, G. A., & Pappas, G. (2008). Comparison of PubMed, Scopus, web of science, and Google Scholar: Strengths and weaknesses. The FASEB Journal, 22(2), 338–342.

    Article  Google Scholar 

  • GarcÇa-Alvarez, D. (2009, January). Fault detection using principal component analysis (PCA) in a wastewater treatment plant (WWTP). In Proceedings of the International Student’s Scientific Conference (Vol. 2009).

  • Harzing, A. W., Alakangas, S., & Adams, D. (2014). hIa: An individual annual h-index to accommodate disciplinary and career length differences. Scientometrics, 99(3), 811–821.

    Article  Google Scholar 

  • Hirsch, J. E. (2005). An index to quantify an individual’s scientific research output. Proceedings of the National Academy of Sciences, 102(46), 16569–16572.

    Article  MATH  Google Scholar 

  • Jin, B. (2006). H-index: An evaluation indicator proposed by scientist. Science Focus, 1(1), 8–9.

    Google Scholar 

  • Jin, B., Liang, L., Rousseau, R., & Egghe, L. (2007). The R-and AR-indices: Complementing the h-index. Chinese Science Bulletin, 52(6), 855–863.

    Article  Google Scholar 

  • Kaushik, R. (2013). The “authorship index”—a simple way to measure an author’s contribution to literature. International Journal of Research in Medical Sciences, 1, 1.

    Article  Google Scholar 

  • Liang, R., & Jiang, X. (2016, February). Scientific ranking over heterogeneous academic hypernetwork. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 30, No. 1).

  • Moed, H. F., Bar-Ilan, J., & Halevi, G. (2016). A new methodology for comparing google scholar and scopus. Journal of Informatics, 10(2), 533–551.

    Google Scholar 

  • Prakash, J., & Kankar, P. K. (2020). Health prediction of hydraulic cooling circuit using deep neural network with ensemble feature ranking technique. Measurement, 151, 107225.

    Article  Google Scholar 

  • Prathap, G. (2010). The 100 most prolific economists using the p-index. Scientometrics, 84(1), 167–172.

    Article  Google Scholar 

  • Raheel, M., Ayaz, S., & Afzal, M. T. (2018). Evaluation of h-index, its variants, and extensions based on publication age & citation intensity in civil engineering. Scientometrics, 114(3), 1107–1127.

    Article  Google Scholar 

  • Sidiropoulos, A., Katsaros, D., & Manolopoulos, Y. (2007). Generalized Hirsch h-index for disclosing latent facts in citation networks. Scientometrics, 72(2), 253–280.

    Article  Google Scholar 

  • Sinatra, R., Wang, D., Deville, P., Song, C., & Barabási, A. L. (2016). Quantifying the evolution of individual scientific impact. Science, 354(6312), aaf5239.

    Article  Google Scholar 

  • Tol, R. (2009). The h-index and its alternatives: An application to the 100 most prolific economists. Scientometrics, 80(2), 317–324.

    Article  Google Scholar 

  • Urbanowicz, R. J., Meeker, M., La Cava, W., Olson, R. S., & Moore, J. H. (2018). Relief-based feature selection: Introduction and review. Journal of Biomedical Informatics, 85, 189–203.

    Article  Google Scholar 

  • Usman, M., Mustafa, G., & Afzal, M. T. (2021). Ranking of author assessment parameters using Logistic Regression. Scientometrics, 126(1), 335–353.

    Article  Google Scholar 

  • Velmurugan, C., & Radhakrishnan, N. (2016). Impact of research productivity on nanotechnology in India: A scientometric profile. International Journal of Multidisciplinary Papers, 49(3), 1–10.

    Google Scholar 

  • Wang, Z., Zhang, Y., Chen, Z., Yang, H., Sun, Y., Kang, J., ... & Liang, X. (2016, July). Application of ReliefF algorithm to selecting feature sets for classification of high-resolution remote sensing image. In 2016 IEEE international geoscience and remote sensing symposium (IGARSS) (pp. 755–758). IEEE.

  • West, R., & Stenius, K. (2004). Publishing addiction science: a guide for the perplexed. In UK7 International Society of Addiction Journal Editors, 387(2), 155–169.

    Google Scholar 

  • Wohlin, C. (2009). A new index for the citation curve of researchers. Scientometrics, 81(2), 521–533.

    Article  Google Scholar 

  • Ye, F., & Rousseau, R. (2010). Probing the h-core: An investigation of the tail–core ratio for rank distributions. Scientometrics, 84(2), 431–439.

    Article  Google Scholar 

  • Zhang, C. T. (2009). The e-index, complementing the h-index for excess citations. PLoS ONE, 4(5), e5429.

    Article  Google Scholar 

Download references

Acknowledgements

This research work was funded by the Institutional Fund Project under grant no. (IFPIP; 812-611-1442). Therefore, the authors gratefully acknowledge technical and financial support from the Ministry of Education and King Abdulaziz University, DSR, Jeddah, Saudi Arabia. Moreover, there is no conflict of interest between the authors and the funding department.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Muhammad Usman.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Alshdadi, A.A., Usman, M., Alassafi, M.O. et al. Formulation of rules for the scientific community using deep learning. Scientometrics 128, 1825–1852 (2023). https://doi.org/10.1007/s11192-023-04633-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11192-023-04633-5

Keywords

Navigation