Skip to main content
Log in

Quantifying the classification of exoplanets: in search for the right habitability metric

  • Regular Article
  • Published:
The European Physical Journal Special Topics Aims and scope Submit manuscript

Abstract

What is habitability? Can we quantify it? What do we mean under the term habitable or potentially habitable planet? With estimates of the number of planets in our Galaxy alone running into billions, possibly a number greater than the number of stars, it is high time to start characterizing them, sorting them into classes/types just like stars, to better understand their formation paths, their properties and, ultimately, their ability to beget or sustain life. After all, we do have life thriving on one of these billions of planets, why not on others? Which planets are better suited for life and which ones are definitely not worth spending expensive telescope time on? We need to find sort of quick assessment score, a metric, using which we can make a list of promising planets and dedicate our efforts to them. Exoplanetary habitability is a transdisciplinary subject integrating astrophysics, astrobiology, planetary science, and even terrestrial environmental sciences. It became a challenging problem in astroinformatics, an emerging area in computational astronomy. Here, we review the existing metrics of habitability and the new classification schemes (machine learning (ML), neural networks, activation functions) of extrasolar planets, and provide an exposition of the use of computational intelligence techniques to evaluate habitability scores and to automate the process of classification of exoplanets. We examine how solving convex optimization techniques, as in computing new metrics such as Cobb–Douglas habitability score (CDHS) and constant elasticity earth similarity approach (CEESA), cross-validates ML-based classification of exoplanets. Despite the recent criticism of exoplanetary habitability ranking, we are sure that this field has to continue and evolve to use all available machinery of astroinformatics, artificial intelligence (AI) and machine learning. It might actually develop into a sort of same scale as stellar types in astronomy, to be used as a quick tool of screening exoplanets in important characteristics in search for potentially habitable planets (PHPs), or Earth-like planets, for detailed follow-up targets.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Data Availability Statement

Data and codes associated with this manuscript are available on several websites: 1. ESI/MSI datasets: https://data.mendeley.com/datasets/c37bvvxp3z/8 2. CDHS/CEESSA Datasets and Catalogs: AstrIRG (Astroinformatics Research Group) website http://astrirg.org/projects.html 3. Codes to analyze exoplanetary data from the PHL HEC catalog: https://github.com/SuryodayBasak/ExoplanetsAnalysis/tree/master/MLAnalysis

Notes

  1. An inscription on the iron pillar at Delhi’s Qutub Minar, probably left in the 4th century BC. Such mentions can also be found in the Mahapuranas—ancient texts of Hinduism (e.g. Rocher [6]).

  2. Life can still exist on Jupiter, in the clouds—the buoyant organisms could exist at depths of several hundred kilometers in the atmosphere; idea originally proposed in 1974 [15]. Though there is no oxygen, even ordinary Earth bacteria was recently shown to live and thrive in a 100% hydrogen atmosphere [16].

  3. Universidad Central “Marta Abreu” de Las Villas, Santa Clara, Villa Clara, Cuba.

  4. http://phl.upr.edu/library/notes/athermalplanetaryhabitabilityclassificationforexoplanets.

  5. Saha Bora Activation Function: https://github.com/sahamath/sym-netv1.

References

  1. J. Pathak, A. Wikner, R. Fussell et al., Chaos Interdiscip. J. Nonlinear Sci. 28(4), 041101 (2018)

    Google Scholar 

  2. Y.G. Zhang, V. Gajjar, G. Foster et al., Astrophys. J. 866, 149 (2018)

    ADS  Google Scholar 

  3. N.M. Batalha, J.F. Rowe, S.T. Bryson et al., Astrophys. J. Suppl. 204, 24 (2013)

    ADS  Google Scholar 

  4. E.A. Petigura, A.W. Howard, G.W. Marcy, Proc. Natl. Acad. Sci. 110(48), 19273–19278 (2013)

    ADS  Google Scholar 

  5. M.C. Turnbull, T. Glassman, A. Roberge et al., Publ. Astron. Soc. Pac. PASP 124, 418 (2012)

    ADS  Google Scholar 

  6. L. Rocher, Puranas (Otto Harrassowitz, Wiesbaden, 1986). ISBN 978-3447025225

    Google Scholar 

  7. G. McColley, Ann. Sci. 1, 385–430 (1936)

    Google Scholar 

  8. A.J. Sternfeld, La Nature, Masson et Cie (eds.), Paris, No. 2956, 1–12 (1935) (in French)

  9. G.A. Tikhov, Priroda (Leningrad) 46(2), 3–6 (1947) (in Russian)

  10. T. Sumi, K. Kamiya, D.P. Bennett et al., Nature 473, 349 (2011)

    ADS  Google Scholar 

  11. M. Safonova, J. Murthy, Y.A. Shchekinov, Int. J. Astrobiol. 15, 93 (2016)

    ADS  Google Scholar 

  12. M. Matsuura, A.A. Zijlstra, F.J. Molster et al., Astrophys. J. 604(7), 91–799 (2004)

    Google Scholar 

  13. E. Schrödinger, What is life? The physical aspect of the living cell (Cambridge University Press, Cambridge, 1944)

    MATH  Google Scholar 

  14. G. Witzany, Front. Astron. Space Sci. 7, 7 (2020). https://doi.org/10.3389/fspas.2020.00007

    Article  ADS  Google Scholar 

  15. W.F. Libby, Life on Jupiter? Orig. Life Evol. Biosph. 5, 483–486 (1974)

    Google Scholar 

  16. S. Seager, J. Huang, J.J. Petkowski et al., Nat. Astron. 4, 802–806 (2020)

    ADS  Google Scholar 

  17. M. Murakami, K. Hirose, H. Yurimoto, S. Nakashima, N. Takafuji, Science 295(5561), 1885 (2002)

    ADS  Google Scholar 

  18. D. Atri, J.R. Soc, Interface 13, 20160459 (2016)

    Google Scholar 

  19. J.R. Spear, J.J. Walker, T.M. McCollom, N.R. Pace, Proc. Natl. Acad. Sci. 102, 2555–2560 (2005)

    ADS  Google Scholar 

  20. Y. Morono, M. Ito, T. Hoshino et al., Nat. Commun. 11(1), 3626 (2020)

    ADS  Google Scholar 

  21. BBC News, Tardigrades: ‘Water Bears’ Stuck on the Moon after Crash (2019), https://www.bbc.com/news/newsbeat-49265125. Accessed 17 Sept 2019

  22. D.S. Spiegel, E.L. Turner, Proc. Natl. Acad. Sci. 109, 395–400 (2012)

    ADS  Google Scholar 

  23. J.J. Swift, J.A. Johnson, T.D. Morton et al., Astrophys. J. 764(1), 105 (2013)

    ADS  Google Scholar 

  24. M. Kunimoto, J.M. Matthews, Astron. J. 159(6), 248 (2020)

    ADS  Google Scholar 

  25. S. Bryson, M. Kunimoto, R.K. Kopparapu, J.L. Coughlin, W.J. Borucki et al., Astron. J. 161, 36 (2021)

    ADS  Google Scholar 

  26. R.H. Dicke, Nature 192, 440–441 (1961)

    ADS  Google Scholar 

  27. P. Dayal, M. Ward, C. Cockell, Preprint arXiv:1606.09224 (2016)

  28. A. Loeb, Int. J. Astrobiol. 13, 337–339 (2014)

    ADS  Google Scholar 

  29. J. Haqq-Misra, R. Kopparapu, E. Wolf, Int. J. Astrobiol. 17(1), 77–86 (2018)

    ADS  Google Scholar 

  30. T.M. McCollom, Proc. Natl. Acad. Sci. 113(49), 13965–13970 (2011)

    ADS  Google Scholar 

  31. F. Klein, N.G. Grozeva, J.S. Seewald, Proc. Natl. Acad. Sci. 116(36), 17666–17672 (2019)

    ADS  Google Scholar 

  32. C. Oze, L.C. Jones, J.I. Goldsmith, R.J. Rosenbauer, Proc. Natl. Acad. Sci. 109(25), 9750 (2012)

    ADS  Google Scholar 

  33. L.O. Stephanie, W.S. Edward, T.R. Christopher et al., Astrophys. J. 858, L14 (2012)

    Google Scholar 

  34. J. Krissansen-Totton, S. Olson, D.C. Catling, Sci. Adv. 4, eaao5747 (2018)

    ADS  Google Scholar 

  35. D. Schulze-Makuch, A. Méndez, A.G. Fairén et al., Astrobiology 11, 1041 (2011)

    ADS  Google Scholar 

  36. R. Barnes, V.S. Meadows, N. Evans, Astrophys. J. 814, 91 (2015)

    ADS  Google Scholar 

  37. J.M. Kashyap, S.B. Gudennavar, U. Doshi, M. Safonova, Astrophys. Space Sci. 362(8), 146 (2017)

    ADS  Google Scholar 

  38. A. Méndez, in Proceedings of sixth astrobiology science conference, Houston, TX, USA, 26–29 April 2010

  39. R. Cardenas, N. Perez, J. Martinez-Frias, O. Martin, Challenges 5, 284 (2014)

    ADS  Google Scholar 

  40. K. Bora, S. Saha, S. Agrawal, M. Safonova, S. Routh, A. Narasimhamurthy, Astron. Comput. 17, 129–143 (2016)

    ADS  Google Scholar 

  41. S. Saha, S. Basak, M. Safonova, K. Bora, S. Agrawal, P. Sarkar, J. Murthy, Astron. Comput. 23, 141 (2018)

    ADS  Google Scholar 

  42. E. Tasker, J. Tan, K. Heng, S. Kane, D.L. Spiege, the ELSI Origins Network Planetary Diversity Workshop, Nat. Astron. 1, 0042 (2017)

    Google Scholar 

  43. J.F. Kasting, D.P. Whitmire, R.T. Reynolds, Icarus 101, 108 (1993)

    ADS  Google Scholar 

  44. D.J. Stevenson, Nature 400, 32 (1999)

    ADS  Google Scholar 

  45. M.A. Limbach, E.L. Turner, Proc. Natl. Acad. Sci. 112, 20 (2015)

    ADS  Google Scholar 

  46. Y. Wang, Y. Liu, F. Tian, Y. Hu, Y. Huang, Preprint arXiv:1710.01405 (2017)

  47. R. Heller, J. Armstrong, Astrobiology 14, 50–66 (2014)

    ADS  Google Scholar 

  48. C.T. Unterborn et al., Nat. Astron. 2, 297 (2018)

    ADS  Google Scholar 

  49. Y.M. Bar-On, R. Phillips, R. Milo, Proc. Natl. Acad. Sci. 115, 6506 (2018)

    Google Scholar 

  50. T.E. Morris, Princ. Planet. Biol. Lecture Notes, Ch. 5 (1999), http://www.planetarybiology.com. Accessed 5 Feb 2018

  51. S.J. Desch, S. Kane, C.M. Lisse, et al., A white paper for the the “Astrobiology Science Strategy for the Search for Life in the Universe” program by the National Academy of Sciences. Preprint arXiv:1801.06935v1 (2018)

  52. D.M. Glaser, H.E. Hartnett, S.J. Desch et al., Astrophys. J. 893(2), 1538–4357 (2020)

    Google Scholar 

  53. O. Abramov, S.J. Mojzsis, Earth Planet. Sci. Lett. 442, 108 (2016)

    ADS  Google Scholar 

  54. R. Citron, M. Manga, D. Hemingway, Nature 555, 643 (2018)

    ADS  Google Scholar 

  55. W. Luo, X. Cang, A. Howard, Nat. Commun. 8, 15766 (2017)

    ADS  Google Scholar 

  56. D.J. Stevenson, Nature 412(6843), 214–219 (2001)

    ADS  Google Scholar 

  57. S. Onofri, J.-P. de Vera et al., Astrobiology 15(12), 1052 (2015)

    ADS  Google Scholar 

  58. F. Martín-Torres, M.-P. Zorzano, P. Valentín-Serrano et al., Nat. Geosci. 8, 357–361 (2015)

    ADS  Google Scholar 

  59. F. Salese, M. Pondrelli, A. Neeseman, G. Schmidt, G.G. Ori, J. Geophys. Res. Planets 124, 374–395 (2019)

    ADS  Google Scholar 

  60. R.D. Wordsworth, The climate of early mars. Annu. Rev. Earth Planet. Sci. 44, 381 (2016)

    ADS  Google Scholar 

  61. M. Safonova, C. Sivaram, in Planet Formation and Panspermia. New Prospects for the Movement of Life through Space, [PNSP, Volume in the series Astrobiology Perspectives on Life of the Universe, Series Eds: R. Gordon & J. Seckbach, 2021] B. Vukotić, ed. by J. Seckbach, R. Gordon (2021) (ISBN: 9781119640394)

  62. J.M. Kashyap, M. Safonova, S.B. Gudennavar, ESI and MSI data sets 2, Mendeley Data, v8 (2020). https://doi.org/10.17632/c37bvvxp3z.8

  63. L.N. Irwin, A. Méndez, A.G. Fairén, D. Schulze-Makuch, Challenges 5, 159 (2014)

    ADS  Google Scholar 

  64. W. von Bloh, C. Bounama, M. Cuntz et al., Astron. Astrophys. 476, 1365 (2007)

    ADS  Google Scholar 

  65. J.M.R. Rodríguez-Mozos, A. Moya, Mon. Not. R. Astron. Soc. 471(4), 4628–4636 (2017)

    ADS  Google Scholar 

  66. G. Ginde, S. Saha, A. Mathur, S. Venkatagiri, S. Vadakkepat, A. Narasimhamurthy, B.S. Daya Sagar, J. Scientometr. 107(1), 1–51 (2016)

  67. S. Saha, J. Sarkar, A. Dwivedi, N. Dwivedi, A.M. Narasimhamurthy, R. Ranjan, J. Cloud Comput. 5(1), 1–23 (2016)

    Google Scholar 

  68. S. Basak, S. Saha, A. Mathur, K. Bora, S. Makhija, M. Safonova, S. Agrawal, Astron. Comput. 30, 100335 (2020)

    Google Scholar 

  69. K.J. Arrow, H.B. Chenery, B.S. Minhas, R.M. Solow, Rev. Econ. Stat. 43, 225 (1961)

    Google Scholar 

  70. S. Saha, A. Mathur, K. Bora, S. Basak, S. Agrawal, in Proc. 2018 international conference on advances in computing, communications and informatics (ICACCI), pp. 1781–1786 (2018), Bangalore. https://doi.org/10.1109/ICACCI.2018.8554460

  71. S. Saha, N. Nagaraj, A. Mathur, R. Yedida, H.R. Sneha, Eur. Phys. J. Spec. Top. 229, 2629–2738 (2020)

    Google Scholar 

  72. R. Yedida, S. Saha, T. Prashanth, Appl. Intell. (2020). https://doi.org/10.1007/s10489-020-01892-0

    Article  Google Scholar 

  73. H.N. Balakrishnan, A. Kathpalia, S. Saha, N. Nagaraj, Chaos Interdiscip. J. Nonlinear Sci. 29, 113125 (2019)

    Google Scholar 

  74. R. Yedida S. Saha, A novel adaptive learning rate scheduler for deep neural networks. Preprint arXiv:1902.07399 (2019)

  75. D.S. Stevenson, S. Large, Int. J. Astrobiol. 18(3), 204–208 (2017)

    ADS  Google Scholar 

  76. E.W. Schwieterman, C.T. Reinhard, S.L. Olson, C.E. Harman, T.W. Lyons, Astrophys. J. 878, 19 (2019)

    ADS  Google Scholar 

  77. A. Minai, R. Williams, Neural Netw. 6, 845–853 (1993)

    Google Scholar 

  78. P. Ramachandran, B. Zoph, Q. Le, Preprint arxiv:1710.05941 (2017)

  79. D. Misra, in Proc. 31st British machine vision conference (BMVC), (2020)

Download references

Acknowledgements

This work was partially funded by the Department of Science and Technology (DST), Government of India, under the Women Scientist Scheme A (WOS-A); project reference number SR/WOS-A/PM-17/2019 (G). This research has made use of the Exoplanets Data Explorer (http://exoplanets.org) and NASA Astrophysics Data System Abstract Service.

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed equally to the present research.

Corresponding author

Correspondence to Margarita Safonova.

Appendices

Appendices

An introduction to machine learning (ML)

ML is a driving component of AI that enable machines to infer information from data. Analogous to the way humans learn from past experiences, ML empowers computers to make use of data and enables them to learn and improve their experiences on continuous basis. Data plays an important role in making a network learn its task and concepts. As more data is fed to machines, they become more experienced (like humans) and eventually become better at the stipulated task, without having to be programmed explicitly. When previously unseen data is fed to a machine, it makes use of information obtained from already perceived data to make decisions. In this manner, the machine learns to perform tasks by gaining insightful information from the data, and eventually evolves itself to accomplish more challenging tasks. The data which a ML algorithm uses is referred as training data. Besides this, there exists test data (unseen data), which is utilized to check predictions made by the trained algorithm. If predictions made by the algorithm are not close to the desired value, the algorithm is trained again on a new set of values till it attains the desired accuracy.

Classical ML algorithms can be categorized as supervised and unsupervised. Supervised learning makes use of labeled data whereas unsupervised learning uses unlabeled data to train the model. In either case, the algorithm tries to search for a pattern in the data, a pattern that can be used later for identification of unknown (test) data. Polynomial regression, Random Forest, Linear Regression, Logistic Regression, Decision Tree, KNN and Naive Bayes are all supervised learning algorithms, while K-Means Clustering, Apriori, Hierarchical Clustering, Principal Component Analysis and Singular Value Decomposition fall under unsupervised learning. A slightly different and less popular category is the Reinforcement Learning. Reinforcement Learning makes use of the three components—agents, environment and actions—to achieve the defined target. An agent is a decision maker which gets to understand behavior of the environment and takes necessary actions to achieve its goal. Rewards are computed for every chosen action. Strategically, the agent interacts with its environment and chooses action that generates maximum rewards. With trial and error, the agent discovers data, generate possible actions, computes reward for each action and keep record of actions that reaches to the target.

1.1 Working of neural networks

A robust ML tool, Neural Network, capable to perform computationally challenging tasks on complex datasets, is inspired from the working of the human brain. The network is made up of processing neurons that are functionally similar to neurons existing in brains. These neurons are arranged in layers across the network, from one end to another and, between two layers, neurons are connected with synaptic weights. A synaptic weight describes the strength of the connection between neurons. This structure enables the processing units to perform high-speed parallel computing on input data. Neurons are equipped with a special function called activation function AF, which squashes the neuron input and brings it to a specific range before propagating it to other layers. There are three types of AFs: binary step function, linear and non-linear AFs. Binary and linear AFs are not suitable to be used in back-propagation (will be explored later) owing to the fact that the derivative of the function becomes constant. On the other hand, non-linear AF allows the network to perform complex mapping of input and output and thus facilitates modelling complex data to solve problems related to the Computer Vision, Artificial Intelligence, Deep Learning and Natural Language Processing.

Essentially, neural networks work on the principle of back-propagation algorithm which runs in two phases to achieve classification on given data set. Figure 6 shows a simple NN, consisting of one hidden layer with 5 hidden units.

Fig. 6
figure 6

A sample neural network

The network weights are initialized with random values. To attain the right set of network weights, a training set is chosen as input such that the network tries to identify patterns from the training values to make correct future predictions. The first phase involves computation of inputs of neurons of first layer. AF converts these inputs into a definite range and propagates it to neurons of connected layer. The process goes on till the output is received at output layer. In the second phase, error (cost) associated with training sample is computed and error gradients are propagated back to the network in the form of weight updates. The process is repeated for multiple training samples in numerous iterations or epochs, till the expected output is received at the output neurons.

1.2 Activation functions (AF)

It must be admitted that the machine classification is a non-trivial task as the classes are non-linearly separated from each other. There is a large number of linear and non-linear AFs currently in use in Neural Networks. These mathematical functions are tied-up to every neuron in the network to facilitate decision making while solving a complex problem at hand. The function output allow neurons to determine whether it should be activated or not. Moreover, the function also normalizes output values of neurons and brings it to either [0, 1] or \([-1,1]\) range. In this section, we explore the non-linear AFs as they speed up solving complex problems by stacking of multiple hidden layers thus forming deep layered architectures.

1.2.1 Sigmoid function

Also known as logistic function, Sigmoid function [77] looks like a S-shaped curve which maps real values between 0 and 1. The mathematical notation of the function is given as

$$\begin{aligned} S(x) = \frac{1}{1 + \mathrm {e}^{-x}}\,. \end{aligned}$$
(A.1)

Sigmoid suffers from vanishing gradient problem as it is evident that for very large and very small values of x, the gradient of the function becomes zero.

1.2.2 ReLU

Rectified linear unit (ReLU) is a function that returns zero for negative and identity for positive inputs. Mathematically, the function looks like the following:

$$\begin{aligned} S(x) = \mathrm{max}(0,x)\,. \end{aligned}$$
(A.2)

Computation of gradient is an crucial step in back propagation and unlike Sigmoid, ReLU provides a constant gradient for positive inputs. But visibly, ReLU’s gradient and also the output from neurons is zero for negative values.

1.2.3 A-ReLU

Approximate ReLu (A-ReLU) is another variant of ReLU that is defined by

$$\begin{aligned} S\left( x \right) = \left\{ \begin{matrix} 0 &{} x\le 0\\ kx^{n} &{} x> 0 \end{matrix}\right. . \end{aligned}$$
(A.3)

The function, unlike ReLU, is differentiable at \(x=0\) and, hence, it can be shown that its derivative is continuous.

1.2.4 Swish

This AF, proposed by the Google Brain team [78], works better than ReLU in various Deep Learning architectures on the most challenging data. The smooth, non-monotonic function is mathematically represented as

$$\begin{aligned} S(x) = \frac{x}{1 + \mathrm {e}^{-x}}\,. \end{aligned}$$
(A.4)

Experimentally, it was observed that Swish and ReLU perform at par till 40 layers in Deep Learning architectures. Beyond 40 layers, Swish outperforms ReLU, which is particularly seen on few datasets while training deep layered architectures. However, on shallow networks this might not be true.

1.3 SBAF—classification of exoplanets for cross-validating metrics

AFs such as Sigmoid suffer from local oscillation and flat mean-square-error (MSE) problems. This means that over successive iterations, the error between the target and predicted labels may not minimize due to the oscillating local minima. Saha et al. [70] mitigated this problem by proposing a new AF, SBAFFootnote 5,

$$\begin{aligned} y = \frac{1}{1 + kx^{\alpha }(1-x)^{1-\alpha }} \,, \end{aligned}$$
(A.5)

where where k and \(\alpha \) are the parameters of the function, and the derivative is computed as

$$\begin{aligned} \frac{dy}{dx} = -y^{2} \cdot \frac{\alpha -x}{x(1-x)} \cdot \frac{1-y}{y} \frac{y(1-y)}{x(1-x)}\cdot (x-\alpha )\,. \end{aligned}$$
(A.6)

SBAF is a non-monotonic function that does not exhibit vanishing gradient problem during training. The characteristics of the function indicate the presence of local minima and maxima not present in other popularly used AFs. It is also shown that unlike the Sigmoid, this function does not have a saddle point [71]. It is also argued that AFs need not be monotonic at all when analyzed over the entire domain. The following properties of SBAF help to achieve a near-perfect classification of exoplanets, matching habitability labels with Earth similarity:

  • Does not admit a saddle point;

  • Has local minima at \(x=\alpha \);

  • Satisfies Universal Approximation theorem;

  • Does not admit flat MSE;

  • Does not suffer from vanishing gradient problem.

An evaluation of SBAF is incomplete without stating its importance in the context of other AFs described above. It is also observed that the classification performance of classical ML algorithms (Support Vector Machine, Random Forest, Naive Bayes, etc.) drops abysmally when applied on the restricted-feature data.

However, pure empiricism does not determine the existence of an AF, its state-of-the-art (SOTA) performance notwithstanding. AFs such as Swish and Mish claim SOTA results on complex neural network architectures. But, are these even AFs? An AF has to be discriminatory in the sense that it should be able to approximate any nonlinear function describing the data over a neural network [71]. There is no discussion in the literature on this property known as the Universal Approximation [71]. Does this mean that we dilute an AF to a level of just being continuous and smooth? In other words, can any function be an AF? The answer is an emphatic negative, but the Deep Learning literature is laden with such empirical observations. Additionally, an AF needs to be smooth everywhere (no singularity), so that the first derivative can be used in back propagation in tandem with the Lipschitz continuity [72] for fast acceleration [73]. RELU suffers from this singularity problem. Sigmoid is not devoid of problems either. It flattens out for most values in the input range and is, therefore, only suited for inputs normalized to [0, 1]. It has its second derivative zero, which explains the difficulty of Sigmoid overcoming local minima/maxima. Another important fact that SBAF helped establish in the significant work of Saha et al. [71] is that we no longer need AFs to be monotone. This was a common wisdom in the Deep Learning community but based purely on empirical observation.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Safonova, M., Mathur, A., Basak, S. et al. Quantifying the classification of exoplanets: in search for the right habitability metric. Eur. Phys. J. Spec. Top. 230, 2207–2220 (2021). https://doi.org/10.1140/epjs/s11734-021-00211-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1140/epjs/s11734-021-00211-z

Navigation