Abstract
What is habitability? Can we quantify it? What do we mean under the term habitable or potentially habitable planet? With estimates of the number of planets in our Galaxy alone running into billions, possibly a number greater than the number of stars, it is high time to start characterizing them, sorting them into classes/types just like stars, to better understand their formation paths, their properties and, ultimately, their ability to beget or sustain life. After all, we do have life thriving on one of these billions of planets, why not on others? Which planets are better suited for life and which ones are definitely not worth spending expensive telescope time on? We need to find sort of quick assessment score, a metric, using which we can make a list of promising planets and dedicate our efforts to them. Exoplanetary habitability is a transdisciplinary subject integrating astrophysics, astrobiology, planetary science, and even terrestrial environmental sciences. It became a challenging problem in astroinformatics, an emerging area in computational astronomy. Here, we review the existing metrics of habitability and the new classification schemes (machine learning (ML), neural networks, activation functions) of extrasolar planets, and provide an exposition of the use of computational intelligence techniques to evaluate habitability scores and to automate the process of classification of exoplanets. We examine how solving convex optimization techniques, as in computing new metrics such as Cobb–Douglas habitability score (CDHS) and constant elasticity earth similarity approach (CEESA), cross-validates ML-based classification of exoplanets. Despite the recent criticism of exoplanetary habitability ranking, we are sure that this field has to continue and evolve to use all available machinery of astroinformatics, artificial intelligence (AI) and machine learning. It might actually develop into a sort of same scale as stellar types in astronomy, to be used as a quick tool of screening exoplanets in important characteristics in search for potentially habitable planets (PHPs), or Earth-like planets, for detailed follow-up targets.
Similar content being viewed by others
Data Availability Statement
Data and codes associated with this manuscript are available on several websites: 1. ESI/MSI datasets: https://data.mendeley.com/datasets/c37bvvxp3z/8 2. CDHS/CEESSA Datasets and Catalogs: AstrIRG (Astroinformatics Research Group) website http://astrirg.org/projects.html 3. Codes to analyze exoplanetary data from the PHL HEC catalog: https://github.com/SuryodayBasak/ExoplanetsAnalysis/tree/master/MLAnalysis
Notes
An inscription on the iron pillar at Delhi’s Qutub Minar, probably left in the 4th century BC. Such mentions can also be found in the Mahapuranas—ancient texts of Hinduism (e.g. Rocher [6]).
Life can still exist on Jupiter, in the clouds—the buoyant organisms could exist at depths of several hundred kilometers in the atmosphere; idea originally proposed in 1974 [15]. Though there is no oxygen, even ordinary Earth bacteria was recently shown to live and thrive in a 100% hydrogen atmosphere [16].
Universidad Central “Marta Abreu” de Las Villas, Santa Clara, Villa Clara, Cuba.
Saha Bora Activation Function: https://github.com/sahamath/sym-netv1.
References
J. Pathak, A. Wikner, R. Fussell et al., Chaos Interdiscip. J. Nonlinear Sci. 28(4), 041101 (2018)
Y.G. Zhang, V. Gajjar, G. Foster et al., Astrophys. J. 866, 149 (2018)
N.M. Batalha, J.F. Rowe, S.T. Bryson et al., Astrophys. J. Suppl. 204, 24 (2013)
E.A. Petigura, A.W. Howard, G.W. Marcy, Proc. Natl. Acad. Sci. 110(48), 19273–19278 (2013)
M.C. Turnbull, T. Glassman, A. Roberge et al., Publ. Astron. Soc. Pac. PASP 124, 418 (2012)
L. Rocher, Puranas (Otto Harrassowitz, Wiesbaden, 1986). ISBN 978-3447025225
G. McColley, Ann. Sci. 1, 385–430 (1936)
A.J. Sternfeld, La Nature, Masson et Cie (eds.), Paris, No. 2956, 1–12 (1935) (in French)
G.A. Tikhov, Priroda (Leningrad) 46(2), 3–6 (1947) (in Russian)
T. Sumi, K. Kamiya, D.P. Bennett et al., Nature 473, 349 (2011)
M. Safonova, J. Murthy, Y.A. Shchekinov, Int. J. Astrobiol. 15, 93 (2016)
M. Matsuura, A.A. Zijlstra, F.J. Molster et al., Astrophys. J. 604(7), 91–799 (2004)
E. Schrödinger, What is life? The physical aspect of the living cell (Cambridge University Press, Cambridge, 1944)
G. Witzany, Front. Astron. Space Sci. 7, 7 (2020). https://doi.org/10.3389/fspas.2020.00007
W.F. Libby, Life on Jupiter? Orig. Life Evol. Biosph. 5, 483–486 (1974)
S. Seager, J. Huang, J.J. Petkowski et al., Nat. Astron. 4, 802–806 (2020)
M. Murakami, K. Hirose, H. Yurimoto, S. Nakashima, N. Takafuji, Science 295(5561), 1885 (2002)
D. Atri, J.R. Soc, Interface 13, 20160459 (2016)
J.R. Spear, J.J. Walker, T.M. McCollom, N.R. Pace, Proc. Natl. Acad. Sci. 102, 2555–2560 (2005)
Y. Morono, M. Ito, T. Hoshino et al., Nat. Commun. 11(1), 3626 (2020)
BBC News, Tardigrades: ‘Water Bears’ Stuck on the Moon after Crash (2019), https://www.bbc.com/news/newsbeat-49265125. Accessed 17 Sept 2019
D.S. Spiegel, E.L. Turner, Proc. Natl. Acad. Sci. 109, 395–400 (2012)
J.J. Swift, J.A. Johnson, T.D. Morton et al., Astrophys. J. 764(1), 105 (2013)
M. Kunimoto, J.M. Matthews, Astron. J. 159(6), 248 (2020)
S. Bryson, M. Kunimoto, R.K. Kopparapu, J.L. Coughlin, W.J. Borucki et al., Astron. J. 161, 36 (2021)
R.H. Dicke, Nature 192, 440–441 (1961)
P. Dayal, M. Ward, C. Cockell, Preprint arXiv:1606.09224 (2016)
A. Loeb, Int. J. Astrobiol. 13, 337–339 (2014)
J. Haqq-Misra, R. Kopparapu, E. Wolf, Int. J. Astrobiol. 17(1), 77–86 (2018)
T.M. McCollom, Proc. Natl. Acad. Sci. 113(49), 13965–13970 (2011)
F. Klein, N.G. Grozeva, J.S. Seewald, Proc. Natl. Acad. Sci. 116(36), 17666–17672 (2019)
C. Oze, L.C. Jones, J.I. Goldsmith, R.J. Rosenbauer, Proc. Natl. Acad. Sci. 109(25), 9750 (2012)
L.O. Stephanie, W.S. Edward, T.R. Christopher et al., Astrophys. J. 858, L14 (2012)
J. Krissansen-Totton, S. Olson, D.C. Catling, Sci. Adv. 4, eaao5747 (2018)
D. Schulze-Makuch, A. Méndez, A.G. Fairén et al., Astrobiology 11, 1041 (2011)
R. Barnes, V.S. Meadows, N. Evans, Astrophys. J. 814, 91 (2015)
J.M. Kashyap, S.B. Gudennavar, U. Doshi, M. Safonova, Astrophys. Space Sci. 362(8), 146 (2017)
A. Méndez, in Proceedings of sixth astrobiology science conference, Houston, TX, USA, 26–29 April 2010
R. Cardenas, N. Perez, J. Martinez-Frias, O. Martin, Challenges 5, 284 (2014)
K. Bora, S. Saha, S. Agrawal, M. Safonova, S. Routh, A. Narasimhamurthy, Astron. Comput. 17, 129–143 (2016)
S. Saha, S. Basak, M. Safonova, K. Bora, S. Agrawal, P. Sarkar, J. Murthy, Astron. Comput. 23, 141 (2018)
E. Tasker, J. Tan, K. Heng, S. Kane, D.L. Spiege, the ELSI Origins Network Planetary Diversity Workshop, Nat. Astron. 1, 0042 (2017)
J.F. Kasting, D.P. Whitmire, R.T. Reynolds, Icarus 101, 108 (1993)
D.J. Stevenson, Nature 400, 32 (1999)
M.A. Limbach, E.L. Turner, Proc. Natl. Acad. Sci. 112, 20 (2015)
Y. Wang, Y. Liu, F. Tian, Y. Hu, Y. Huang, Preprint arXiv:1710.01405 (2017)
R. Heller, J. Armstrong, Astrobiology 14, 50–66 (2014)
C.T. Unterborn et al., Nat. Astron. 2, 297 (2018)
Y.M. Bar-On, R. Phillips, R. Milo, Proc. Natl. Acad. Sci. 115, 6506 (2018)
T.E. Morris, Princ. Planet. Biol. Lecture Notes, Ch. 5 (1999), http://www.planetarybiology.com. Accessed 5 Feb 2018
S.J. Desch, S. Kane, C.M. Lisse, et al., A white paper for the the “Astrobiology Science Strategy for the Search for Life in the Universe” program by the National Academy of Sciences. Preprint arXiv:1801.06935v1 (2018)
D.M. Glaser, H.E. Hartnett, S.J. Desch et al., Astrophys. J. 893(2), 1538–4357 (2020)
O. Abramov, S.J. Mojzsis, Earth Planet. Sci. Lett. 442, 108 (2016)
R. Citron, M. Manga, D. Hemingway, Nature 555, 643 (2018)
W. Luo, X. Cang, A. Howard, Nat. Commun. 8, 15766 (2017)
D.J. Stevenson, Nature 412(6843), 214–219 (2001)
S. Onofri, J.-P. de Vera et al., Astrobiology 15(12), 1052 (2015)
F. Martín-Torres, M.-P. Zorzano, P. Valentín-Serrano et al., Nat. Geosci. 8, 357–361 (2015)
F. Salese, M. Pondrelli, A. Neeseman, G. Schmidt, G.G. Ori, J. Geophys. Res. Planets 124, 374–395 (2019)
R.D. Wordsworth, The climate of early mars. Annu. Rev. Earth Planet. Sci. 44, 381 (2016)
M. Safonova, C. Sivaram, in Planet Formation and Panspermia. New Prospects for the Movement of Life through Space, [PNSP, Volume in the series Astrobiology Perspectives on Life of the Universe, Series Eds: R. Gordon & J. Seckbach, 2021] B. Vukotić, ed. by J. Seckbach, R. Gordon (2021) (ISBN: 9781119640394)
J.M. Kashyap, M. Safonova, S.B. Gudennavar, ESI and MSI data sets 2, Mendeley Data, v8 (2020). https://doi.org/10.17632/c37bvvxp3z.8
L.N. Irwin, A. Méndez, A.G. Fairén, D. Schulze-Makuch, Challenges 5, 159 (2014)
W. von Bloh, C. Bounama, M. Cuntz et al., Astron. Astrophys. 476, 1365 (2007)
J.M.R. Rodríguez-Mozos, A. Moya, Mon. Not. R. Astron. Soc. 471(4), 4628–4636 (2017)
G. Ginde, S. Saha, A. Mathur, S. Venkatagiri, S. Vadakkepat, A. Narasimhamurthy, B.S. Daya Sagar, J. Scientometr. 107(1), 1–51 (2016)
S. Saha, J. Sarkar, A. Dwivedi, N. Dwivedi, A.M. Narasimhamurthy, R. Ranjan, J. Cloud Comput. 5(1), 1–23 (2016)
S. Basak, S. Saha, A. Mathur, K. Bora, S. Makhija, M. Safonova, S. Agrawal, Astron. Comput. 30, 100335 (2020)
K.J. Arrow, H.B. Chenery, B.S. Minhas, R.M. Solow, Rev. Econ. Stat. 43, 225 (1961)
S. Saha, A. Mathur, K. Bora, S. Basak, S. Agrawal, in Proc. 2018 international conference on advances in computing, communications and informatics (ICACCI), pp. 1781–1786 (2018), Bangalore. https://doi.org/10.1109/ICACCI.2018.8554460
S. Saha, N. Nagaraj, A. Mathur, R. Yedida, H.R. Sneha, Eur. Phys. J. Spec. Top. 229, 2629–2738 (2020)
R. Yedida, S. Saha, T. Prashanth, Appl. Intell. (2020). https://doi.org/10.1007/s10489-020-01892-0
H.N. Balakrishnan, A. Kathpalia, S. Saha, N. Nagaraj, Chaos Interdiscip. J. Nonlinear Sci. 29, 113125 (2019)
R. Yedida S. Saha, A novel adaptive learning rate scheduler for deep neural networks. Preprint arXiv:1902.07399 (2019)
D.S. Stevenson, S. Large, Int. J. Astrobiol. 18(3), 204–208 (2017)
E.W. Schwieterman, C.T. Reinhard, S.L. Olson, C.E. Harman, T.W. Lyons, Astrophys. J. 878, 19 (2019)
A. Minai, R. Williams, Neural Netw. 6, 845–853 (1993)
P. Ramachandran, B. Zoph, Q. Le, Preprint arxiv:1710.05941 (2017)
D. Misra, in Proc. 31st British machine vision conference (BMVC), (2020)
Acknowledgements
This work was partially funded by the Department of Science and Technology (DST), Government of India, under the Women Scientist Scheme A (WOS-A); project reference number SR/WOS-A/PM-17/2019 (G). This research has made use of the Exoplanets Data Explorer (http://exoplanets.org) and NASA Astrophysics Data System Abstract Service.
Author information
Authors and Affiliations
Contributions
All authors contributed equally to the present research.
Corresponding author
Appendices
Appendices
An introduction to machine learning (ML)
ML is a driving component of AI that enable machines to infer information from data. Analogous to the way humans learn from past experiences, ML empowers computers to make use of data and enables them to learn and improve their experiences on continuous basis. Data plays an important role in making a network learn its task and concepts. As more data is fed to machines, they become more experienced (like humans) and eventually become better at the stipulated task, without having to be programmed explicitly. When previously unseen data is fed to a machine, it makes use of information obtained from already perceived data to make decisions. In this manner, the machine learns to perform tasks by gaining insightful information from the data, and eventually evolves itself to accomplish more challenging tasks. The data which a ML algorithm uses is referred as training data. Besides this, there exists test data (unseen data), which is utilized to check predictions made by the trained algorithm. If predictions made by the algorithm are not close to the desired value, the algorithm is trained again on a new set of values till it attains the desired accuracy.
Classical ML algorithms can be categorized as supervised and unsupervised. Supervised learning makes use of labeled data whereas unsupervised learning uses unlabeled data to train the model. In either case, the algorithm tries to search for a pattern in the data, a pattern that can be used later for identification of unknown (test) data. Polynomial regression, Random Forest, Linear Regression, Logistic Regression, Decision Tree, KNN and Naive Bayes are all supervised learning algorithms, while K-Means Clustering, Apriori, Hierarchical Clustering, Principal Component Analysis and Singular Value Decomposition fall under unsupervised learning. A slightly different and less popular category is the Reinforcement Learning. Reinforcement Learning makes use of the three components—agents, environment and actions—to achieve the defined target. An agent is a decision maker which gets to understand behavior of the environment and takes necessary actions to achieve its goal. Rewards are computed for every chosen action. Strategically, the agent interacts with its environment and chooses action that generates maximum rewards. With trial and error, the agent discovers data, generate possible actions, computes reward for each action and keep record of actions that reaches to the target.
1.1 Working of neural networks
A robust ML tool, Neural Network, capable to perform computationally challenging tasks on complex datasets, is inspired from the working of the human brain. The network is made up of processing neurons that are functionally similar to neurons existing in brains. These neurons are arranged in layers across the network, from one end to another and, between two layers, neurons are connected with synaptic weights. A synaptic weight describes the strength of the connection between neurons. This structure enables the processing units to perform high-speed parallel computing on input data. Neurons are equipped with a special function called activation function AF, which squashes the neuron input and brings it to a specific range before propagating it to other layers. There are three types of AFs: binary step function, linear and non-linear AFs. Binary and linear AFs are not suitable to be used in back-propagation (will be explored later) owing to the fact that the derivative of the function becomes constant. On the other hand, non-linear AF allows the network to perform complex mapping of input and output and thus facilitates modelling complex data to solve problems related to the Computer Vision, Artificial Intelligence, Deep Learning and Natural Language Processing.
Essentially, neural networks work on the principle of back-propagation algorithm which runs in two phases to achieve classification on given data set. Figure 6 shows a simple NN, consisting of one hidden layer with 5 hidden units.
The network weights are initialized with random values. To attain the right set of network weights, a training set is chosen as input such that the network tries to identify patterns from the training values to make correct future predictions. The first phase involves computation of inputs of neurons of first layer. AF converts these inputs into a definite range and propagates it to neurons of connected layer. The process goes on till the output is received at output layer. In the second phase, error (cost) associated with training sample is computed and error gradients are propagated back to the network in the form of weight updates. The process is repeated for multiple training samples in numerous iterations or epochs, till the expected output is received at the output neurons.
1.2 Activation functions (AF)
It must be admitted that the machine classification is a non-trivial task as the classes are non-linearly separated from each other. There is a large number of linear and non-linear AFs currently in use in Neural Networks. These mathematical functions are tied-up to every neuron in the network to facilitate decision making while solving a complex problem at hand. The function output allow neurons to determine whether it should be activated or not. Moreover, the function also normalizes output values of neurons and brings it to either [0, 1] or \([-1,1]\) range. In this section, we explore the non-linear AFs as they speed up solving complex problems by stacking of multiple hidden layers thus forming deep layered architectures.
1.2.1 Sigmoid function
Also known as logistic function, Sigmoid function [77] looks like a S-shaped curve which maps real values between 0 and 1. The mathematical notation of the function is given as
Sigmoid suffers from vanishing gradient problem as it is evident that for very large and very small values of x, the gradient of the function becomes zero.
1.2.2 ReLU
Rectified linear unit (ReLU) is a function that returns zero for negative and identity for positive inputs. Mathematically, the function looks like the following:
Computation of gradient is an crucial step in back propagation and unlike Sigmoid, ReLU provides a constant gradient for positive inputs. But visibly, ReLU’s gradient and also the output from neurons is zero for negative values.
1.2.3 A-ReLU
Approximate ReLu (A-ReLU) is another variant of ReLU that is defined by
The function, unlike ReLU, is differentiable at \(x=0\) and, hence, it can be shown that its derivative is continuous.
1.2.4 Swish
This AF, proposed by the Google Brain team [78], works better than ReLU in various Deep Learning architectures on the most challenging data. The smooth, non-monotonic function is mathematically represented as
Experimentally, it was observed that Swish and ReLU perform at par till 40 layers in Deep Learning architectures. Beyond 40 layers, Swish outperforms ReLU, which is particularly seen on few datasets while training deep layered architectures. However, on shallow networks this might not be true.
1.3 SBAF—classification of exoplanets for cross-validating metrics
AFs such as Sigmoid suffer from local oscillation and flat mean-square-error (MSE) problems. This means that over successive iterations, the error between the target and predicted labels may not minimize due to the oscillating local minima. Saha et al. [70] mitigated this problem by proposing a new AF, SBAFFootnote 5,
where where k and \(\alpha \) are the parameters of the function, and the derivative is computed as
SBAF is a non-monotonic function that does not exhibit vanishing gradient problem during training. The characteristics of the function indicate the presence of local minima and maxima not present in other popularly used AFs. It is also shown that unlike the Sigmoid, this function does not have a saddle point [71]. It is also argued that AFs need not be monotonic at all when analyzed over the entire domain. The following properties of SBAF help to achieve a near-perfect classification of exoplanets, matching habitability labels with Earth similarity:
-
Does not admit a saddle point;
-
Has local minima at \(x=\alpha \);
-
Satisfies Universal Approximation theorem;
-
Does not admit flat MSE;
-
Does not suffer from vanishing gradient problem.
An evaluation of SBAF is incomplete without stating its importance in the context of other AFs described above. It is also observed that the classification performance of classical ML algorithms (Support Vector Machine, Random Forest, Naive Bayes, etc.) drops abysmally when applied on the restricted-feature data.
However, pure empiricism does not determine the existence of an AF, its state-of-the-art (SOTA) performance notwithstanding. AFs such as Swish and Mish claim SOTA results on complex neural network architectures. But, are these even AFs? An AF has to be discriminatory in the sense that it should be able to approximate any nonlinear function describing the data over a neural network [71]. There is no discussion in the literature on this property known as the Universal Approximation [71]. Does this mean that we dilute an AF to a level of just being continuous and smooth? In other words, can any function be an AF? The answer is an emphatic negative, but the Deep Learning literature is laden with such empirical observations. Additionally, an AF needs to be smooth everywhere (no singularity), so that the first derivative can be used in back propagation in tandem with the Lipschitz continuity [72] for fast acceleration [73]. RELU suffers from this singularity problem. Sigmoid is not devoid of problems either. It flattens out for most values in the input range and is, therefore, only suited for inputs normalized to [0, 1]. It has its second derivative zero, which explains the difficulty of Sigmoid overcoming local minima/maxima. Another important fact that SBAF helped establish in the significant work of Saha et al. [71] is that we no longer need AFs to be monotone. This was a common wisdom in the Deep Learning community but based purely on empirical observation.
Rights and permissions
About this article
Cite this article
Safonova, M., Mathur, A., Basak, S. et al. Quantifying the classification of exoplanets: in search for the right habitability metric. Eur. Phys. J. Spec. Top. 230, 2207–2220 (2021). https://doi.org/10.1140/epjs/s11734-021-00211-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1140/epjs/s11734-021-00211-z