Abstract
Background
Motivated by the size and availability of cell line drug sensitivity data, researchers have been developing machine learning (ML) models for predicting drug response to advance cancer treatment. As drug sensitivity studies continue generating drug response data, a common question is whether the generalization performance of existing prediction models can be further improved with more training data.
Methods
We utilize empirical learning curves for evaluating and comparing the data scaling properties of two neural networks (NNs) and two gradient boosting decision tree (GBDT) models trained on four cell line drug screening datasets. The learning curves are accurately fitted to a power law model, providing a framework for assessing the data scaling behavior of these models.
Results
The curves demonstrate that no single model dominates in terms of prediction performance across all datasets and training sizes, thus suggesting that the actual shape of these curves depends on the unique pair of an ML model and a dataset. The multiinput NN (mNN), in which gene expressions of cancer cells and molecular drug descriptors are input into separate subnetworks, outperforms a singleinput NN (sNN), where the cell and drug features are concatenated for the input layer. In contrast, a GBDT with hyperparameter tuning exhibits superior performance as compared with both NNs at the lower range of training set sizes for two of the tested datasets, whereas the mNN consistently performs better at the higher range of training sizes. Moreover, the trajectory of the curves suggests that increasing the sample size is expected to further improve prediction scores of both NNs. These observations demonstrate the benefit of using learning curves to evaluate prediction models, providing a broader perspective on the overall data scaling characteristics.
Conclusions
A fitted power law learning curve provides a forwardlooking metric for analyzing prediction performance and can serve as a codesign tool to guide experimental biologists and computational scientists in the design of future experiments in prospective research studies.
Background
Human cancer cell lines remain a primary cancermimicking environment in a laboratory setting for understanding the molecular biology of this complex disease [1,2,3]. In the search for anticancer treatments, in vitro drug sensitivity assays serve as a standard, highthroughput experimental platform for measuring the response of cancer cells to drug treatments. The standardized protocols of sensitivity assays, along with rapid improvement of technologies for genomic profiling, have led researchers to generate large pharmacogenomic drug response datasets for anticancer drug discovery [4,5,6]. Considering the scale and diversity of tumors and compounds in these datasets, machine learning (ML) techniques have become a natural fit for analytically predicting the response of cell lines to drug treatments. By maneuvering through a landscape of computational approaches and numerical representations of tumors and drugs, researchers strive to develop highly predictive ML drug response models [7,8,9]. Demonstrating the accuracy and robustness of prediction models is essential in order to identify their potential utility for clinical applications in cancer treatment including precision oncology and drug repurposing.
In MLdriven cancer research, a common question is whether existing predictive models can be further improved with more training data. Given recent advances in artificial neural networks (NNs), deep learning (DL) methods have become a favorite approach across a variety of scientific disciplines for discovering hidden patterns in large volumes of complex data. This trend is also observed in medical applications, including the prediction of drug response in cancer cell lines [10,11,12,13,14]. Regardless of the learning algorithm, supervised learning models are expected to improve generalization performance with increasing amounts of highquality labeled data. Generalization performance refers to the aggregated accuracy of model predictions on a set of unseen data samples. Analytically estimating the learning capacity of models is a challenging task. Alternatively, given a dataset and a learning algorithm, the projected improvement of predictions with increasing number of training samples can be empirically estimated by using learning curves.
A learning curve is a plot of the generalization performance of a supervised learning model as a function of training set size (Fig. 1). These curves have been explored as an efficient method for modeling the power law relationship, \(s(m)\propto am^{b}\), between the generalization score s (such as generalization error or accuracy) and the number of training samples m, where a and b are two parameters of the power law model. The power law characteristics of learning curves can provide insights into the data scaling behavior of drug response prediction models, which otherwise could not be investigated by merely analyzing singlevalue performance measures obtained with the full training set size.
A main bottleneck of utilizing learning curves, however, is often the limited availability of sufficient computational resources for performing the analysis. Particularly challenging is analysis with DL models and large datasets because of the large computational cost. While learning curves have been explored in a variety of smallscale applications with classical ML [15,16,17,18], only a few recent studies have applied DL methods to large benchmark datasets in vision and text applications [19,20,21]. To the best of our knowledge, learning curves of drug response prediction models have not been previously explored.
In this paper, we utilize learning curves to evaluate the data scaling properties of drug response prediction models in cancer cell lines. The primary objective of fitting a power law expression to raw data is twofold: (1) efficient and accurate estimation of prediction performance with a larger, not yet available or computationally prohibitive dataset, and (2) fair and systematic comparison of prediction models across various learning algorithms and datasets. To that end, we perform a systematic comparison between classical ML and DL models, implemented with large pharmacogenomic drug response datasets. To accomplish these objectives, we develop an efficient computational workflow, leveraging highperformance computing (HPC) resources to conduct the largescale analysis. We use this workflow for generating learning curve data with gradient boosting decision tree (GBDT) models and NNs, where each model is trained on four large drug response datasets of cancer cell lines. To assess the data scaling trajectory of each datasetmodel pair, the power law expression is fitted to the raw learning curve data to generate a learning curve and uncertainty estimates of the curve. We apply this methodology to analyze sixteen datasetmodel combinations.
Learning curves
Theoretical [22, 23] and empirical [15, 20, 21] studies demonstrate that learning curves of predictive models are characterized by a power law relationship between the training set size m and the generalization score s,
where \(\beta =(a,b,c)\) is the set of the power law parameters. The parameters in \(\beta\) determine the shape of the curve and typically vary for individual combinations of learning algorithm, prediction task, and data.
An empirical learning curve often exhibits three primary learning regions: smalldata, power law, and irreducible error [20]. Figure 1 provides a glimpse into the experimental results discussed later. All three learning regions of the power law expression in Eq. (1) are distinguishable on the loglog plot in Fig. 1b. Alternatively, the linear scale representation in Fig. 1a visually obscures these learning trends.
The smalldata region is attributed to the lack of an appropriate amount of training samples for learning a sufficiently generalizable model. This region produces models that perform as good as, or slightly better than, random (or best) guessing and therefore is commonly known as the region of random guess [20, 21]. On a loglog plot, the random guessing is characterized primarily by a horizontal flat region, followed by a transition to the power law region.
From random guessing, the curve transitions to the power law region described by the \(am^{b}\) term. In this region, the curve maintains a steady trajectory, as shown by the approximately constant slope on the loglog plot in Fig. 1b. The parameter b, ranging between \(0<b<1\), is the scaling exponent and determines the steepness of the curve in the power law region [16, 20, 24].
As the training size increases, the model starts to exhaust its learning capacity, gradually approaching a plateau, known as the irreducible asymptotic error [16, 20, 21]. The constant term c in Eq. (1) accounts for a smooth transition from the power law region into this plateau. When this convergence region becomes apparent in the plot, it implies that the model is not expected to significantly improve with more training data, providing researchers with valuable information for future directions in their attempts to improve model predictions.
Learning curves provide intuitive insight into the data scaling behavior of prediction performance, as opposed to singlevalue performance measures obtained with the entire set of training samples. The shape of these curves facilitates comparison between ML models by illustrating a global trajectory of model improvement. Thus, learning curves can be utilized for quantifying the learning capacity of prediction models with increasing amounts of training data.
Methods
This section describes the drug response datasets, learning algorithms, training procedures, and methodology for generating and fitting learning curves.
Drug response datasets
The four datasets used for the experiments are listed in Table 1. The data comes from public repositories of drug sensitivity studies: the Genomics of Drug Sensitivity in Cancer project, which includes GDSC1 and GDSC2 datasets [4]; the Cancer Therapeutics Response Portal (CTRP v2) [5]; and the NCI60 Human Tumor Cell Lines Screen (NCI60) [6].
In drug sensitivity data, the drug response of a cancer cell line to a drug treatment is measured by the percentage of viable cells at multiple drug doses. A threeparameter Hill–Slope model was used to fit the dose response curve for each celldrug pair. To retain highquality response data, we removed samples in which the \({R^2}\) of the dose response curve fit was lower than 0.3. The fitted curve was used to calculate the area under the curve (AUC) over a dose range of \([10^{10}\)M, \(10^{4}\)M]. The AUC value was then normalized by the dose range to take a value between 0 and 1. We note that lower AUC values indicate higher growth inhibition (i.e., stronger response to drug treatment). Figure 2 shows the histograms of the response values.
The NCI60 dataset originally included more than 3 million samples, where a sample refers to a drug response of a celldrug pair. We randomly selected 750,000 samples from the full NCI60 collection for our analysis. For the GDSC1, GDSC2, and CTRP datasets, we collected all the available samples for which we were able to retrieve or calculate feature representations of cells and drugs.
Representation of cell lines
A variety of genomic representations of cancer cells have been used as input features for drug response prediction [25,26,27]. A number of studies have shown, however, that gene expression exhibits a superior predictive power for modeling drug response in cancer cells [7, 25, 28]. For our analysis, we used gene expression data generated by the RNASeq technology. The expression data was collected from two public data repositories: NCI60 and Cancer Cell Line Encyclopedia (CCLE). For the cells in the NCI60 dataset in Table 1, the RNASeq was retrieved from the NCI60 repository. For the GDSC1, GDSC2, and CTRP datasets, the RNASeq was retrieved from the NCI60 and CCLE repositories by matching the cell line names across the databases. Drug response samples for which the cell gene expressions were not available in CCLE or NCI60 were excluded from the datasets.
The retrieved RNASeq data, provided in FPKM (fragments per kilobase million) values, was transformed into TPM (transcripts per kilobase million) values. Instead of using expressions of more than 20,000 available genes for modeling drug response, we used the expressions of 976 landmark genes as identified by the Library of Integrated NetworkBased Cellular Signatures (LINCS) project [29]. The LINCS gene set has been shown to comprehensively characterize transcriptomic changes under various chemical and genetic perturbations [29].
Representation of drugs
Classical ML algorithms such as GBDT ignore the arrangement of features in datasets while utilizing the feature values only. Since we want to compare the learning curves of classical ML and DL models, we used molecular descriptors as drug representations in which the ordering of features is not intended to carry meaningful information. The descriptors were generated by using the Mordred software package [30]. The full descriptor set comprises 1,826 features, including both 2D and 3D molecular structure descriptors. Since most of the 3D descriptors resulted in invalid (NaN) values for the majority of compounds, we retained only the 2D descriptors, providing a total of 1,613 drug features.
Machine learning for drug response prediction
This section first presents the formulation of drug response prediction as a supervised learning problem and then describes the learning algorithms and model training procedures.
Drug response prediction as a supervised learning task
Consider the following definitions for a supervised learning problem of drug response prediction. Given a training set T with M samples, \(T=\{\boldsymbol{x}_{i},y_i\}_{i=1}^M\), \(\boldsymbol{x}_i\) is a feature vector for celldrug pair i, and \(y_i\) is the corresponding doseindependent drug response. The prediction task is to learn a mapping function \(f:{\mathbb {R}}^{C+D}\rightarrow {\mathbb {R}}\), where C and D are the number of gene expressions and drug descriptors, respectively. Drug response datasets comprise a unique set of celldrug pairs, \(X=[X_{c}\; X_{d}]=\{[{\boldsymbol{x}}_{c,i}\; {\boldsymbol{x}}_{d,i}]\}_{i=1}^M\), where \({\boldsymbol{x}}_{c,i}\) and \({\boldsymbol{x}}_{d,i}\) are, respectively, the cell and drug feature vectors of the \(i^\text {th}\) pair. The algorithm learns the mapping function (i.e., prediction model) by minimizing a regression loss function, which in our analysis is the mean squared error of prediction outcomes.
Note that each cancer cell line was screened against multiple drugs and, vice versa, each drug was tested on multiple cell lines. Thus, although each celldrug combination is unique in the training set T, the feature vectors of individual cells, \({\boldsymbol{x}}_c\), and drugs, \({\boldsymbol{x}}_d\), appear multiple times in T. The analysis of how this redundancy in feature space affects prediction models and learning curves is beyond the scope of this paper and provides a topic for further investigation.
Machine learning models and training procedures
NNs can be designed to enhance learning from a particular feature type of cell or drug [11, 31]. In such models, the prediction performance depends on the availability, quality, and diversity of that specific feature type in a training set. In contrast, our primary objective is to gain insight into the overall prediction improvement with an increasing number of training samples. Thus, we refrained from using architectures that focus on learning from specific feature types of cells or drugs.
Two NN architectures and one classical ML algorithm were used for the analysis. The two NNs differ primarily in the way the features were fed to the network input, as shown in Fig. 3. In the singleinput NN (sNN), gene expressions and drug descriptors were concatenated to form an input feature vector. In the multiinput NN (mNN), expression features and descriptors were first encoded by separate subnetworks before being concatenated and subsequently propagated to the output. Both models contain approximately the same number of trainable parameters (sNN: 4.254 million parameters, mNN: 4.250 million parameters). All fully connected layers, excluding the output layer, were followed by batch normalization [32], ReLu activation, and a dropout layer [33]. A batch size of 32 and the Adam optimizer [34] were used for model training.
For classical ML, we used the GBDT algorithm, implemented in the LightGBM library [35]. We used two versions of this algorithm: (1) dGBDT, a GBDT with default hyperparameters (HPs), and (2) hGBDT, where we optimized the HPs via a randomized search [36]. GBDT is an ensemble of decision trees in which a series of tree learners is optimized via a gradient descent optimization. Every subsequent tree improves inaccurate predictions of previous learners, boosting the predictive performance of the final model.
The GBDT and NNs were trained with, respectively, the LightGBM [35] and Keras [37] software libraries. The best set of HPs for each combination of a model and dataset was determined based on a randomized HP search by training on 80% of the available data and validating on the remaining 20%. Note that default HPs, as provided by the LightGBM library, were used for training dGBDT.
To mitigate overfitting, we used the early stopping functionality, available in both LightGBM and Keras. With early stopping, the model training procedure is terminated if the prediction performance on a validation set has not improved for a specified number of training iterations. We set the early stopping parameter to 25 epochs for the NNs and 50 boosting rounds for GBDT models. To guarantee convergence of NNs, we used a sufficiently large number of 500 epochs to ensure that the early stopping function was triggered.
Workflow for generating and fitting learning curves
This section lays out the methodology for generating learning curve data and fitting a power law model. A schematic of the workflow is illustrated in Fig. 4.
Data partitioning
A dataset D is randomly shuffled and split into three disjoint sets: training T, validation V, and test E. Shuffling D before generating the splits increases the likelihood that the three partitions exhibit a similar distribution of drug responses. A total of \(N=20\) combinations, \(\{T, V, E\}_{n=1}^{N}\), were generated by shuffling D with different random seeds. Each set \(\{T, V, E\}\) maintains the same size proportion of (0.8, 0.1, 0.1) as a fraction of the total number of samples D.
For each of the N splits, we form a set of K training subsets, \(\{T_k\}_{k=1}^{K}\), of increasing sizes \(m_{k1}<m_k\), where \(m_k\) is the sample size of \(T_{k}\). The samples are sequentially pulled from T to form each subset \(T_k\), as illustrated in Fig. 4. This process ensures that each subset contains all the training samples from the immediately preceding subset such that \(T_{k1}\subseteq T_k\). The inclusion of the entire subset \(T_{k1}\) in \(T_{k}\) mimics the temporal evolution of a research study where new samples are added to an already available dataset. Note that shuffling D at the first step of this data partitioning workflow is essential to enable the sequential sampling from T while eliminating potential biases associated with the original ordering of samples in D.
Generating learning curve data
Once the N data splits and the K training subsets for each split are generated, a total of \(N \times K\) prediction models are trained. In order to support early stopping, the training is terminated if the generalization performance on the validation set V has not improved over a predefined number of training iterations. Upon training completion, the model predicts the response for each celldrug pair in the test set E. Note that within a given data split n, all K models use the same validation and test sets (V and E) irrespective of \(m_k\).
The predictions on test set E are aggregated into a single score, \(s_k^n\), by computing the mean absolute error of predictions, where \({s_k^n}\) is the generalization error for a subset of size \(m_k\) of split n. As a result, the raw learning curve data, \({LC_{raw}=\{s_k^n\}, k=1,\ldots ,K; n=1,\ldots ,N}\), is produced. The total number of models trained for each combination of a dataset and a model depends on the values of N and K. For most datasetmodel pairs, we trained \(K=10\) models per split, resulting in a total of 200 models (200 raw LC error scores).
There is no sequential dependency in terms of model training with different subsets and data splits for generating the raw LC data. Thus, the proposed workflow enables the trainings to be distributed across multiple processors. The workflow was parallelized on appropriate platforms depending on the ML software framework. The NNs were trained on the Summit HPC, while the GBDT models were trained on a CPU cluster.
Curve fitting
The raw learning curve data, \({LC_{raw}}\), contains \(N \times K\) error scores that were obtained for N data splits and K training sizes. The power law expression in Eq. (1) primarily accounts for error scores that span the power law (\(am^{b}\) term) and plateau (constant c) regions, while mostly excludes the region of smalldata. Thus, an inadequate choice of data points can lead to a bad fit.
As a first step in producing a reliable fit, we visually identified error scores in the smalldata region and excluded them from \(LC_{raw}\), as shown in Fig. 5a, where \(m_{kmin}\) is the smallest sample size that was considered for further analysis. The remaining scores were considered next for fitting the power law expression, \(s(m) = am^{b}+c\), with \(\beta =(a,b,c)\) being the parameters to be estimated. We used the weighted version of the leastsquares fit to prioritize the contribution of larger subsets (i.e., larger weights, \(\alpha _k\), were assigned to larger values of \(m_k\)) which essentially reduces the effect of small training subsets that are in close proximity with random guessing. Specifically, each score was assigned a weight that is normalized by the sample size, \(\alpha _k=m_k/m_K\) [18, 38]. The parameter estimates \({\hat{\beta }}\) were obtained for each split n by fitting the remaining scores with a weighted nonlinear leastsquares minimization method as described in [38], producing N combinations of parameter estimates, \(\{{\hat{\beta }}_n\}_{n=1}^N\), where \({\hat{\beta }}_n\) are the estimated power law parameters for split n. For each n, we used \(\{m_k\}_{k=k_{min}}^K\) to compute a set of power law values, \(s_n(m_k) = s(m_k;{\hat{\beta }}_n)\), based on Eq. 1. The N sets of all the computed power law values, \(y(m_k)=\{s_n(m_k)\}_{n=1}^N\), as well as the corresponding median (i.e., the \(0.5^\text {th}\) quantile) values computed across the N sets, \(q_{0.5}(m_k) \triangleq \tilde{y}(m_k)=\{\tilde{s}_n(m_k)\}_{k=k_{min}}^K\), are shown as gray and black dots, respectively, in Fig. 5b. In a similar manner of computing \(\tilde{y}(m_k)\), we computed the \(0.1^{\text {th}}\) quantile, \(q_{0.1}(m_k)\), and \(0.9^{\text {th}}\) quantile, \(q_{0.9}(m_k)\). Finally, we fit the power law to \(\tilde{y}(m_k)\), \(q_{\text {0.1}}(m_k)\), and \(q_{\text {0.9}}(m_k)\), which represent, respectively, the error score estimates for a given datasetmodel combination (black curve), and the corresponding lower and upper variability estimates of the fit (shaded region bounded by the blue and green curves).
We used the \(MAE_{fit}\) and \(R_{fit}^2\) as the goodnessoffit measures. The \(MAE_{fit}\) is the mean of absolute value of residuals between the values representing the error estimates (black dots) and the values representing the fit (black curve), where smaller values of \(MAE_{fit}\) indicate a closer fit. These measures have been shown as appropriate metrics for evaluating the quality of the power law fit [17, 18].
Results
A single experiment refers to the workflow of generating \(LC_{raw}\) and fitting the power law expression in Eq. (1) to \(\tilde{y}\), \(q_{0.1}\), and \(q_{0.9}\), for a pair of a dataset and an ML model. Pairing all the possible combinations results in a total of sixteen experiments. The prediction error scores of models trained with the full training set, \(\tilde{y}_K=\tilde{y}(m=m_K)\), are listed in Table 2. The dGBDT serves as the baseline model in this comparison and is used to calculate the reduction in error score of other models with
where \(\alpha\) is the evaluated ML model. All models yield lower \(\tilde{y}_K\) as compared with dGBDT, exhibiting an improvement of 8% up to 42% across the datasets. While these results were expected, the observed improvements render dGBDT as an inadequate baseline for drug response prediction.
Because of the large difference in performance between dGBDT and the other models, we analyze the learning curves of dGBDT separately in Fig. 5. Since LightGBM is highly parallelizable and allows faster model convergence as compared with NNs, we train 1,000 dGBDT models (\(N=20\), \(K=50\)) with each of the four datasets, as shown in Fig. 5a. Note that the three regions of the learning curve are apparent in each plot.
The smalldata region especially stands out because of the large spread of scores on the vertical axis. The power law region follows next and can be identified as the linear region on the \(log{}log\) plot. The curve then starts to converge and progresses to the range of irreducible error. We visually identified the data points within the smalldata region and excluded them from \(LC_{raw}\). The remaining points were used to obtain the power law fits for \(\tilde{y}\), \(q_{0.1}\), and \(q_{0.9}\), resulting in a substantially close fit with a maximum \(MAE_{fit}\) of \(2.62 \times 10^{5}\) and a minimum \(R^2_{fit}\) of 0.98 across the four datasets, as shown in Fig. 5b. In each case, the curve exhibits a trajectory of convergence at higher \(m_k\) values, suggesting that additional data is not expected to substantially reduce the prediction error. These plots are the first observation demonstrating that the power law appropriately characterizes the data scaling behavior of drug response prediction models.
To assess the utility of learning curves as a global metric for evaluating prediction models, we collected \(LC_{raw}\) for the hGBDT, sNN, and mNN models for each dataset. To obtain error scores for an appropriate powerlaw fit, we qualitatively selected, based on empirical observations of plots in Fig. 5b, a range of \(m_k\) that excludes the smalldata region for each dataset. The remaining data points were used to obtain \(\tilde{y}\), \(q_{0.1}\) and \(q_{0.9}\), and the corresponding power law fits, as shown in Fig. 6. The \(\tilde{y}\) values and the corresponding power law fits are shown in Fig. 6, including the fits for \(q_{0.1}\) and \(q_{0.9}\) which are represented by the shaded regions. The selected \(m_k\) range is summarized in Table 3 for each dataset, including the goodnessoffit measure \(MAE_{fit}\) for the power law fits.
As Fig. 6 indicates, mNN outperforms sNN across the entire range of the explored training sizes on every dataset, albeit with the similar number of trainable parameters in these NNs. This superiority of mNN can be attributed to the separate encoding of gene expressions and drug descriptors within the individual input subnetworks, enhancing the overall model learning. Moreover, mNN exhibits the lowest prediction error at the full sample size for all datasets. On both GDSC datasets, however, no single model dominates across the entire \(m_k\) range: hGBDT outperforms both NNs at a lower range but performs worse than NNs as the training size increases. Another important observation is the different trajectories of the curves among the datasets. On CTRP, for example, the slope of mNN is considerably steeper than that of sNN. Thus, the mNN is expected to exhibit a higher rate of improvement on prediction score if the training size further increases. On NCI60, however, while the NNs exhibit a similar curve for the majority of the observed range, mNN shows a sign of convergence and begins to transition from the power law region to plateau.
The power law fit can be used to address questions such as the following that allow forecasting the prediction performance beyond the available training set size. (1) What is the expected prediction error if the training size is doubled, namely, \(\tilde{y}(m=2T)\)? (2) What is the training set size required to reduce the prediction error by 10%, namely, \(m(\tilde{y}=0.9\tilde{y}_K)\)? These questions are addressed by plugging the appropriate values for \(m_k\) or \(\tilde{y}_k\) in Eq. (1) while using the power law parameter estimates for each datasetmodel pair. The rightmost two columns in Table 2 list the computed values addressing these two questions. The observations and results in this section directly demonstrate the benefit of using learning curves to evaluate prediction models, which provide a broader perspective on the overall scaling trajectory of these models for drug response prediction, and the utility of the powerlaw fits for addressing important questions in prospective research.
Discussion
The analysis across sixteen experiments (with four datasets and four models) demonstrates that no single model dominates in terms of prediction performance across all datasets and training sizes. This result supports the assumption that the actual shape of learning curves depends on both the dataset and the model. For example, hGBDT exhibits lower error scores for a lower range of training sizes on two datasets, as shown in Fig. 6a, b. Alternatively, mNN outperforms all the investigated models (dGBDT, hGBDT, sNN) across all datasets for sufficiently large training sizes. Moreover, both NNs maintain data scaling properties that are characterized by the power law region, demonstrating a promising trajectory of further improvement. These observations indicate that the power law fits can be used to project the expected error score beyond the available training size or, alternatively, calculate the sample size required to achieve a specific performance. These uses of learning curves can aid in collaboration between experimental biologists and computational scientists to shape a global vision of how predictive models can be further improved. This valuable perspective can guide the process of new data generation, either through lab experiments or synthetically, via simulations or resampling methods.
One should be cautious, however, in generalizing the data scaling characteristics when building prediction models with subsets of the investigated datasets or modified architectures of sNN or mNN. In this study, we explored a general case where both cell and drug features predict the drug response. Alternatively, models that focus on a specific cancer type are usually trained by using an appropriate subset of cell lines of a much smaller sample size. Moreover, to mitigate overfitting with the reduced sample size, researchers may choose to limit the feature space by using the drug features only. These dataset and model changes may produce a different layout of learning curves and, therefore, different conclusions and downstream actions. We therefore recommend generating learning curves for every datasetmodel pair that is being analyzed.
In Fig. 6b, d, we observe a moderate improvement at the full training set size with mNN as compared with hGBDT. Similarly, certain studies proposing novel DL methods for drug response prediction demonstrate moderate improvement as compared with classical ML [10, 31, 39]. This is in contrast to vision and text applications where DL methods represent the state of the art [40, 41]. The complexity in design and training of models such as mNN and hGBDT is higher than that of their respective counterparts, sNN and dGBDT. While using simpler models as a demonstration vehicle might be tempting, such models typically result in a poor baseline for objectively evaluating the prediction performance of proposed models. Similarly, generating learning curves requires significant computational resources for performing a thorough analysis across multiple data splits and training sizes, as demonstrated in this study. While often time and resourceconsuming, a rigorous comparison of novel models with strong baselines is necessary for producing a significant impact and visibility within the ML community.
Contributed to the diversity and complexity of cancer pathologies, no single combination of a dataset, feature type, and prediction target serves as a universal benchmark for modeling drug response. Various learning methodologies and data preprocessing techniques have been explored to enhance the predictive capabilities of models, including multimodal learning [11, 26], feature encoding schemes [27, 31], and transfer learning [13, 42]. The performance of these models is assessed by comparing singlevalue performance measures such as prediction error or accuracy against baseline models obtained at the full sample size. The lack of unified benchmarks for the drug response problem, alongside a moderate improvement in predictions as compared with baseline models, may cast doubt on the true potential of proposed models. In such cases, learning curves can better highlight the potential impact of models by depicting the improvement of predictions with increasing sample size. For example, assume that we analyze the performance of mNN against sNN using CTRP and NCI60 datasets. On NCI60, the trajectory of the curves suggests that both models are expected to converge with more training data (see Fig. 6d), whereas on CTRP, the curves portray a more favorable trajectory for mNN to potentially further improve the performance with more data (see Fig. 6c). Alternatively, when using the prediction performance at the full training set size, \(\tilde{y}_K\), both CTRP and NCI60 exhibit a comparable superiority of mNN as compared with the sNN baseline, failing to demonstrate the additional dimension of comparison provided by the learning curves. To the best of our knowledge, this is the first work that proposes the use of learning curves for systematic evaluation and comparison of machine learning algorithms in drug response problem.
Similarly to learning curve studies from other scientific domains, we demonstrate that the power law in Eq. (1) closely models the data scaling for the application of drug response prediction. While other works focus primarily on a single family of ML models, we have investigated both classical ML and DL models, with a primary observation that no single model is superior to other models for all datasets and training sizes. Moreover, the extent of the three learning regions (described in Fig. 1) significantly differs among the different applications. For example, Mukherjee et al. [16] accurately fit the power law of eight cancerrelated classification tasks with DNA microarray datasets ranging between 53 and 280 samples. The prediction of drug response in cancer cell lines is presumably a more challenging task, since our models require thousands of training samples to reveal the learning regions.
While cell lines remain a primary environment for mimicking cancer, alternative biological models are being investigated as closer surrogates of human cancer. These alternatives include patientderived xenografts (PDXs), which are cancer implants in animals, and patientderived organoids (PDOs), which are 3D cultures of cancer cells from patients. ML analysis with these emerging cancer environments is essential for future development of cancer treatment. At this point, however, drug response data for these biological models is scarce compared with the relatively abundant cell line screening data. The scarcity of PDXs and PDOs data imposes a challenging search for suitable methods across the entire space of learning algorithms. Therefore, learning curves can serve as a useful codesign tool for comparing predictive performance of learning algorithms and facilitate the design of future experiments in a prospective research setting.
Conclusions
We demonstrate that learning curves of drug response prediction models using both classical ML and DL methods follow a power law expression. The specific trajectory of the curves depends on the dataset and the learning algorithm and therefore should be obtained empirically. While hGDBT exhibits superior performance at a lower range of training sizes, the mNN outperforms all the investigated models as the training size increases. These observations demonstrate the benefits of learning curves in modeling the data scaling properties of prediction models, while the proposed methodology allows to quantify the expected prediction performance across the entire range of training sizes. The power law fit can also be utilized to forecast the behavior of learning curves beyond the available training size. The fitted power law curve provides a forwardlooking metric for analyzing prediction performance and can serve as a codesign tool to guide experimental biologists and computational scientists in the design of future experiments in prospective research studies.
Availability of data and materials
The full code is available at https://github.com/adpartin/drlearningcurves. The data that was used to train the machine learning models is available at https://ftp.mcs.anl.gov/pub/pilot1/publications/drlearningcurves/data/.
Abbreviations
 GBDT:

Gradient boosting decision tree
 dGBDT:

Gradient boosting decision tree with default hyperparameters
 hGBDT:

Gradient boosting decision tree with hyperparameters tuning
 sNN:

Singleinput neural network
 mNN:

Multiinput neural network
 ML:

Machine learning
 DL:

Deep learning
 MAE:

Mean absolute error
 \(LC_{raw}\) :

Raw learning curve data.
References
Sharma SV, Haber DA, Settleman J. Cell linebased platforms to evaluate the therapeutic efficacy of candidate anticancer agents. Nat Rev Cancer. 2010;10:241–53. https://doi.org/10.1038/nrc2820.
Gillet JP, Varma S, Gottesman MM. The clinical relevance of cancer cell lines. JNCI J Natl Cancer Inst. 2013;105(7):452–8. https://doi.org/10.1093/jnci/djt007.
BenDavid U, et al. Genetic and transcriptional evolution alters cancer cell line drug response. Nature. 2018;560:325–30. https://doi.org/10.1038/s4158601804093.
Yang W, et al. Genomics of drug sensitivity in cancer (GDSC): a resource for therapeutic biomarker discovery in cancer cells. Nucleic Acids Res. 2013;41:955–61. https://doi.org/10.1093/nar/gks1111.
SeashoreLudlow B, et al. Harnessing connectivity in a largescale smallmolecule sensitivity dataset. Cancer Discov. 2015;5(11):1210–23. https://doi.org/10.1158/21598290.CD150235.
Grever MR, Schepartz SA, Chabner BA. The national cancer institute: cancer drug discovery and development program. Semin Oncol. 1992;19:622–38.
Costello JC, et al. A community effort to assess and improve drug sensitivity prediction algorithms. Nat Biotechnol. 2014;32:1202–12. https://doi.org/10.1038/nbt.2877.
Niz CD, Rahman R, Zhao X, Pal R. Algorithms for drug sensitivity prediction. Algorithms. 2016;9(77):1202–12. https://doi.org/10.3390/a9040077.
Adam G, Rampášek L, Safikhani Z, Smirnov P, HaibeKains B, Goldenberg A. Machine learning approaches to drug response prediction: challenges and recent progress. NPJ Precis Onc. 2020. https://doi.org/10.1038/s4169802001221.
Rampášek L, Hidru D, Smirnov P, HaibeKains B, Goldenberg A. Dr.VAE: improving drug response prediction via modeling of drug perturbation effects. Bioinformatics. 2019;35(19):3743–51. https://doi.org/10.1093/bioinformatics/btz158.
Manica M, Oskooei A, Born J, Subramanian V, SáezRodríguez J, Rodríguez MM. Towards explainable anticancer compound sensitivity prediction via multimodal attentionbased convolutional encoders. Mol Pharm. 2019;16(12):4797–806. https://doi.org/10.1021/acs.molpharmaceut.9b00520.
Bazgir O, Zhang R, Dhruba SR, Rahman R, Ghosh S, Pal R. Representation of features as images with neighborhood dependencies for compatibility with convolutional neural networks. Nat Commun. 2020;11:4391. https://doi.org/10.1038/s4146702018197y.
Zhu Y, et al. Ensemble transfer learning for the prediction of anticancer drug response. Sci Rep. 2020;10:18040. https://doi.org/10.1038/s41598020749210.
Vougas K, et al. Machine learning and data mining frameworks for predicting drug response in cancer: An overview and a novel in silico screening process based on association rule mining. Pharmacol Therapeut. 2019. https://doi.org/10.1016/j.pharmthera.2019.107395.
Cortes C, Jackel LD, Solla SA, Vapnik V, Denker JS. Learning curves: Asymptotic values and rate of convergence. Adv Neural Inf Process Syst. 1994;6:327–34.
Mukherjee S, Tamayo P, Rogers S, Rifkin R, Engle A, Campbell C, Golub TR, Mesirov JP. Estimating dataset size requirements for classifying DNA microarray data. J Comput Biol. 2003;10(2):119–42. https://doi.org/10.1089/106652703321825928.
Last M. Predicting and optimizing classifier utility with the power law. In: Seventh IEEE International Conference on Data Mining Workshops (ICDMW 2007), 2007; pp. 219–224. https://doi.org/10.1109/ICDMW.2007.31.
Figueroa RL, ZengTreitler Q, Kandula S, Ngo LH. Predicting sample size required for classification performance. BMC Med Inform Decis Mak. 2012. https://doi.org/10.1186/14726947128.
Sun C, Shrivastava A, Singh S, Gupta A. Revisiting unreasonable effectiveness of data in deep learning era. In: 2017 IEEE international conference on computer vision (ICCV), 2017; pp. 2961–2969. https://doi.org/10.1109/ICCV.2017.97.
Hestness J et al. Deep learning scaling is predictable, empirically. arXiv eprints; 2017. arXiv:712.00409.
Rosenfeld JS, Rosenfeld A, Belinkov Y, Shavit N. A constructive prediction of the generalization error across scales. In: International conference on learning representations 2020.
Amari SI, Fujita N, Shinomoto S. Four types of learning curves. Neural Comput. 1992;4(4):605–18. https://doi.org/10.1162/neco.1992.4.4.605.
Haussler D, Kearns M, Seung HS, Tishby N. Rigorous learning curve bounds from statistical mechanics. Mach Learn. 1996;25:195–236. https://doi.org/10.1023/A:1026499208981.
Anzanello MJ, Fogliatto FS. Learning curve models and applications: Literature review and research directions. Int J Ind Ergon. 2011;41(5):573–83. https://doi.org/10.1016/j.ergon.2011.05.001.
Xia F, et al. Predicting tumor cell line response to drug pairs with deep learning. BMC Bioinform. 2018;19:486. https://doi.org/10.1186/s1285901825093.
SharifiNoghabi H, Zolotareva O, Collins CC, Ester M. MOLI: multiomics late integration with deep neural networks for drug response prediction. Bioinformatics. 2019;35(14):501–9. https://doi.org/10.1093/bioinformatics/btz318.
Zhu Y, et al. Enhanced coexpression extrapolation (coxen) gene selection method for building anticancer drug response prediction models. Genes. 2020;11:1070. https://doi.org/10.3390/genes11091070.
Jang IS, Neto EC, Guinney J, Friend SH, Margolin A. Systematic assessment of analytical methods for drug sensitivity prediction from cancer cell line data. In: Pacific symposium on biocomputing, 2014; pp. 63–74.
Subramanian A, et al. A next generation connectivity map: L1000 platform and the first 1,000,000 profiles. Cell. 2017;171(6):1437–52. https://doi.org/10.1016/j.cell.2017.10.049.
Moriwaki H, Tian YS, Kawashita N, Takagi T. Mordred: a molecular descriptor calculator. J Cheminform. 2018. https://doi.org/10.1186/s133210180258y.
CortésCiriano I, Bender A. Kekulescope: prediction of cancer cell line sensitivity and compound potency using convolutional neural networks trained on compound images. J Cheminform. 2019. https://doi.org/10.1186/s1332101903645.
Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Proceedings of the 32nd international conference on machine learning; 2015, vol. 37, pp. 448–456.
Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res. 2014;15(56):1929–58.
Kingma DP, Ba JL. Adam: a method for stochastic optimization. In: International conference on learning representations (ICLR) 2015.
Ke G, Meng Q, Finley T, Wang T, Chen W, Ma W, Ye Q, Liu TY. Lightgbm: a highly efficient gradient boosting decision tree. In: Proceedings of the 31st international conference on neural information processing systems, 2017; pp. 3149–3157.
Bergstra J, Bengio Y. Random search for hyperparameter optimization. J Mach Learn Res. 2012;13:281–305.
Chollet F et al. Keras. https://keras.io 2015.
Johnson M, Anderson P, Dras M, Steedman M. Predicting accuracy on large datasets from smaller pilot data. In: Proceedings of the 56th annual meeting of the association for computational linguistics (volume 2: short papers); 2018, pp. 450–455. https://doi.org/10.18653/v1/P182072.
Chiu YC, Chen HIH, Zhang T, Zhang S, Gorthi A, Wang LJ, Huang Y, Chen Y. Predicting drug response of tumors from integrated genomic profiles by deep neural networks. BMC Med Genomics. 2019;12:18. https://doi.org/10.1186/s1292001804609.
Tan M, Pang R, Le QV. Efficientdet: scalable and efficient object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (CVPR), 2020.
Howard J, Ruder S. Universal language model finetuning for text classification. In: Proceedings of the 56th annual meeting of the association for computational linguistics (volume 1: long papers), 2018; pp. 328–339 . https://doi.org/10.18653/v1/P181031
Rahman Dhruba S, Rahman R, Matlock K, Ghosh S, Pal R. Application of transfer learning for cancer drug sensitivity prediction. BMC Bioinform 2018. doi: 10.1186/s128590182465y.
Acknowledgements
The authors thank the Argonne Leadership Computing Facility (ALCF) and Oak Ridge Leadership Computing Facility (OLCF) for providing computing resources for this research.
Funding
This work has been supported in part by the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) program established by the U.S. Department of Energy (DOE) and the National Cancer Institute (NCI) of the National Institutes of Health. This work was performed under the auspices of the U.S. Department of Energy by Argonne National Laboratory under Contract DEAC0206CH11357, Lawrence Livermore National Laboratory under Contract DEAC5207NA27344, Los Alamos National Laboratory under Contract DEAC5206NA25396, and Oak Ridge National Laboratory under Contract DEAC0500OR22725. This project has also been funded in whole or in part with federal funds from the NCI, National Institutes of Health (NIH), under Contract No. HHSN261200800001E. The content of this publication does not necessarily reflect the views or policies of the Department of Health and Human Services, nor does mention of trade names, commercial products, or organizations imply endorsement by the U.S. Government.
Author information
Authors and Affiliations
Contributions
AP designed the study; AP, MS, AC and FX collected and preprocessed the data; AP, HY and TB conducted the experiments; AP, YZ, TB and RLS wrote and edited the article; AP, SJ and YZ performed data analysis; AP, TB, YAE, YZ, HY, FX, SJ, AC, MS, MF, JHD and RLS interpreted results and derived conclusions; RLS and JHD conceived the idea. All authors have read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Partin, A., Brettin, T., Evrard, Y.A. et al. Learning curves for drug response prediction in cancer cell lines. BMC Bioinformatics 22, 252 (2021). https://doi.org/10.1186/s1285902104163y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1285902104163y