Managerial efficiency and efficiency differentials in adult education: a conditional and bias-corrected efficiency analysis

In efficiency studies, inputs and outputs are often noisy measures, especially when education data is used. This article complements the conditional efficiency model by correcting for bias within conditional draws, using the m out of n bootstrap procedure. With a unique panel dataset, we estimate managerial efficiency, which accounts for nondiscretionary variables, and explain efficiency differentials of adult education programs in Flanders. Our results suggest that the characteristics of learners in a program matter for managerial efficiency, and that teacher characteristics are strongly correlated with efficiency differentials, as more homogeneity in the teacher workforce appears to result in higher program efficiency.

education is particularly important given the increased outsourcing, continuous developments in technology, and a trend towards job-polarization.
This article evaluates the efficiency of adult education programs. Since the seminal work by Charnes et al. (1978), applications of efficiency models have emerged in various sectors, ranging from health care (e.g. Cordero et al. 2015), water (Cherchye et al. 2014), banking (Puri et al. 2017), public sector (De Witte and Geys 2013), farming (Asmild et al. 2016) or energy (e.g. Chen et al. 2017;Hampf et al. 2015). In addition, a particularly rich efficiency literature focuses on education (see De Witte and López-Torres 2014; Johnes 2015; Thanassoulis et al. 2016 for a review). Although efficiency has been estimated in primary (Blackburn et al. 2014), secondary (Portela et al. 2012) and higher education (Johnes 2006), adult education has received no attention in the efficiency literature. To the best of our knowledge, no article in the efficiency literature has estimated the efficiency of adult education programs. Outside the efficiency literature, a number of articles have estimated the causal effects of such programs (Blanden et al. 2012;Martin et al. 2013;Oosterbeek et al. 2010). These studies generally conclude that adult education increases labour market outcomes. Yet, evaluating the efficiency of adult education programs is particularly important to ensure that budget resources are allocated in a way that maximizes outcomes.
This article contributes to the literature in at least three ways. First, from a methodological perspective, we introduce an innovative approach to estimate efficiency in the presence of heterogeneity and sampling noise. In particular, we extend the commonly used conditional model (Daraio and Simar 2007) by correcting for bias within conditional draws, using the m out of n bootstrap procedure (Simar and Wilson 2011). Inputs and outputs are often noisy measures, especially when education data is used, due to uncertainty when statistical procedures are not adhered to. For instance, inputs and outputs may be noisy due to imputation errors from using plausible values (PVs) in PISA (Jerrim et al. 2017), 1 due to inaccurate use of survey tests (von Davier et al. 2009), or due to measurement errors and outlying observations. As a result, efficiency estimates can be biased, despite the use of conditional models that account for observed heterogeneity. 2 As a second contribution, we distinguish between managerial efficiency and efficiency differentials. The former does not allow us to assume separability as nondiscretionary variables need to be accounted for in a conditional model. The latter can be regressed on a set of discretionary variables using a two-stage approach, as the attainable set is not affected by those variables (separability) (as in Cherchye et al. 2014). As a third contribution, we extend the efficiency of education literature to the yet unexplored field of adult education. We do so by exploiting a unique panel dataset from the Belgian region of Flanders. This dataset comprises detailed information on program costs and learners' outcomes, in addition to a rich set of both learners' and teachers' characteristics. Using this dataset, managerial efficiency estimates can be obtained while accounting for the observed characteristics of learners in a program, and efficiency differentials can be explained using discretionary teacher and program characteristics.
Our results indicate that a larger share of low educated learners has an unfavorable influence on the evaluation of adult education programs. By contrast, a larger share of men appears to have a favorable influence on the evaluation of adult education programs, while the age of learners does not appear decisive. Overall, we observe that characteristics of learners matter significantly for managerial efficiency, motivating the need for a conditional model. In 1 The use of plausible values in PISA data results in imputation errors when constructing observations for students that did not complete specific subjects (for a technical description, see OECD (2009)). 2 The suggested approach can also be applied in other fields where sampling noise is present. To facilitate further applications, the R code is available upon request. addition, we find that a considerable amount (over 30%) of the bias in our unconditional estimates persists when a conditional model is used to obtain managerial efficiency estimates, despite the subsampling approach implemented to obtain the latter. This highlights the importance of a bias-correction in conditional models. With respect to the efficiency differentials, we find that teacher and program characteristics are able to explain efficiency differentials between adult education programs. In particular, female teachers and a more homogenous teacher composition, in terms of age and number of hours allocated to teachers, are likely to correspond with a higher program efficiency.
The remainder of the article is structured as follows. Section 2 proposes the conditional and bias-corrected efficiency model. Section 3 provides a brief introduction to the adult education system in Flanders and presents the data. Section 4 presents the results and Sect. 5 concludes.

Efficiency in adult education
Data Envelopment Analysis (DEA) models have received considerable attention in the efficiency literature, considering their ability to aggregate information on different inputs and outputs. The DEA model emanates from work by Farrell (1957), later implemented by Charnes, et al. (1978Charnes, et al. ( , 1981. Our study further extends this non-parametric methodology and applies it on the yet relatively unexplored field of adult education. In particular, we estimate the efficiency of adult education program o using the following baseline output-oriented model: max v,u θ o (y) u 1 y 1o + · · · + u q y qo (1) s.t. υ 1 x 1o + · · · + υ p x po 1 ( 1 a ) u 1 y 1 j + · · · + u q y q j ≤ υx 1 j + · · · + υx pj ( j 1, . . . , n) Solving the above linear program essentially constructs a production frontier composed of efficient education programs, and benchmarks all programs relative to this frontier. The DEA model in (1) yields an efficiency score θ o for program o, with q different outputs y, weighted by u, after normalizing the p different inputs x, weighted by v, in (1a). The weights u and v are determined endogenously to maximize this score. Condition (1b) requires the efficiency score of each education program o to attain at most 1, while (1c) implies that efficiency is increasing in the outputs used to construct it-i.e. weights are non-negative. The final condition (1d) imposes an integer formulation, i.e. the Free Disposal Hull (FDH) model of Deprins et al. (1984), rather than imposing a convexity assumption on the production frontier. The FDH model is a more general model than DEA, and performs better in the absence of convexity in the underlying data. An additional benefit of FDH estimates is that the omission of convexity allows the use of ratio measures in our output-oriented model (Olesen et al. 2015). Nevertheless, condition (1d) can be relaxed such that the DEA model is formulated. 3 The baseline FDH model, as presented above, can be improved along three lines. First, environmental variables which are not at the discretion of managers need to be included in the analysis to obtain an estimate of managerial efficiency. In the application at hand, the observed differences in learners' characteristics between programs call for a method that accounts for the environment when assessing performance, as the inflow of learners cannot be directly manipulated. Second, in the presence of noisy data, emanating from inaccurate use of survey tests (e.g. Jerrim et al. 2017;von Davier et al. 2009), invalid assessments, measurement errors or outliers in the inputs or outputs, conditional efficiency estimates should be corrected for bias. Parametric evidence suggests that ignoring similar uncertainty in the data results in biased estimates (e.g. Jerrim et al. 2017;von Davier et al. 2009). Third, variables at the discretion of managers can be used to explain differentials in conditional and bias-corrected efficiency differentials and draw policy conclusions. The next subsections provide more details on these extensions.

Conditional efficiency
One possible approach to account for nondiscretionary variables that are beyond the influence of the observation's managers (the 'environment') would be to obtain efficiency estimates using our baseline model (1) and regress these estimates on nondiscretionary variables (Z). However, this approach would impose a 'separability' condition, claiming that variables in Z are not affecting the attainable set, or production frontier (Simar and Wilson 2007). Clearly, this assumption is unlikely to hold when different types of learners sort into programs (see Sect. 3). Therefore, we evaluate adult education programs relative to a sample of programs with similar characteristics z. For each program o, we draw B subsamples of size m where the likelihood of being drawn depends on z and the kernel function estimated around z. 4 Each subsample is then used to compute a 'conditional' efficiency score 5 : For example, education programs with mostly low educated and male learners are more likely to be drawn into the reference set of program o when this program has a similar inflow of learners. 6 This contrasts to an unconditional model where the likelihood of being included in the reference set is independent of Z, and hence equal for all units. As a result, when applying the conditional model, the assumption that variables in Z are independent of the attainable set (i.e. separability) need not be imposed as programs are 'liked with likes'. In our application, we only include continuous (Epanechnikov kernel function) and ordered discrete variables (Li and Racine 2007;De Witte and Kortelainen 2013) and estimate bandwidths using least squares cross-validation (Li and Racine 2008). The issue of environmental variables in the estimation of efficiency has received considerable attention in the literature (see for example, Dyson et al. 2001;Estelle et al. 2010;Ruggiero 1996Ruggiero , 1998 and has been identified as a major concern in many efficiency studies in the field of, amongst others, education (De Witte and López-Torres 2017) and local governments (Narbón-Perpiñá and De Witte 2018a, b).

Conditional and bias-corrected efficiency
We correct for possible remaining sampling noise in the conditional sample. A naive bootstrap is invalid for FDH (Model 1) estimators as the sample maximum is drawn too frequently (Simar and Wilson 2011). Therefore, we follow Simar and Wilson (2011) by implementing the m out of n bootstrap procedure. As we are drawing B subsamples (of size s) from within the conditional draws (of size m), without replacement, we essentially implement an s out of m approach. 7 Denote the bias-corrected efficiency score as follows: Correcting for bias can prove useful in applications with noisy data in general and education data in particular. In education, inputs and outputs are often noisy measures (e.g. Jerrim et al. 2017). Bias-correction is especially important when using education data, as noise resulting from measurement error is widespread (see for example De Witte and Schiltz (2018) who illustrate issues when using survey data). Moreover, often the design of data collection methods leads to substantial bias in education data. For example, the use of plausible values in PISA data results in imputation errors when constructing observations for students that did not complete specific subjects (for a technical description, see OECD (2009)).
Despite mitigating the influence of outlying observations by the B draws of size m in the conditional efficiency estimator, the noisy input and output variables can result in biased estimates. The proposed method, combining conditional and bias-corrected efficiency, can be seen as an alternative to partial frontier approaches such as the robust order-m approach (Cazals et al. 2002;Daraio and Simar 2007) or order-α approach (Aragon et al. 2005). Although the conditional estimates account for the operational environment an observation 7 We set the subsample size s m k equal to 118, which corresponds to k 0.9. In unconditional bias-corrected estimates, the subsample size is equal to 644 ( n k ). is operating in, subsampling within conditional draws allows us to correct for remaining bias in the input and output variables of the conditional estimates. As we will illustrate in Sect. 4, bias-corrected estimates of the conditional model can still differ considerably from their conditional counterparts, despite an m out of n procedure being used to obtain the latter. In the application at hand, over 30% of the initially estimated bias remains after conditionally subsampling. Also, we compared the variation in bootstrapped conditional efficiency estimates to the bias estimates and conclude that bias-correction is required for almost all programs (see Sect. 4.2.). As an additional advantage, confidence intervals can be constructed from the distribution of bootstrap estimates, which facilitates the comparison and ranking of evaluated units (e.g. Johnes 2006).

Statistical inference for managerial efficiency and efficiency differentials
Once conditional and unconditional bias-corrected efficiency estimates are obtained for every program o, we can compute the ratio of both: Using this ratio, it is possible to gain insights into which nondiscretionary characteristics are favorable and unfavorable when measuring managerial efficiency (Daraio and Simar 2005). For example, when Q is increasing with the number of learners in program, it suggests that having more learners in a program is favorable. Nonparametric regression plots uncover such (nonlinear) relationships and corresponding significance levels can be computed (Li and Racine 2007).
In order to explain efficiency differentials, we regress the conditional and bias-corrected efficiency estimates on a set of explanatory variables. We do so in a nonparametric manner as a linear, a parabolic, or even a Fourier relationship (Schiltz and De Witte 2017) might be unable to capture the true nature of patterns in data. In fact, our flexible specification avoids imposing any functional form assumptions on the relationship between efficiency differentials and explanatory variables. As noted in Sect. 3, this set of variables consists of teacher and program characteristics that are discretionary, and hence subject to optimization by program providers. In contrast to the composition of learners choosing to enroll in a program, which cannot be directly manipulated, these variables are not included in the set of conditional variables. Again, nonparametric regression plots and significance levels can be obtained (Li and Racine 2007).

Summary of the methodology
Summarizing our approach, we proceed in four steps to measure and explain managerial efficiency. First, a subsample of size m is drawn based on a set of nondiscretionary characteristics (Z). Second, using this sample, the efficiency score is corrected for bias in inputs and outputs using bootstrapped (size s) estimates. This procedure is repeated many times to mitigate the influence of a single conditional draw. 8 Third, after averaging the results for each procedure, we compute the ratio of unconditional over conditional estimates to gain insights into which characteristics are favorable and unfavorable when measuring managerial efficiency (Daraio and Simar 2005), and compute significance levels (Li and Racine 2007). Finally, we explain

Data and background
Adult education programs in Flanders aim to provide lifelong learning opportunities to adults in six sectors (i.e. technology, management, environment, food, design, and metal and wood), and are designed to foster entrepreneurship within these sectors. Following the definition of Eurydice (Eurydice 2018) which distinguishes between adult basic education, secondary adult education, and adult higher vocational schools, we focus on programs offered in adult higher vocational schools in this study. All adult education programs are privately organized and publicly funded. Programs are subsidized in an output-oriented manner by the Flemish government. Centers of adult education receive funding per learner obtaining his or her certificate, depending on the length of the program and whether or not it is an officially recognized program. In contrast to the financing of compulsory education, funding is not earmarked such that resources can be allocated freely by the centers. No prerequisites are required for enrolment, although multiple programs can be organized in an adjacent manner, increasing in the difficulty level. Completion of previous programs will then be needed to proceed to the next program. The length of programs can vary considerably, but is at the discretion of centers providing adult education. Adult education centers (with different physical locations per center) are available throughout Flanders, as displayed in Fig. 2.
Our dataset covers the period 2006-2015 and includes detailed information on learners, teachers, and programs. We analyze efficiency at the program level to explain the differences in the ability of programs to transform inputs into outputs. All models estimated here are output-oriented, as our focus is on the maximization of learners' outcomes for a given set of inputs. Accordingly, our aim is not to estimate a cost function (e.g. Costrell et al. 2008;Duncombe et al. 1995;Imazeki 2008) but rather to provide a framework which allows policy makers to benchmark adult education programs. 9 The first panel of Table 1 reports summary statistics for the input and the outputs. As an input, we consider the cost per learner for each session. This input is measured in euros and computed by dividing the total cost of a program by the number of learners in a program and the number of sessions in a program. The total cost is obtained by multiplying the average hourly wage of teachers assigned to the program and the total number of hours taught. Hence, the cost per learner for each session is affected by the number of sessions, the number of learners, the number of hours taught, and the wage of the teachers hired for the program. Note that only the latter cost component cannot be altered by providers of entrepreneurial programs, as the teachers' wage is set by the central government. Summary statistics on the number of sessions per program and the number of learners per program are provided in Table 1. The selection of the outputs follows the insights obtained by De Leebeeck et al. (2018), who analyse adult education by a qualitative research design. They report the outcomes of 5 focus groups and interviews with 13 program directors of adult education programs, and several interviews of representatives of the national association of adult education (SYN-TRA). De Leebeeck et al. (2018) indicate that adult education programs aim to maximize the participation of learning both in classes and in evaluations in order to make all learners succeed. Therefore, subject to data limitations, we use three outputs: the reported presence during classes (reported by teachers and very accurate as it is used for funding purposes), the share of learners participating in exams and the average score on the final exam. It should be noted that these outputs are also in accordance with the criteria for evaluation put forward by the Flemish government. All outputs are measured as ratios, ranging from 0 to 1.
The second panel of Table 1 presents summary statistics for learners enrolled in an adult education program between 2006 and 2015. Our dataset includes information on the share of boys relative to girls ('gender balance'), the average age of learners, the share of low educated (at most high school) learners, the number of sessions per program, and the size of the program-measured as the number of learners. Apart from the two latter variables, learner characteristics cannot be directly manipulated by providers of adult education programs. Therefore, in order to benchmark programs in a more equitable manner, we compare programs that started with a similar composition of learners using a conditional efficiency model. It can be seen intuitively that non-discretionary variables (e.g. prior education of learners choosing a program) should be taken into account when benchmarking programs, while discretionary variables (e.g. the size of the groups created per program) should not. Along the same lines, the panel data structure requires programs to be evaluated relative to programs instructed in the same reference period to account for time-varying efficiency.
In addition to characteristics of learners following adult education programs, our unique dataset allows the inclusion of the gender balance of teachers, their average age, but also the variation in age between teachers and the way teaching hours are allocated between teachers. In the third panel, we present this set of variables related to programs' teacher workforce and the corresponding summary statistics. It is debatable whether teacher characteristics can be altered at will. Evaluating adult education programs in Flanders, two arguments can be raised to not include them as conditions in our model. First, the employment restrictions in adult education are less strict than in the compulsory education system. Teachers in adult education can be hired and fired more easily compared to those employed in the compulsory education system, making them more prone to managerial forces-and hence can be considered discretionary. Second, using our dataset, it is unfeasible to include all learners' and teachers' characteristics as conditions, as the curse of dimensionality would render all programs efficient. Therefore, once managerial efficiency estimates are obtained for all programs accounting for the variation in the inflow of learners, the teacher variables will be used in a second stage to explain efficiency differentials across programs. Our final sample includes 1200 observations, which corresponds to 120 programs observed over a 10 years' time span. Table 2 presents the efficiency scores of adult education programs under different model assumptions, introduced in the previous section. We start from our baseline FDH model without bias-correction (Model 1) and gradually extend it to measure managerial efficiency with bias-correction. On average, the FDH unconditional efficiency score amounts to 0.87, implying that programs could improve their performance by 13% if they were as efficient as those programs in their reference set. This rate of improvement amounts to over 50% for the worst performing programs. We enrich this unconditional FDH model to account for sampling noise in the data by adding bias-correction (Model 2), using the s out of m bootstrap, and identify an upward bias of around 1 percentage points.

Measuring efficiency
However, as performance is expressed in terms of learners' presence, participation in exams, and exam results, the inflow of learners in a program will be decisive for its evaluation. Failing to account for differences in inflow will complicate the comparison of programs. For example, two programs with the same budget (input) but with a different composition in learners might perform differently, which does not necessarily lead to the conclusion that one program is more efficient than the other. It might be that the program deemed efficient is attracting more motivated or mature learners who always attend classes and study well for their exams. In order to account for programs' environment, not at the discretion of providers of adult education, we employ conditional models to obtain managerial efficiency scores. Efficiency differences are then due to managerial decisions as similar programs are used to benchmark each other. Note Conditional models are obtained by including the following set of variables as conditions: the year of assessment, the gender balance, the average age, and the share of low educated learners. Bounds are constructed at the 95% level

Measuring managerial efficiency
Model 3 in Table 2 presents the results from the conditional model, accounting for the age, gender, and education background of learners in a program. In addition, we include the year of assessment as an ordered conditional variable. On average, conditional efficiency estimates are higher and less dispersed. This follows directly from programs being benchmarked to more similar programs in smaller subsamples. Nonetheless, efficiency estimates still vary considerably, ranging from full efficiency (efficiency score equal to 1) to a projected improvement of 43% for the worst performing programs. In Model 4, we further extend the conditional model to its bias-corrected form by accounting for sampling noise. Bias-corrected estimates again suggest an upward bias of conditional estimates, around 0.2 percentage points. Comparing the ratio of bias in the unconditional model to bias in the conditional model reveals that over 30% of the initial bias persists in the conditional model, despite the subsampling approach implemented to obtain the conditional estimates. 10 Accordingly, 95% confidence intervals of conditional estimates are markedly narrower on average. The final row of Table 2 presents the ratio of unconditional over conditional bias-corrected efficiency estimates. Figure 3 presents the plots of nonparametric regressions, relating this ratio to the set of conditional variables. 11 By considering the slope graphically, an interpretation can be inferred for the marginal effect of a given variable on the attainable set. The share of low educated learners appears to have an unfavorable influence on the evaluation of adult education programs, although its influence only appears once a certain threshold is reached (around 40% low educated learners in a program). This negative correlation implies that the frontier used to evaluate programs with many low educated learners is positioned below the frontier used in the unconditional model. The timing of data collection and the age of learners does not appear to be of major Fig. 3 Conditional efficiency. Plots of nonparametric regressions are presented linking the ratio of unconditional to conditional estimates and the conditional variables included in the analysis. A negative slope suggests an unfavorable influence on efficiency, while a positive slope suggests a favorable influence on efficiency. We obtained p values using nonparametric bootstraps (Li and Racine 2007), and conclude that all conditional variables are significantly related to the ratio of unconditional over conditional estimates (p < 0.05), except the share of low educated learners in a program (p 0.234) importance, whereas the gender balance of learners appears important for the assessment of programs. Overall, as the share of men increases, the evaluation of programs appears to be more favorable.

Explaining efficiency differentials
Once we obtained efficiency estimates that measure the managerial efficiency-i.e. accounting for non-discretionary variables-of adult education programs, we are interested in which discretionary variables are able to explain efficiency differentials. In particular, we focus on teacher and program characteristics (see Sect. 2). Figure 4 presents the plots of nonparametric regressions, relating the conditional and bias-corrected efficiency estimates to the size of the program (in number of learners and number of sessions), and the composition of the teacher workforce (age, gender and hours). 12 To reduce the influence of outliers, the number of learners per program and the number of sessions per program are expressed in logs.
Our results suggest that efficiency is increasing in the number of learners in a program and decreasing in the number of sessions per program, although the longest programs appear to be efficient again. In terms of demographics, it can be seen that, in contrast to learners' Fig. 4 Explaining efficiency differentials. Plots of nonparametric regressions are presented, relating the conditional and bias-corrected efficiency estimates to teacher and program characteristics. A negative slope suggests an unfavorable influence on efficiency, while a positive slope suggests a favorable influence on efficiency. We obtained p values using nonparametric bootstraps (Li and Racine 2007), and conclude that all program and teacher characteristics are significantly related to the managerial efficiency estimates (p < 0.05), except the way teaching hours are allocated between teachers (p 0.132) gender, the most efficient programs are characterized by mostly female teachers. With respect to age, a U-shaped curve is found, where the least efficient programs are taught by teachers averaging age 40. Finally, the composition of teachers seems to matter for efficiency. In general, more homogeneous groups (little variation in age, and number of hours allocated to teachers) result in more efficient adult education programs.
In order to test the robustness of our results, we repeated second stage regressions for several alternative specifications of the conditional efficiency estimates. First, we added an additional specification by also including the share of non-Belgians as a conditional variable. When non-natives sort into specific programs, this is expected to affect efficiency scores-if the share of non-natives affects the production function. Second, we take care of heterogeneity among programs by specifying an additional model with sector fixed effects instead of year fixed effects. This allows us to take out any sector-level variations, essentially comparing programs within the same sector, and hence similar learners and teachers. Third, we include center fixed effects. This specification captures all variation at the center level (taking out urban-rural variation, for example), while exploiting variation across programs within centers. For all specifications, the explained variation (R 2 ) is rather high (between 0.417 and 0.471, and 0.454 in the original model), suggesting the set of six explanatory variables is able to explain the efficiency differentials relatively well. Empirically, nonparametric plots are equivalent to Fig. 4 (see Figs. 5,6,7 in the "Appendix") when running the second stage regressions using these alternative efficiency estimation procedures. These findings imply that the set of conditional variables in the original model captures the heterogeneity in programs rather well.

Discussion and conclusion
This article introduced an innovative approach to estimate managerial efficiency in the presence of sampling noise. Using the m out of n bootstrap procedure (Simar and Wilson 2011), we extended the conditional efficiency model (Daraio and Simar 2007) by correcting for bias within conditional draws. Inputs and outputs are often noisy measures, which is especially the case when education data is used due to uncertainty emanating from survey sample designs, measurement errors or outlying observations. As a result, efficiency estimates can be biased, also when conditional models are used. We further distinguished between managerial efficiency and efficiency differentials. The former does not allow separability such that nondiscretionary variables need to be accounted for in a conditional model. The latter can be regressed on a set of discretionary variables using a two-stage approach to provide an insight into factors that explain why some units are efficient and others are not.
Using a rich panel dataset from Flanders, we illustrate the usefulness of this framework to measure and explain managerial efficiency of adult education programs. Our results suggest that a higher share of low educated learners has an unfavorable influence on the evaluation of adult education programs. By contrast, a higher share of men appears to have a favorable influence on the evaluation of adult education programs. Furthermore, we find that teacher characteristics influence the efficiency of adult education programs. In particular, more female teachers and a more homogenous teacher composition in terms of age and number of hours allocated to teachers are likely to result in a higher program efficiency. Finally, the estimated bias in our conditional model amounts to 30% of the initial bias estimate in the unconditional model, despite the subsampling approach used when estimating the conditional model.
Nonetheless, several limitations of our approach are worth noting. First, we do not claim to present causal evidence, but we contribute by using a novel empirical framework to benchmark adult education programs. Future research might expand the range of environmental variables or exploit exogenous shocks that may address causality. For example, including socio-economic characteristics (e.g. ethnicity or income) that influence both the costs of adult education (inputs) and the in-class performance (outputs) might prove an interesting robustness check beyond the ones presented here-considering our data limitations.
Second, although the education system in Flanders is similar to the education system in many OECD countries, results cannot be simply extrapolated to other countries. Nonetheless, the framework proposed here can serve as a starting point for further research. As the main use of conditional models is to account for the exogenous environment, the methodology can be applied to other educational systems as well. Finally, a possible avenue to explore is to extend the proposed model to the linearized version of FDH ('LFDH'), introduced by Jeong and Simar (2006). This extension provides a means to overcome limitations of the subsampling approach when constructing confidence intervals for the FDH estimator in finite sample applications, as suggested by Jeong and Simar (2006). Further research should address these topics.

Fig. 6
Explaining efficiency differentials (including sector fixed effect as additional conditional variable. Plots of nonparametric regressions are presented, relating the conditional and bias-corrected efficiency estimates to teacher and program characteristics. A negative slope suggests an unfavorable influence on efficiency, while a positive slope suggests a favorable influence on efficiency. We obtained p values using nonparametric bootstraps (Li and Racine 2007), and conclude that all program and teacher characteristics are significantly related to the managerial efficiency estimates (p < 0.1), except the way teaching hours are allocated between teachers (p 0.394)

Fig. 7
Explaining efficiency differentials (including center fixed effect as additional conditional variable. Plots of nonparametric regressions are presented, relating the conditional and bias-corrected efficiency estimates to teacher and program characteristics. A negative slope suggests an unfavorable influence on efficiency, while a positive slope suggests a favorable influence on efficiency. We obtained p values using nonparametric bootstraps (Li and Racine 2007), and conclude that all program and teacher characteristics are significantly related to the managerial efficiency estimates (p < 0.1)