Motivation

With the growth in world economy and population, the global energy demand is projected to rise from 570 exajoules in 2015 to 1000 exajoules by 2070 [1, 2]. As a result, the energy availability and usage will continue to be key challenges our society faces. Today’s heavily fossil fuel-based energy supply chain has developed successfully over decades to produce reliable, available, and affordable energy to various industries and sectors. Unfortunately, this dependence on fossil resources results in the release of large amounts of greenhouse gas (GHG) emissions that affect the environment and accelerate climate change [3, 4]. Meeting the increasing energy demand, while reducing GHG emissions, will arguably be one of the biggest challenges for 21st century engineers, scientists, economists, and policy makers.

Energy production, conversion, and delivery systems of the future should not only help us meet the increasing energy demand and be economically feasible, but also (i) reduce GHG emissions and environmental pollution, (ii) increase energy savings while using less resources, and (iii) shift from fossil fuel-based technologies to larger shares of renewable resources [57]. These concerns prevail in various energy sectors such as power & electricity generation, transportation, heavy industry, and residential & commercial [8]. Additionally, operation of one sector affects the others since all these energy systems are to some degree connected [9]. In the past, each of these energy systems have been treated separately by their own technical community or political groups; however, holistic solution strategies are becoming more popular recently due to the possibility of exploiting the similarities, interconnections, and synergies between different energy systems [10].

As the integrated systems grow in complexity, a traditional method for energy systems design such as using heuristics that rely on rules of thumb become less useful to a decision-maker. While heuristics combined with experience can generate quick solutions that are often reasonably good, it does not provide a way to establish the quality of the solution. Furthermore, conflicting objectives or comparison of alternatives in a problem might require arbitrary decisions. As an alternative, high-efficiency targeting approaches that focus on reaching thermodynamic limits can reduce energy consumption, but most of the time ignore capital cost considerations. They might also require considerable insight and trial & error [11, 12]. In contrast to these two approaches, mathematical optimization-based methods that rely on algorithms and simultaneous consideration of physics, chemistry, biology, and economics in a system have proven themselves as promising tools to help decision-makers generate design and operational strategies for integrated systems. Previously mentioned approaches can find good and near optimal solutions to a difficult problem. On the other hand, an optimization approach aims to find the best possible solution to the problem by quantifying the “goodness” of solutions. Optimization methods thrive when tackling systems with high degrees of freedom. Since integration means an increase in the degrees of the freedom, this translates into bigger room for improvement for energy systems. Rigorous optimization methods do not rely on trial & error; instead, they are build on systematic solution strategies [13, 14].

Energy systems engineering [15] methods aim to provide a generic framework to arrive at realistic integrated solutions to complex energy and environmental problems. Energy systems engineering puts mathematical optimization at its core to make systematic and quantitative analysis of design and operational decisions for energy systems ranging from nanoscale to megascale levels over time horizons that range from milliseconds to months or years [10, 16]. Energy systems engineering has been successfully applied to optimizing design and operation in various sectors such as fuels and chemicals production and distribution [1722], conventional and unconventional oil production [2326], biofuels and biorefineries [2729], and urban energy systems [3033].

In this commentary, we aim to introduce some key methodologies of energy systems engineering to show the versatility and resourcefulness of its tools. Then, we present some representative case studies highlighting the application of these methodologies into energy systems of different scales. Finally, we will briefly comment on a few directions that will be explored more rigorously in the upcoming years.

Methodologies of energy systems engineering

The primary aims of energy systems engineering are the design and operation of energy intensive processes in a more efficient and economic manner through mathematical optimization. In this section, we present some of the important tools used in energy systems engineering. Figure 1 summarizes the concepts discussed below.

Fig. 1
figure 1

Key methodologies of energy systems engineering. The judicious utilization of energy feedstocks requires optimal energy system design and operation, incorporating important concepts from modeling & data analysis, multi-objective optimization, and optimization under uncertainty

Mathematical modeling

Before we discuss the design and operation of energy systems, we first briefly mention how the equations describing energy systems (i.e. the constraints for optimization problems) are developed. Approaches for modeling energy systems [34] are: (1) first-principles, (2) data-driven, and (3) hybrid. First-principles modeling, also referred to as white-box, is using theory and mechanistic insights to derive the mathematical equations like mass, momentum, and energy balances that govern the energy system [35, 36]. A complete physical understanding of the energy system is presumed. On the other hand, data-driven or black-box models assume no physical insights and construct mathematical relationships solely based upon historical data from the energy system. Data-driven modeling is effective when a mechanistic understanding of the energy system is either not achievable or computationally too expensive. These data-inspired surrogate models include regression, classification, interpolation, or artificial neural network techniques [3739].

Incorporating concepts from both methods, hybrid or grey-box modeling is using theory and data to build a mathematical representation of the energy system. Some physical understanding is presumed; data is utilized to guide and adjust the first-principles equations in areas where there is a lack of insight [4042]. Hybrid models are increasingly becoming a mainstay in energy systems engineering as energy systems grow more complex and pure theoretical approaches are not sufficient [43, 44], especially in applications such as renewable energy infrastructural design [45] and refinery manufacturing operations [46].

Optimal design

In the process systems engineering (PSE) community [47], design of energy systems is traditionally performed through a superstructure-based approach [48]. A superstructure is a systematic abstraction that consists of all possible alternatives in an energy system design including different system configurations, process integration, pathway interactions, operating conditions, and other important design parameters. For example, in designing a process for manufacturing a chemical product, possible alternatives could be different feedstocks, heat and power generation sources, technological units, and operational modes. This is the classical process synthesis problem [49, 50] that originates from optimizing the design of heat exchanger networks in the 1980s [51]. Other energy system design problems include molecular design [5254], material discovery [5557], process intensification [5860], and supply chain networks [6163]. From representing an energy system design as a superstructure through mathematical equations, an optimization problem is formulated and solved to determine the optimal design from all other candidates.

Selection between design options is a discrete decision, while continuous modeling is needed to capture first principles like mass and energy balances. Linear programming (LP), first introduced in the 1930s and 40s [64, 65] to optimize military efforts during the war, handles continuous variables, but cannot illustrate discrete decision-making. By introducing binary (0-1) variables into the LP formulation, mixed-integer programming (MIP) is well-suited to model energy system superstructures [11]. In a MIP model, binary variables capture the discrete decisions and their realizations correspond to the selection (or not) of an alternative option. In fact, any logical condition between discrete events in a superstructure can be expressed through binary variables [66]. Depending on the nature of energy system, the continuous variables could be linearly or nonlinearly related, resulting in mixed-integer linear (MILP) or mixed-integer nonlinear (MINLP) problems. While solving MIP models has been a challenge, commercial solvers have dramatically improved over the years, especially for MILP problems, due to the significant developments in solution algorithms and increases in computational power [67, 68]. Large-scale MILP problems and modestly-sized MINLP problems are now routinely solved using commercial software. Nevertheless, customized algorithms are still necessary to solve specific instances of MIP problems, especially large-scale nonconvex MINLP types to global optimality [6971].

Optimal operation

Once an energy system design has been implemented, the focus switches to its operation. Unlike the design phase, determining the optimal operation of an energy system involves temporal and dynamic considerations. An energy system is usually designed to operate at defined setpoints, and given certain conditions that actualize, it may operate at other setpoints as well. Control strategies are required to keep the energy system at these setpoints. For example, in keeping a reactor at a specific temperature or its product output at a targeted purity, proportional-integral-derivative (PID) controllers [72] are typically implemented. While these controllers are proven solutions, they do not provide any guarantee of optimality or stability because they are empirical and not model-based. In the PSE community, model predictive control (MPC) is a model-based control algorithm, first conceived in the 1979 by Cutler and Ramaker at Shell [73], for optimal and stable operation. State-space models derived from the energy system design are used for the MPC [74]. Analogies between control and other energy system operational issues at different timescales, such as production planning and scheduling [7579], have been mentioned.

A complete rigorous modeling of energy system operations would necessitate dynamic programming (DP), where equations for each time period are fed as inputs into the equations for the next time period and so on in a recursive fashion. DP traces its development back to Bellman in the 1950s [80]. While a thorough approach, DP models are typically too large and complex to solve in a timely manner. They suffer from the curse of dimensionality as the amount of equations explodes exponentially as the number of time periods grows. As a workaround, instead of capturing the entire timeline, operational models for planning, scheduling, and control are usually applied in a discrete-time rolling or moving horizon manner [74]. Here, only a smaller horizon of the entire timeline is modeled initially, and then as time passes, the model is stepped forward in time by advancing this horizon. Operational models can be either linear or nonlinear; however, they are typically linear [81] as the time component already places a heavy computational burden on the solution of DP models and finding a feasible operation is sometimes more important than locating the absolute best, especially in scheduling problems [82]. Large-scale linear DP models are now regularly solved for energy system operations [8386].

Simultaneous design and operation

Up until now, the design and operation of an energy system has been constructed as two separate problems; in actuality, they are intricately related and solution to one depends on the solution to the other. When considered alone, it is possible that an optimal design may not have a feasible operation, or vice versa. Moreover, separate optimal design and optimal operation may not reflect the true optimum of the overall energy system and lead to suboptimal results. Therefore, it is paramount to develop strategies to integrate together energy system design and operation [87, 88]. One such method is multi-parametric programming (mpP), where an optimization problem is solved for a range and as a function of multiple parameters [89]. mpP dates back to the 1950s, and early work originated in the field of sensitivity analysis for LP [90].

In a mpP model combining energy system design and operation, the operation can be optimized in terms of the design parameters – that is the optimal operation is expressed as a function of the design (i.e. a design-dependent operation). In other words, mpP maps lower level decisions such as design to upper level decisions like control and scheduling [91]. An mpP approach has been utilized to investigate the simultaneous design and operation of a continuously stirred tank reactor, a binary distillation column, and a combined heat and power generation unit through design-dependent controllers [92]. Furthermore, as renewable energies consumption grows, the integration of design and operation will become an even more important research area within energy systems engineering for addressing intermittency [9396].

Multiple criteria decision making

In constructing and solving these optimization problems, model equations describing the energy system become constraints, and the criterion one wishes to optimize over is the objective function. Typical criteria for energy systems are energy efficiency, economic performance, carbon footprint, and environmental impact indicators. Depending on the choice of objective function, the optimal solution will vary – there are competing interests among the different criteria. For example, in designing a gas combustion engine, an economic profit maximization may determine coal feedstock to be most lucrative, but an environmental emissions minimization may choose biomass as the eco-friendliest. While a single objective function, usually an economic one, is characteristic of most energy systems engineering problems, challenges with resource supplies and greenhouse gas emissions require energy systems that are holistically designed and operated with respect to efficiency, economics, and environmental factors [9799]. Multi-objective optimization (MO), the simultaneous optimization according to two or more conflicting criteria, is one suitable strategy to consider the existing trade-offs among different objective functions [100102]. The goal of MO is to calculate a Pareto front, a set of equally good optimal solutions, that captures these trade-offs between conflicting criteria [103]. A decision-maker can then make an optimal decision based upon his individual interests and preferences for different target levels for each criterion. The epsilon-constraint method is the most popular method for solving MO problems [104].

Uncertainty in design and operation

In all the optimization problems discussed thus far, parameter values have been assumed to be definitively known. In reality, this is often not the case, and there is some uncertainty associated with the parameters. For example, crude oil prices can fluctuate geographically and throughout the year; this influences the optimal design and operation of a refinery. Depending on what parameter values are realized, a solution to an optimization problem may no longer be optimal or even feasible because the parameter in a constraint has taken on a different value than what was first assumed when solving the problem. Actual realizations of uncertain parameters can affect the solution’s quality. Therefore, it is important to account for sources of uncertainty in energy systems during the model development phase, especially since the uncertainty could propagate between different levels of the design or operation.

From a modeling prospective, uncertainty can be addressed either stochastically or probabilistically [105107]. Stochastic programming (SP) treats the uncertain parameters as random variables that when considered together generate different scenarios over which to model and optimize [108, 109]. SP problems are typically solved using Monte Carlo [110] or stage-wise decomposition techniques like Bender’s [111]. Robust optimization (RO) bounds uncertain parameters within uncertainty sets consisting of all possible realizations and assigns probabilities to parameter violation of bound [112114]. While this guarantees feasibility, resulting robust solutions are often overly conservative. For this reason, probabilistic guarantees on constraint violation are implemented to improve the performance of robust solutions [115120]. RO problems can be solved with commercial MIP solvers, after reformulating them as deterministic optimization problems using strong duality properties of LP. Notwithstanding, there is no agreed upon standard method to account for uncertainty. In the literature, both SP and RO have been employed to study energy systems of industrial importance [121, 122]. The right technique likely depends on the specific energy system being studied [123].

These methodologies presented above form the fundamental basis of energy systems engineering. While not exhaustively comprehensive, the goal was to describe with sufficient detail the essence of energy systems engineering toward addressing the complex design and operation of energy systems. In the following sections, applications in several interesting energy systems are presented to highlight the utility and power of energy systems engineering.

Applications of energy systems engineering

The methodologies listed in the previous section have been applied to various energy systems. Interested readers are encouraged to read the following publications from our research group on the listed topics below:

  1. 1)

    Optimal production of fuels and chemicals through process synthesis [124127]

  2. 2)

    Supply chain analysis of fuels and chemicals [128, 129]

  3. 3)

    Polygeneration energy systems [130132]

  4. 4)

    Combined heat and power generation systems [92, 133, 134]

  5. 5)

    Design and operation of fuel cells and electrolyzers [135138]

  6. 6)

    Food-energy-water nexus [139141]

  7. 7)

    Fault detection and diagnosis in chemical processes [142, 143]

To show the power of energy systems engineering analyses, we will present a few studies in detail in this chapter. Selected are three examples of an energy systems engineering approach to tackle the multi-faceted and multi-scale challenges in the design, supply chain, and operation of producing energy carriers. We will first present our optimal energy carrier production process design using process synthesis, then follow with a supply chain analysis of Texas. After that, we will show how design and operation of a PEM electrolyzer can be considered simultaneously using mpP techniques.

Process synthesis and global optimization for sustainable ammonia production

Ammonia is one of the most widely produced chemicals in the world. Global ammonia production in 2015 was reported to be over 140 million tons [144]. While currently more than 80% of the produced ammonia is used for fertilizer production, it also offers a promising potential as a renewable energy carrier. If produced from renewable resources, ammonia does not produce any GHGs when converted back to power. It has a high hydrogen content (17.8 wt.%) and more favorable storage and transportation characteristics compared to other energy carriers like pressurized or liquefied hydrogen [145]. Due to this dual opportunity, demand for ammonia in the future is expected to grow. While industrial ammonia synthesis (Eq. 1) by the famous Haber-Bosch process is very-well established and has been finely optimized during its 100 years of practice, it is energy intensive (requires 28-30 GJ/ton of ammonia) and has a significant carbon footprint (on average 2.8 tons of CO 2/ton of ammonia) due to its dependence on fossil feedstocks for hydrogen and power generation [146, 147].

$$\begin{array}{@{}rcl@{}} N_{2} + 3H_{2} \rightleftharpoons 2NH_{3}, \;\;\; \Delta H^{\circ} = -91.4 \; kJ/mol \end{array} $$
(1)

In their work, Demirhan et al. [127] used process synthesis and superstructure optimization to compare ammonia production from different renewable feedstocks and production routes. They analyzed the economic feasibility of sustainable ammonia production through comparing the effects of GHG emission restrictions, plant location (i.e. different utility and feedstock prices and availability), and plant scales on production costs. The natural gas-based production route is used as a reference case.

A conceptual design of the ammonia production facility is illustrated in Fig. 2. It consists of three main components: (i) plant, (ii) utility system, and (iii) heat recovery system. These components are highly integrated; they exchange power, heat, and process streams. The plant takes in raw materials and converts them to products. Depending on the process, the plant can consume or generate electricity and/or heat. The ammonia plant has subsections in itself: (1) natural gas reforming, (2) biomass gasification, (3) water electrolysis, (4) synthesis gas cleaning, (5) air separation unit, and (6) Haber-Bosch process for ammonia synthesis. Each of these subsections can involve reactor, separation, and recycle subsystems. Process alternatives of the plant and the connections are presented in Fig. 3. The utility system consists of heat and power generation units and waste water treatment facilities. It takes fuel, air, and water to provide the ammonia plant with electricity, power, and steam. It also provides the heat recovery system with hot and cold utilities. Heat recovery system plays a very important role in utilizing the wasted heat from the ammonia plant to minimize the hot and cold utility requirements. Process synthesis strategies can generate optimal process flowsheets with simultaneous heat, power, and water integration by exploiting the interactions between these three components [148, 149].

Fig. 2
figure 2

Conceptual design of a process. Overall process consists of three main components: plant, utility system, and heat recovery system

Fig. 3
figure 3

Process superstructure for ammonia synthesis. Process superstructure for ammonia synthesis consists of natural gas reforming, biomass gasification, water electrolysis, synthesis gas cleaning, air separation unit, and Haber-Bosch process sections

When all the technology alternatives, operating condition options, interconnections, and heat, power, and water integration systems are embedded in the postulated process superstructure, a large scale nonconvex MINLP is obtained in the form shown in Eq. 2,

$$\begin{array}{@{}rcl@{}} \min_{x,y} & & f(x,y) \\ s.t. & & h(x,y) = 0 \\ & & g(x,y) \leq 0 \\ & & x \in R^{n} \\ & & y \in \{0,1\}^{k} \end{array} $$
(2)

where x is a vector of continuous variables that represents the molar flow rates, compositions of the process streams, split fractions, total enthalpy flows, transferred or absorbed heat, and costs of the processing units.

y is a vector of 0-1 variables that denote the potential existence of a process unit (e.g. 1 if a unit is selected, 0 else).

f(x,y), the objective function, is the performance criterion that is the levelized total cost of ammonia production.

h(x,y) are the equality constraints that denote stream connections, total mass/component/atomic balances, energy balances, equilibrium relationships, input-output relationships for black-box units which constitute the process constraints as well as unit investment costs functions.

g(x,y) are the inequality constraints which correspond to design specifications, restrictions (e.g. GHG emissions, product purities, etc.), feasibility constraints, and logical constraints (e.g. select only one type of reactor). Solution of such large nonconvex MINLPs requires global optimization techniques and tailored algorithms. Interested readers are encourage to read work by Floudas and coworkers to learn more about the global optimization methods used in this work [70, 148].

One case study from the work focused on ammonia production in Texas (TX), where GHG emissions are restricted to 25% of a traditional natural gas-based ammonia plant and production capacity is set as 500 metric tons/day. Considered production routes include natural gas reforming (NG), hardwood-type (HW) forest residue gasification, municipal solid waste (MSW) type biomass gasification, wind-powered water electrolysis (W), and solar-powered water-electrolysis (S). Tables 1 and 2 show the total production cost and investment cost breakdowns.

Table 1 Total production cost breakdown of ammonia plants for Texas
Table 2 Investment cost breakdown of ammonia plants for Texas

TX-HW-500 production cost is lower than that of the base case TX-NG-500. TX-MSW-500 has higher production costs, mainly due to expensive cleaning operation used for MSW processing. Wind- or solar-powered water electrolysis-based ammonia production has high production costs, due to high electricity consumption of the electrolyzers. Sensitivity studies show that water electrolysis-based ammonia production only becomes competitive when renewable electricity prices are very low. This is important to note, because the excess electricity production from renewables can often be sold at a negative price in states like California to prevent overloading the grid. Such excess production can be used to power electrolysis to store intermittent solar resources in renewable ammonia.

Energy carriers supply chain optimization

A challenging barrier to greater integration of renewable energies such as solar and wind is their intermittency. Solar irradiation and wind speeds fluctuate hourly, daily, seasonally, and geographically. Moreover, solar and wind availabilities are often asynchronous with consumer energy demands. One potential solution to the intermittency problem is storing energy during periods and in areas of excess supply. Later on, the stored energy can be utilized when renewable energies are not directly available. The DOE [150], IRENA [151], and IEA [152] have acknowledged that developing cost-effective energy storages is a crucial step for the wider adoption of renewable energies. Options for electrical energy storage include pumped-storage hydroelectricity (PSH), compressed air energy storage (CAES), batteries, and chemical compounds [153, 154].

While they are mature and already deployed large-scale technologies, PSH and CAES are geographically limited in their suitable construction sites. On the other hand, the storage capacity of batteries is much smaller. At current costs, the scale-up of batteries is prohibitively expensive, and they are more fit for distributed applications [152, 155]. Energy can be stored in chemical compounds through renewable energy powered water electrolysis to produce hydrogen [156]. Other chemicals, such as ammonia [157, 158] and methanol [159], can also be synthesized from hydrogen. Storing energy in chemicals is attractive because their production is well-studied, can be easily scaled up to large volumes, and benefits from economies of scale. Moreover, chemicals have higher energy content than batteries and are geographically flexible in terms of where they can be produced and consumed. In this latter regard, chemicals can act as energy carriers, storing and transporting renewable energy from regions of excess supply to demand areas (Fig. 4). Energy carriers are then converted back to electricity on-demand through fuel cells or gas turbines. Compared to other storage media, energy carriers have several more intermediary steps, and this is an existing cost barrier to overcome.

Fig. 4
figure 4

Geographic mismatch of renewable energy and population in Texas. Solar and wind energy is concentrated away from the five most populous cities: Houston, Dallas, Ft. Worth, San Antonio, and Austin. Energy carriers can be used to bridge this mismatch

An infrastructure that coordinates the logistics of carriers storing and transporting energy is a complex energy system in which the optimal design is not self-evident. Among other decisions, it requires the following key considerations: which renewable resources to utilize as feedstock, which energy carriers to produce, what types of production facilities to build, where to build, what means to transport carriers, where to send them, and what conversion technologies to use. Methodologies from energy system engineering are needed to design a cost-effective energy carrier supply chain network that maximizes the carriers’ potential and is competitive with PSH, CAES, and batteries. Previous works [129, 160162] have only consider single energy carriers in the supply chain, when in fact the optimal may include a combination of multiple ones. Here, all carrier options are collected into a network superstructure and modeled using a MILP formulation. Figure 5 shows the necessary input parameters into an energy carrier supply chain model. Binary variables determine the location and type of production facilities and conversion technologies, whereas continuous variables are associated with network flows and power capacities. Equality constraints denote network flow balances, while inequalities govern resource limitations and logical constraints. The overall objective function of the model (Eq. 3) is to minimize the levelized cost of electricity (LCOE) delivered at the demand locations.

$$\begin{array}{@{}rcl@{}} \text{min} & & {Cost}_{Invest} + {Cost}_{O\&M} + {Cost}_{Feedstock} + {Cost}_{Transport} \\ & & + {Cost}_{Storage} + {Cost}_{Land} - {Sales}_{Oxygen} \end{array} $$
(3)
Fig. 5
figure 5

Inputs for energy carrier supply chain. Important to accurately incorporate relevant parameters from databases into the model

A preliminary case study implementing the aforementioned MILP model is considered for designing an energy carrier supply chain network in Texas. Wind energy is concentrated in the central and northern part of the state, whereas solar energy is most abundant in the west; however, the majority of the population lives closer to the eastern part. In this setup, energy carriers are storing and transporting renewable energy from resource-rich areas to the five most populous cities in Texas. Using energy carriers to replace 100% of the electricity demand, assuming 3-month storage time, the LCOE is $0.556/kWh with the total energy carrier profile being 0.2% hydrogen, 55.5% ammonia, and 44.3% methanol (Fig. 6). About 91.7% of the renewable energy utilized for energy carrier production comes from wind, and the production facilities are expectedly built near concentrated wind resources. Mostly rail is used for long distance transport, while truck is deployed for shorter distance deliveries. This optimal, though conservative, LCOE is very competitive with projected numbers from PSH, CAES, and batteries. Other investigated scenarios will be part of a subsequent publication [163].

Fig. 6
figure 6

LCOE of energy carriers for 3-month storage time. Hybrid mix of hydrogen, ammonia, and methanol supply the energy to the five cities. Capital costs of production and conversion of energy carriers dominate the total cost. Transportation costs contribute a minor amount

Simultaneous design and control of PEM electrolyzers

Water electrolysis for producing hydrogen is crucial to realizing sustainable ammonia production and energy carriers because it is the most upstream component in either energy system. Likewise, the electrolyzer alone is an important energy system. There are three major kinds of water electrolysis: alkaline, proton exchange membrane (PEM), and solid oxide [164]. The PEM water electrolysis process involves three major process units as shown in Fig. 7: oxygen and water management unit, electrolyzer stack, and hydrogen management unit. Water is first purified in the water management unit where ions are removed to prevent catalyst poisoning. Dissolved oxygen byproduct is also removed here. The water then goes into the electrolyzer stack where the reaction takes place. Produced hydrogen is finally separated from the unreacted water in the hydrogen management.

Fig. 7
figure 7

PEM water electrolysis. An overview of the major parts comprising the system

An advantage for PEM electrolyzers is their ability to operate at high current density, which increases the hydrogen production rate [165]. While this potentially reduces the operating cost of the electrolyzer, high current density also lowers the efficiency of the system due to increased energy losses due to faradaic resistance and overpotentials [165167]. Therefore, there is an optimal current density to operate at. Another operational consideration for PEM electrolyzers is the inlet water flowrate. The water electrolysis reaction is theoretically endothermic, but because of these energy losses, heat is generated in the process due to Joule heating. Some of the reactant water is thereby used to cool and regulate the temperature across the electrolyzer stacks. Safe operation of PEM electrolyzers requires maintaining this temperature below a certain threshold. An increased water flowrate can also reduce the effects of overpotentials such as bubble coverage [168]. Consequently, there is an optimal operating point for inlet water flowrate as well.

Because of these different operational objectives, modeling approaches [169171] from energy system engineering are needed to develop optimal operating strategies that integrate with the PEM electrolyzer design. For example, it is unnecessary to overdesign an electrolyzer stack (increasing its capital investment cost) to handle a maximum current density and water flowrate if these values will never be realized during the operation. First, high-fidelity dynamic models of PEM electrolysis are developed to represent the electrochemistry and mass & energy balances. This allows for an accurate simulation and digital twin of the PEM electrolyzer. Next, these models are reduced through statistical data methods (i.e. system identification of input/output data) to create approximate models to be used as input in the controller design. A multi-parametric model predictive control (mpMPC) approach is used to construct operating strategies that account for the electrolyzer design. The mpMPC [172] is a mpP-inspired exact reformulation of the classical linear quadratic regulator (LQR) problem (Eq. 4), allowing control laws to be explicitly expressed as a function of design parameters such as electrolyzer size and membrane material. The classical LQR for optimal operation is:

$$\begin{array}{@{}rcl@{}} \min_{u} J & & = \sum_{k=1}^{N-1}\left(y_{k}-y_{k}^{R}\right)^{T}QR\left(y_{k}-y_{k}^{R}\right)+ \sum_{k=0}^{M-1}\Delta u_{k}^{T}R\Delta u_{k} \\ s.t.& & x_{k+1} = {Ax}_{k} + {Bu}_{k} \\ & &y_{k} = {Cx}_{k} + {Du}_{k} + e \\ & &u_{min} \leq u_{k} \leq u_{max} \\ & &\Delta u_{min} \leq \Delta u_{k} \leq \Delta u_{max} \\ & &x_{min} \leq x_{k} \leq x_{max} \\ & &y_{min} \leq y_{k} \leq y_{max} \\ & &u = \left[u_{0},u_{1},...,u_{m-1}\right] \end{array} $$
(4)

where x is the state variables; uk are the control variables; Δuk denotes the difference between two consecutive control actions; yk and \(y^{R}_{k}\) are the outputs and their respective set points; R and QR are the corresponding weights in the quadratic objective function; N and M are the output horizon and control horizon, respectively; k is the time step; A, B, C, and D are the matrices of the discrete linear state-space model; and e denotes the mismatch between the actual system output and the predicted output at initial time.

In this manner, the simultaneous optimization of energy system design and operation is performed. These collective steps comprise the Parametric Optimization and Control (PAROC) framework (Fig. 8), an integrated software platform that facilitates this simultaneous optimization [89].

Fig. 8
figure 8

PAROC framework. An overview of the steps. Explicit MPC models, designed from reduced models of the energy system, are used to operate the system

Ogumerem and Pistikopoluos [138] applied the PAROC framework toward optimizing the control strategy for a PEM electrolyzer. They observed that a mpMPC approach allowed them to optimally operate a PEM electrolyzer below cell voltage and temperature limits, while using inlet water flowrate as the manipulated variable. It is also confirmed that the water flowrate is better at responding to temperature changes, while current density is better for adjusting operation for changes in cell voltage. The mpMPC model elucidated explicit control laws depending on the state of the electrolyzer (Fig. 9) and setups subsequent work on simultaneous optimization for the design and operation of PEM electrolyzers. The mathematical equations expressing the control laws can be included as additional constraints in a superstructure MIP formulation for PEM electrolyzer design. Thereby, in this fashion, the design will be optimized with regards to the optimal operation determined from the MPC. This will minimize the capital and operating costs of PEM electrolyzers, making them more competitive with their alkaline counterparts.

Fig. 9
figure 9

Critical regions for simultaneous design and operation of PEM electrolyzer. Each region represents a separate optimal operating strategy for an unique set of design parameters for PEM electrolysis. mpMPC determines the design-dependent operation

New directions

In the previous sections, we described energy systems engineering methodologies and showed a couple of examples to highlight their usefulness in the analysis of an energy carrier system. As the energy sources and systems continue to evolve, so will the energy systems engineering methods and application areas. Exciting developments in various disciplines and fields have occurred in recent years and before concluding this article, we want to touch upon a few directions that we think will be explored more vigorously in the upcoming years by energy systems engineers.

Information technologies in design and operation of energy systems

Manufacturing facilities collect large amounts of operational data thanks to improved sensor and monitoring technologies. However, usefulness of data is limited without any strong data integration, classification, visualization, and analysis methods [173]. In their 2018 article, Edgar and Pistikopoulos [174] report that many U.S. manufacturing operations are data rich and knowledge poor. They indicate that while operations use sophisticated modeling and control technologies, usage of data analytics tools in the decision-making phase is still constrained. Integrating manufacturing intelligence in real-time across an entire production operation does not currently exist. The concept of smart manufacturing (SM) is defined as using the right data in the right form, the right technology and the right operations, wherever and whenever needed throughout the manufacturing enterprise. SM combines operations technology with information technology to improve the manufacturing platforms. Integrated modeling approaches that combine sensors and monitoring, data analytics, real-time data management and cloud technologies with control and automation will become more important in the future.

Implications of artificial intelligence and machine learning for energy systems engineering

While artificial intelligence and machine learning have been ongoing research areas for many decades [175177], only recently have they gained wider attention due to the information age’s explosion of data and increasing computational power [178]. Recent notable achievements with IBM’s Watson and Google’s AlphaGo have even garnered the interest of the greater public and brought promises of how "big data" can revolutionize the way we understand and study the world [179, 180]. However, as with any new technological development, it might be better to remain cautiously optimistic to not overhype the fruits and oversell the perils, since there is still much progress to be made for artificial intelligence and machine learning to mature, penetrate, and spread to greater adoption. With enough time and directed efforts, we expect the advances in the artificial intelligence and machine learning community will migrate into energy systems engineering and become as commonly utilized methodologies as mixed-integer optimization and MPC have become in the last few decades [181]. Research fields such as catalyst design [182, 183] and drug discovery [184] are actively developing open databases to simulate data-driven and hybrid model building. Artificial intelligence and machine learning will be especially needed in areas, such as process operation and fault diagnosis, where traditional mathematical approaches are not suitable due to a lack of a first-principles basis for model development.

There are now more ample resources and accessible software to learn and implement artificial intelligence and machine learning applications compared to earlier eras. The increasing number of such tools allow users to easily code their own machine learning or neural network models with little effort. In his engaging perspective on the status of artificial intelligence in chemical engineering, Venkatasubramanian [181] points out that there is a risk of substituting the well-trained usage of such tools for actual mastery of artificial intelligence and machine learning knowledge. We agree with his assessment that it is important for future engineers to be properly educated in the “know-why” rather than just the “know-how”. We think there is a need to reform the current engineering curriculum and graduate training to include more artificial intelligence and machine learning material in order to better prepare individuals for advancing the energy systems of the future.

Extending the boundaries: Increased interdisciplinary work

In this article, we presented applications focusing on mainly chemical engineering topics. Other disciplines such as electrical, mechanical, and civil engineering, physics, chemistry, biology, operations research, statistics, computer science, agriculture, economics, political science, and law also conduct extensive research on energy systems. Each of these disciplines has its own areas of focus, goals, solution strategies, and challenges related its problems. However, as the energy systems get more complex and interconnected, close cooperation of experts from such fields becomes a necessity. One such area that has manifested itself as an interdisciplinary field is the food, energy, and water nexus (FEW-N). In FEW-N, needs for each resource are linked to both global demands as well as their interdependency. This field brings researchers, stakeholders, and policy makers together to tackle problems that are too big to deal with by each of the individual community [185]. Recent work showed that energy systems engineering methods can be useful in modeling and exploiting the interconnections between these resources. By coming up with metrics to make sure that all perspective are equally considered, energy systems engineering methods can provide solutions that are feasible, sustainable, and effective [139, 186].

We expect new interdisciplinary fields to emerge in the following years. Academic and government initiatives play a key role here to advocate interdisciplinary work. Over the recent years, we have seen an increase in the activities of university institutes focusing on energy research across the USA. Such institutes try to bring experts from the aforementioned fields together on joint projects, so that the capabilities of each discipline are understood and synergies between collaborations can be exploited. A further indicator of academia’s emphasis on interdisciplinary work is the rise in the number of master of science and master of engineering programs on energy systems in the USA. These programs have curriculum that cover topics from various engineering disciplines and social sciences to cultivate a generation of decision-makers with holistic and broader views of the energy landscape. It is our opinion, that energy systems engineering courses can play a pivotal role in bringing different disciplines together.

Extending the scope: Increased industry-academia collaboration

Another field that we expect to grow in the future is the industry-academia collaborations. So far, aims of the optimization community and industry have been different. The value of novel optimization-based energy systems engineering tools have been somewhat underestimated by industry, and therefore, unevenly utilized [187]. Academia products are high impact and peer-reviewed open publications so that funding can be ensured. The peer-review process favors fast publications with novel & sophisticated methodologies. For this reason, academia products, especially in energy systems engineering, are in the form of prototypical software. On the other hand, industry is interested in basic ideas and their resulting benefits. Industry does not show interest in migrating information between multiple tools and software packages. Additionally, most companies limit their information exchange with academia due to confidentiality reasons. As a result, academia has limited access to realistic cases [7]. However, this picture is likely to change in the future. As Department of Energy supported initiatives like Rapid Advancement in Process Intensification (RAPID) Manufacturing Institute and Clean Energy Smart Manufacturing Innovation Institute (CESMII) show, collaborations between industry and academia can work effectively when the research objectives are clearly presented for both parties [188, 189].

New modeling environments and strategies for optimal design and operation

Two of the most widely established modeling environments used for the formulation and optimization of algebraic problems are GAMS [190] and AMPL [191]. These are commercial high-level programming platforms where modeling is done by explicitly declaring all the variables, constraints, and parameters in an optimization program. While these environments have access to numerous solvers and have been used extensively by the optimization community, they suffer from difficulties in data input, manipulation, visualization, and implementation of advanced algorithms. Introduction of recent modeling environments, such as PyOMO [192] and JuMP [193] that are built upon Python and Julia programming languages, respectively, present an alternative approach via use of object-oriented programming (OOP). Both environments allow users direct access to modeling objects that are given in a model library. By doing that, the users do not have to specify governing equations, every time they add a unit to a process. This will hopefully help modelers make designs in a more standardized, intuitive way, and an easier way. PyOMO and JuMP currently have limited access to optimization solvers. However, they are license-free and open-source environments. Another difficulty with the traditional way of formulating optimization problems is modeling of process alternatives using integer variables. There is no unique way to express the logic encapsulated in the superstructure as a set of variables and constraints. Generalized disjunctive programming (GDP) techniques offer new alternatives to traditional mixed-integer modeling approaches, by directly addressing the relationships between two distinct alternatives (disjunctions) via logic-based methods [194, 195]. OOP can work well with GDP techniques to formulate more standardized mixed-integer programs [196]. As a result, we expect the formulation of optimization models to become easier in the upcoming years. This can increase the interest and accessibility of other communities in energy systems engineering tools as well.

Closing the loop: Experimental expertise

Finally, research expertise in computational modeling and experimentation are often concentrated in separate groups and housed in different locations. However, advancing energy systems to greater heights will depend on both quantitative and empirical knowledge and experience. It is usually the role of principal investigators of these groups to facilitate any teamwork between them. We have emphasized that increased interdisciplinary and industrial collaborations are necessary, and a significant reason for this is to close the loop between modeling and experimentation. Models can be unguided efforts toward abstract understanding if not supplemented by real-life results, and experiments can be tedious trial & error excursions toward physical understanding if not supported by quantitative tools. While collaborations can help close the loop, they are difficult to secure due to lack of appropriate funding and limited by networks between researchers. Therefore, it is also imperative that modelers gain some empirical familiarity and experimentalists become more versed in computation. In this way, knowledge gaps between modeling and experimentation is minimized and progress in energy systems engineering is accelerated. We strongly believe that a holistic approach to energy systems engineering necessitates knocking down walls between modeling and experimentation.

Conclusions

In this commentary, we introduced the methodologies, applications, and a few possible future directions of energy systems engineering. We hope the methods and results show the importance and strength of an energy systems engineering approach to improve the efficiency of tomorrow’s energy systems.