The paper concerns the study of new classes of parametric optimization problems of the so-called infinite programming that are generally defined on infinite-dimensional spaces of decision variables and contain, among other constraints, infinitely many inequality constraints. These problems reduce to semi-infinite programs in the case of finite-dimensional spaces of decision variables. We focus on DC infinite programs with objectives given as the difference of convex functions subject to convex inequality constraints. The main results establish efficient upper estimates of certain subdifferentials of (intrinsically nonsmooth) value functions in DC infinite programs based on advanced tools of variational analysis and generalized differentiation. The value/marginal functions and their subdifferential estimates play a crucial role in many aspects of parametric optimization including well-posedness and sensitivity. In this paper we apply the obtained subdifferential estimates to establishing verifiable conditions for the local Lipschitz continuity of the value functions and deriving necessary optimality conditions in parametric DC infinite programs and their remarkable specifications. Finally, we employ the value function approach and the established subdifferential estimates to the study of bilevel finite and infinite programs with convex data on both lower and upper level of hierarchical optimization. The results obtained in the paper are new not only for the classes of infinite programs under consideration but also for their semi-infinite counterparts.
Variational analysis and parametric optimizationWell-posedness and sensitivityMarginal and value functionsGeneralized differentiationOptimality conditionsSemi-infinite and infinite programmingConvex inequality constraintsBilevel programming