Skip to main content

State Constraints and Safety Consideration

  • Chapter
  • First Online:
Reinforcement Learning for Sequential Decision and Optimal Control
  • 4212 Accesses

Abstract

Controlling a real-world system with state constraints has drawn increasing attention due to practical needs, such as operating limits and safety guarantees. Equipping RL/ADP with the ability to handle constrained behaviors is of practical significance in both training process and controller implementation. Basically, there are three constrained RL/ADP methods, including penalty function method, Lagrange multiplier method, and feasible descent direction method. The phenomenon of infeasibility occurs when the constrained OCP has no solution due to overly tight state confinement, i.e., there is no available policy that can satisfy the strict constraint. Hence, handling constrained OCP is a systematic task, in which an optimal policy and its feasible working region must be simultaneously learned to ensure the recursive feasibility. A new three-element learning architecture called actor-critic-scenery (ACS) is proposed to address the issue, whose elements include policy improvement (PIM), policy evaluation (PEV), and a newly added region identification (RID) step. By equipping an OCP with hard state constraint, the safety guarantee is equivalent to solving this constrained control task to output its safe policy. Two basic training modes are proposed for safe policy search, and their corresponding safety-critical ACS algorithms can be designed in both model-free and model-based settings.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 119.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Li, S.E. (2023). State Constraints and Safety Consideration. In: Reinforcement Learning for Sequential Decision and Optimal Control. Springer, Singapore. https://doi.org/10.1007/978-981-19-7784-8_9

Download citation

  • DOI: https://doi.org/10.1007/978-981-19-7784-8_9

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-19-7783-1

  • Online ISBN: 978-981-19-7784-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics