서브메뉴
검색
Specification-Guided Reinforcement Learning- [electronic resource]
Specification-Guided Reinforcement Learning- [electronic resource]
- 자료유형
- 학위논문
- Control Number
- 0016931764
- International Standard Book Number
- 9798379751203
- Dewey Decimal Classification Number
- 004
- Main Entry-Personal Name
- Jothimurugan, Kishor.
- Publication, Distribution, etc. (Imprint
- [S.l.] : University of Pennsylvania., 2023
- Publication, Distribution, etc. (Imprint
- Ann Arbor : ProQuest Dissertations & Theses, 2023
- Physical Description
- 1 online resource(189 p.)
- General Note
- Source: Dissertations Abstracts International, Volume: 84-12, Section: A.
- General Note
- Advisor: Alur, Rajeev.
- Dissertation Note
- Thesis (Ph.D.)--University of Pennsylvania, 2023.
- Restrictions on Access Note
- This item must not be sold to any third party vendors.
- Summary, Etc.
- 요약Recent advances in Reinforcement Learning (RL) have enabled data-driven controller design for autonomous systems such as robotic arms and self-driving cars. Applying RL to such a system typically involves encoding the objective using a reward function (mapping transitions of the system to real values) and then training a neural network controller (from simulations of the system) to maximize the expected reward. However, many challenges arise when we try to train controllers to perform complex long-horizon tasks-e.g., navigating a car along a complex track with multiple turns. Firstly, it is quite challenging to manually define well-shaped reward functions for such tasks. It is much more natural to use a high-level specification language such as Linear Temporal Logic (LTL) to specify these tasks. Secondly, existing algorithms for learning controllers from logical specifications do not scale well to complex tasks due to a number of reasons including the use of sparse rewards and lack of compositionality. Furthermore, existing algorithms for verifying neural network controllers (trained using RL) cannot be easily applied to verify controllers for complex long-horizon tasks due to large approximation errors.This thesis proposes novel techniques to overcome these challenges. We show that there are inherent limitations in obtaining theoretical guarantees regarding RL algorithms for learning controllers from temporal specifications. We then preset compositional RL algorithms that achieve state-of-the-art performance in practice by leveraging the structure in the given logical specification. Finally, we show that compositional approaches to learning enable faster verification of learned controllers containing neural network components.
- Subject Added Entry-Topical Term
- Computer science.
- Subject Added Entry-Topical Term
- Information science.
- Index Term-Uncontrolled
- Formal specifications
- Index Term-Uncontrolled
- Reinforcement learning
- Index Term-Uncontrolled
- Reward shaping
- Index Term-Uncontrolled
- Temporal logic
- Added Entry-Corporate Name
- University of Pennsylvania Computer and Information Science
- Host Item Entry
- Dissertations Abstracts International. 84-12A.
- Host Item Entry
- Dissertation Abstract International
- Electronic Location and Access
- 로그인을 한후 보실 수 있는 자료입니다.
- Control Number
- joongbu:643201
detalle info
- Reserva
- 캠퍼스간 도서대출
- 서가에 없는 책 신고
- Mi carpeta