Meta-optimization in safe reinforcement learning: Enhancing safety at training and deployment with fewer hyperparameters

dc.contributor.authorHonari, Homayoun
dc.contributor.supervisorNajjaran, Homayoun
dc.date.accessioned2024-09-25T18:36:31Z
dc.date.available2024-09-25T18:36:31Z
dc.date.issued2024
dc.degree.departmentDepartment of Mechanical Engineering
dc.degree.levelMaster of Applied Science MASc
dc.description.abstractReinforcement learning (RL) is a trial-and-error framework for enabling intelligent systems to learn the optimal behaviour based on the feedback from the environment. In recent years, successful application of RL in controlling various embodied systems have been observed. However, the real-world deployment and training of RL methods require paying attention to certain limitations imposed by the robot and its surroundings. To address these limitations, safe RL algorithms aim to define safety constraints based on the physics of the system and modify the training regime of the RL methods to satisfy them during training and inference. While safe RL offers a promising direction for achieving real-world deployability, challenges such as sample efficiency and hyperparameter tuning hinders its applicability in real-world scenarios. To address these challenges, this thesis proposes various approaches. First, a metagradient-based training pipeline called Meta Soft Actor-Critic Lagrangian (Meta SAC-Lag) is proposed which aims to optimize the aforementioned safety-related hyperparameters under the conventional Lagrangian framework. To study the performance, the proposed method is evaluated in various safety-critical simulated environments. In addition, a real-world task is designed, and the algorithm is successfully deployed on a Kinova Gen3 robotic arm to showcase its real-world deployability with minimal hyperparameter tuning requirements. Furthermore, a multi-objective policy optimization framework is proposed which specifies the trade-off between optimality and safety directly and optimizes both of them simultaneously. The competitive performance of the proposed algorithm compared to the state-of-the-art safe RL methods with fewer hyperparameters showcases its potential in providing a powerful alternative framework for safe RL.
dc.description.scholarlevelGraduate
dc.identifier.urihttps://hdl.handle.net/1828/20452
dc.languageEnglisheng
dc.language.isoen
dc.rightsAvailable to the World Wide Web
dc.titleMeta-optimization in safe reinforcement learning: Enhancing safety at training and deployment with fewer hyperparameters
dc.typeThesis

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Honari_Homayoun_MASc_2024.pdf
Size:
4.53 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.62 KB
Format:
Item-specific license agreed upon to submission
Description: