Safety Constrained Decision Making with Large Language Models
Title: |
Safety Constrained Decision Making with Large Language Models |
DNr: |
NAISS 2024/22-1025 |
Project Type: |
NAISS Small Compute |
Principal Investigator: |
Rishi Hazra <rishi.hazra@oru.se> |
Affiliation: |
Örebro universitet |
Duration: |
2024-07-31 – 2024-10-01 |
Classification: |
10201 |
Keywords: |
|
Abstract
This project seeks to advance constrained decision-making, particularly in Reinforcement Learning, by integrating Large Language Models (LLMs) with probabilistic logic-based safety shields. These safety shields will constrain the LLMs to only produce actions that comply with safety norms and other specific standards defined by the user.
We plan to implement this integration by employing probabilistic logic shields, as described in "Safe Reinforcement Learning via Probabilistic Logic Shields" by Yang et al. (IJCAI 2023). Additionally, we will utilize techniques from "SayCanPay: Heuristic Planning with Large Language Models using Learnable Domain Knowledge" by Hazra et al. (AAAI 2024) to further refine the functionality of LLM-based agents. This integrated methodology is expected to significantly improve the safety of LLMs in navigating decision-making scenarios.