Gosia Lazuka, Andreea Simona Anghel, et al.
SC 2024
Although the rise of Large Language Models (LLMs) in en- terprise settings brings new opportunities and capabilities, it also brings challenges, such as the risk of generating inap- propriate, biased, or misleading content that violates regu- lations and can have legal concerns 1. To alleviate this, we present “LLMGuard”, a tool that monitors user interactions with an LLM application and flags content against specific behaviours or conversation topics. To do this robustly, LLM- Guard employs an ensemble of detectors.
Gosia Lazuka, Andreea Simona Anghel, et al.
SC 2024
Yidi Wu, Thomas Bohnstingl, et al.
ICML 2025
Ben Fei, Jinbai Liu
IEEE Transactions on Neural Networks
Robert Farrell, Rajarshi Das, et al.
AAAI-SS 2010