What is red teaming for generative AI?ExplainerKim Martineau11 Apr 2024Adversarial Robustness and PrivacyAIAI TestingFairness, Accountability, TransparencyFoundation ModelsNatural Language ProcessingSecurityTrustworthy AI
An open-source toolkit for debugging AI models of all data typesTechnical noteKevin Eykholt and Taesung Lee08 Sep 2023Adversarial Robustness and PrivacyAI TestingData and AI Security
AI diffusion models can be tricked into generating manipulated imagesNewsKim Martineau05 Jun 2023AIAI TestingData and AI SecurityFoundation ModelsGenerative AISecurity
DOFramework: A testing framework for decision optimization model learnersTechnical noteOrit Davidovich02 Feb 2023AIAI TestingMathematical Sciences
Breaking ReAct Agents: Foot-in-the-Door Attack Will Get You InItay NakashGeorge Kouret al.2025NAACL 2025
Exploring Straightforward Methods for Automatic Conversational Red-TeamingGeorge KourNaama Zwerdlinget al.2025NAACL 2025
ASTER: Natural and Multi-language Unit Test Generation with LLMsRangeet PanMyeongsoo Kimet al.2025ICSE 2025
Workshop on Neuro-Symbolic Software EngineeringChristian Medeiros AdrianoSona Ghahremaniet al.2025ICSE 2025
Combinatorial Test Design Model Creation using Large Language ModelsDebbie FurmanEitan Farchiet al.2025IWCT 2025
Evolution of catalysis at IBM: From microelectronics to biomedicine to sustainability with AI-driven innovationJames HedrickTim Erdmannet al.2025ACS Spring 2025