Python Fuzzing for Trustworthy Machine Learning Frameworks Paper • 2403.12723 • Published Mar 19, 2024 • 2
Red Teaming GPT-4V: Are GPT-4V Safe Against Uni/Multi-Modal Jailbreak Attacks? Paper • 2404.03411 • Published Apr 4, 2024 • 8
Teams of LLM Agents can Exploit Zero-Day Vulnerabilities Paper • 2406.01637 • Published Jun 2, 2024 • 1
AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases Paper • 2407.12784 • Published Jul 17, 2024 • 48
Cybench: A Framework for Evaluating Cybersecurity Capabilities and Risk of Language Models Paper • 2408.08926 • Published Aug 15, 2024 • 5