MART: Improving LLM Safety with Multi-round Automatic Red-Teaming Paper • 2311.07689 • Published Nov 13, 2023 • 7
Unveiling Safety Vulnerabilities of Large Language Models Paper • 2311.04124 • Published Nov 7, 2023 • 6