ai-safety Operationalizing a Threat Model for Red-Teaming Large Language Models (LLMs) Paper • 2407.14937 • Published Jul 20 • 1
Operationalizing a Threat Model for Red-Teaming Large Language Models (LLMs) Paper • 2407.14937 • Published Jul 20 • 1