« Back to Glossary Index
An attack in which a prompt sent to a large language model contains malicious input intended to override existing safeguards and allow the attacker to manipulate the model’s responses or behavior to generate harmful or unintended outputs or actions
See Indirect Prompt Injection Attack and Passive Prompt Injection Attack
« Back to Glossary Index