LLM-PIEval: a benchmark for indirect prompt injection attacks in large language models

Large language models have increased interest in Artificial Intelligence and their integration with external tools introduces risks such as direct and indirect prompt injection. LLM-PIEval provides a framework and test set to measure indirect prompt injection risk and the authors release API specifications and prompts to support wider assessment.

Large language models have become widely used in applications such as virtual assistants and smart home agents, driving broader interest in Artificial Intelligence. That same integration with external tools creates attackers’ opportunities, including direct prompt injection when malicious instructions appear in a user query and indirect prompt injection when harmful instructions are present in the retrieved information payload of retrieval augmented generation systems. The article notes indirect prompt injection carries particular risk because end users may not be aware of new attacks when they occur and detailed benchmarking of models on this threat remains limited.

To address that gap, the authors develop LLM-PIEval, a framework designed to measure any candidate large language model for its vulnerability to indirect prompt injection attacks. Using the framework the team created a new test set and used it to evaluate several state of the art large language models. The reported results show strong attack success rates across most evaluated models, demonstrating that indirect prompt injection is an active and measurable threat to current model deployments.

The authors release their generated test set together with API specifications and prompts to enable broader assessment of this risk in current large language models. By publishing these artifacts the work aims to make it easier for researchers and practitioners to evaluate model robustness to indirect prompt injection and to compare defenses and mitigations across systems. The paper frames LLM-PIEval as a practical, shareable resource to support more systematic security testing in conversational and retrieval augmented workflows.

58

Impact Score

Google Vids opens free video generation to all Google users

Google has made Google Vids available to anyone with a Google account, adding free access to video generation with its latest models. The move expands Google’s end-to-end video workflow and increases pressure on rivals that charge for similar tools.

Court warns against chatbot legal advice in Heppner case

A federal court found that chats with a publicly available generative Artificial Intelligence tool were not protected by attorney-client privilege or the work-product doctrine. The ruling highlights litigation risks when executives or employees use chatbots for legal guidance without lawyer supervision.

Newsom orders California to weigh Artificial Intelligence harms in contract rules

Gov. Gavin Newsom has signed an executive order directing California agencies to account for potential Artificial Intelligence harms in state contracting while expanding approved use of generative tools across government. The move follows a dispute involving Anthropic and reflects a broader split between California and the Trump administration on Artificial Intelligence oversight.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.