RedTeamGuides

RedTeamGuides

Large Language Model Prompts

Reza's avatar
Reza
Jan 26, 2025
∙ Paid

LLM01:2023 - Prompt Injections

In this example, the injected prompt includes a malicious command disguised as part of the translation request. The LLM, when not properly protected against prompt injections, may execute the command and delete files from the system, leading to potential data loss or unauthorized actions.

Original prompt: user_prompt = "Tran…

User's avatar

Continue reading this post for free, courtesy of Reza.

Or purchase a paid subscription.
© 2025 RedTeamGuides · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture