0 / 0
Prompt leaking risk for AI

Prompt leaking risk for AI

Risks associated with input
Inference
Robustness
Amplified by generative AI

Description

A prompt leak attack attempts to extract a model's system prompt (also known as the system message).

Why is prompt leaking a concern for foundation models?

A successful attack copies the system prompt used in the model. Depending on the content of that prompt, the attacker might gain access to valuable information, such as sensitive personal information or intellectual property, and might be able to replicate some of the functionality of the model.

Parent topic: AI risk atlas

Generative AI search and answer
These answers are generated by a large language model in watsonx.ai based on content from the product documentation. Learn more