0 / 0
Unrepresentative data risk for AI

Unrepresentative data risk for AI

Risks associated with input
Training and tuning phase
Accuracy
Traditional AI risk

Description

Unrepresentative data occurs when the training or fine-tuning data is not sufficiently representative of the underlying population or does not measure the phenomenon of interest.

Why is unrepresentative data a concern for foundation models?

If the data is not representative, then the model will not work as intended.

Parent topic: AI risk atlas

We provide examples covered by the press to help explain many of the foundation models' risks. Many of these events covered by the press are either still evolving or have been resolved, and referencing them can help the reader understand the potential risks and work towards mitigations. Highlighting these examples are for illustrative purposes only.

Generative AI search and answer
These answers are generated by a large language model in watsonx.ai based on content from the product documentation. Learn more