About cookies on this site Our websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising. For more information, please review your options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.
Last updated: Feb 06, 2025
Description
Models might generate code that causes harm or unintentionally affects other systems.
Why is harmful code generation a concern for foundation models?
The execution of harmful code might open vulnerabilities in IT systems.

Example
Generation of Less Secure Code
According to their paper, researchers at Stanford University investigated the impact of code-generation tools on code quality and found that programmers tend to include more bugs in their final code when they use AI assistants. These bugs might increase the code's security vulnerabilities, yet the programmers believed their code to be more secure.
Parent topic: AI risk atlas
We provide examples covered by the press to help explain many of the foundation models' risks. Many of these events covered by the press are either still evolving or have been resolved, and referencing them can help the reader understand the potential risks and work towards mitigations. Highlighting these examples are for illustrative purposes only.