Threat modeling for generative AI applications
Threats on AI and machine learning (AI/ML) systems are becoming more frequent, moving beyond controlled environments to real-world production deployments. These threats target vulnerabilities such as exposure to personally identifiable information (PII), lack of oversight in decision-making, and insufficient logging and monitoring.
Conducting a thorough threat model for your generative AI application is essential. Begin by defining the level of agency that you will provide to the LLM and any AI agents that you might use. This involves determining the extent of autonomy and decision-making power the AI system will have.
Next, clearly define where authentication and authorization should
be performed. For guidance on this, see
this
blog post
Implement applicable traditional security controls for data security
and deploy AI-specific mitigations for AI safety risks. For example,
consider implementing controls such as
Amazon
Bedrock Guardrails
Carefully consider the pros and cons of logging in generative AI systems and determine the appropriate level of logging for your application. This decision will directly impact your ability to monitor, audit, and respond to incidents.
Finally, establish incident response plans that align with your logging capabilities and the specific risks associated with your generative AI application.