GENSEC01-BP02 Implement private network communication between foundation models and applications - Generative AI Lens

GENSEC01-BP02 Implement private network communication between foundation models and applications

Implementing a scoped down data perimeter on foundation model endpoints helps reduce the surface-area of potential threat vectors and encourages a zero-trust security architecture. This best practice describes how to implement private network communications for your generative AI workloads.

Desired outcome: When implemented, this best practice reduces the risk of unauthorized access to a foundation model endpoint. It also helps create a process to grant least privileged access to authorized parties.

Benefits of establishing this best practice:

  • Apply security at all layers - Private network communications facilitate an additional layer of security within your application.

  • Protect data in transit and at rest - Using private networks instead of the default public endpoints helps to protect data in transit, especially when combined with encryption techniques.

Level of risk exposed if this practice is not established: High

Implementation guidance

Without private network communication between foundation model endpoints and generative AI applications, access to these endpoints would be available through the public internet, increasing exposure. Implementing a private network between a foundation model and a generative AI application requires full control over application hosting and network traffic configuration.

AWS PrivateLink supports a range of AWS generative AI managed services, including Amazon Bedrock and the Amazon Q family of services. AWS PrivateLink facilitates private network communications for customers across the AWS network within their own account. This capability enables customers to maintain private network communication between generative AI managed services and applications making the request without using the public internet. AWS PrivateLink works for self-managed services as well, like Amazon SageMaker AI. Hosted inference endpoints in Amazon SageMaker AI can be deployed in a Virtual Private Cloud (VPC). In addition to network controls which protect and secure infrastructure, endpoints deployed in a VPC can be made private using AWS PrivateLink. AWS PrivateLink enables VPC instances to communicate with service resources without the need for public IP addresses, reducing potential threats from public internet exposure.

Implementation steps

  1. Determine the VPC you need to create a private endpoint in.

  2. Select the service you wish to create a private route to from your VPC.

  3. Configure the endpoint to allow least privilege access for your services.

    • Network access to the interface endpoint is controlled using Security Groups and Policy Documents.

Resources

Related practices:

Related guides, videos, and documentation:

Related examples: