Skip to main content

Generative AI platform on AWS

Large language models (LLMs) are growing in popularity but it also surfaced new challenges. Multiple teams need to collaborate and have the right workflows and platforms in place for the deployment and management of generative AI solutions in production. Join this session as we dive into the generative platform capabilities and key considerations for deployment and management of generative AI solutions in production. Learn how to leverage various capabilities at different stages when consuming and iterating on LLMs for different use cases, such as prompt templates management, validation system, feedback system, conversation management, caching and more. The session covers MVS (Minimum Variable Service) and best practices in applying the different LLM models for your use cases. This session also covers the approaches to remove heavy lifting in infrastructure management, enabling your technical teams to focus on core tasks instead.
Speakers: 
Hao Fei Feng, Senior Cloud Architect, AWS Professional Services
Bin Liu, Senior Cloud Architect, AWS Professional Services