Enterprise-ready generative AI chatbot with RAG capabilities.
The AWS GenAI LLM Chatbot is a production-ready solution that enables organizations to deploy a secure, feature-rich chatbot powered by large language models (LLMs) with Retrieval Augmented Generation (RAG) capabilities.
- Multiple LLM Support: Amazon Bedrock (Claude, Llama 2), SageMaker, and custom model endpoints
- Nexus Gateway Integration: Connect to Nexus Gateway for additional model access
- Comprehensive RAG Implementation: Connect to various data sources for context-aware responses
- Enterprise Security: Fine-grained access controls, audit logging, and data encryption
- Conversation Memory: Full conversation history with persistent storage
- Web UI and API Access: Modern React interface and API endpoints for integration
- Cost Optimization: Token usage tracking and cost management features
- Deployment Flexibility: Multiple deployment options to fit your needs
This blueprint deploys the complete AWS GenAI LLM Chatbot solution in your AWS account.
- AWS Account with appropriate permissions
- AWS CLI configured with credentials
- Node.js 18+ and npm
- Python 3.8+
The deployment process is fully automated using AWS CDK and SeedFarmer.
The solution architecture includes:
- Amazon Bedrock for LLM access
- Amazon OpenSearch for vector storage
- Amazon S3 for document storage
- Amazon Cognito for authentication
- AWS Lambda for serverless processing
- Amazon API Gateway for API access
- React-based web interface
For complete documentation, visit the GitHub repository.
This project is licensed under the MIT-0 License.