aws-samples/reinvent2023-aim-329
How to build a chat assistant with Amazon Bedrock using LLMs, Embeddings model, and Knowledge Bases for Amazon Bedrock.
This helps developers create custom chat assistants that can answer questions using up-to-date, domain-specific information. You provide your proprietary documents or data, and the assistant uses this knowledge to respond to user queries, avoiding generic or outdated information. This is for developers building AI-powered conversational tools for their applications or internal systems.
No commits in the last 6 months.
Use this if you need to build a secure, accurate chat assistant that leverages your own specific data to provide answers, rather than relying solely on a large language model's general knowledge.
Not ideal if you're looking for a pre-built, ready-to-deploy chat assistant without any custom development or integration work.
Stars
18
Forks
5
Language
Jupyter Notebook
License
—
Category
Last pushed
Jun 17, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/aws-samples/reinvent2023-aim-329"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
aws-samples/generative-ai-use-cases
Application implementation with business use cases for safely utilizing generative AI in...
aws-samples/serverless-rag-demo
Amazon Bedrock Foundation models with Amazon Opensearch Serverless as a Vector DB
aws-samples/amazon-bedrock-rag
Fully managed RAG solution implemented using Knowledge Bases for Amazon Bedrock
IBM/granite-workshop
Source code for the IBM Granite AI Model Workshop
aws-samples/rag-with-amazon-bedrock-and-opensearch
Opinionated sample on how to build and deploy a RAG application with Amazon Bedrock and OpenSearch