Рет қаралды 141
This tutorial show you how to build a serverless GenAI RAG solution to implement a document chat feature using Amazon Bedrock Converse API and Lambda. Also, I will apply one of the newest features introduced in July 2024 which is apply guardrail so that we have control over input prompt as well as response being returned to the calling app/consumer.
Join WhatsApp: www.whatsapp.c...
👉Get CloudWays ➜platform.cloud...
💥Save big with 30% for 4 months + 10 free migrations. Offer valid till 31st August, 2024
☝️☝️ USE THE EXCLUSIVE COUPON CODE ABOVE TO GET 40% OFF FOR 4 MONTHS💥along with 40 FREE migrations handled by our expert engineers (valid till 1st December, 2023).
👉Get Digital Ocean ➜ digitalocean.pxf.io/ZQERvQ
💥Get $200 FREE Credits for signup. So, hurry up!💥
╔═╦╗╔╦╗╔═╦═╦╦╦╦╗╔═╗
║╚╣║║║╚╣╚╣╔╣╔╣║╚╣═╣
╠╗║╚╝║║╠╗║╚╣║║║║║═╣
╚═╩══╩═╩═╩═╩╝╚╩═╩═╝
Guardrail is a much needed feature supported by Amazon Bedrock to guard the contents while using a Generative AI solution.
'Chat With Document' features supported by Amazon Bedrock is a form of RAG and allows you to have a contextual conversation and ask questions based on the data in the document augmented with LLM for Generative AI.
RAG, which stands for Retrieval Augmented Generation, is becoming increasingly popular in the world of Generative AI. It allows organizations to overcome the limitations of LLMs and utilize contextual data for their Generative AI solutions.
I will use the recently released Anthropic Sonnet foundation model and invoke it via the Amazon Bedrock Converse using Lambda and API.
#aws #genai #bedrock #generativeai #amazonbedrock #awsbedrock #cloudguru #genaionaws #llm #ai