A. Amazon Bedrock Guardrails
B. Amazon Bedrock Agents
C. Amazon Bedrock inference APIs
D. Amazon Bedrock custom models
Explanation:
The ecommerce company is deploying a chatbot that needs safeguards to filter harmful content from input prompts and responses. Amazon Bedrock Guardrails provide mechanisms to ensure responsible AI usage by filtering harmful content, such as hate speech, violence, or misinformation, making it the appropriate feature for this requirement.
Exact Extract from AWS AI Documents:
From the AWS Bedrock User Guide:
"Amazon Bedrock Guardrails enable developers to implement safeguards for generative AI applications, such as chatbots, by filtering harmful content in input prompts and model responses. Guardrails include content filters, word filters, and denied topics to ensure safe and responsible outputs."
(Source: AWS Bedrock User Guide, Guardrails for Responsible AI)
Detailed
Option A: Amazon Bedrock Guardrails This is the correct answer. Amazon Bedrock Guardrails are specifically designed to filter harmful content from chatbot inputs and responses, ensuring safe interactions for users.
Option B: Amazon Bedrock Agents Amazon Bedrock Agents are used to automate tasks and integrate with external tools, not to filter harmful content. This option does not meet the requirement.
Option C: Amazon Bedrock inference APIs Amazon Bedrock inference APIs allow users to invoke foundation models for generating responses, but they do not provide built-in content filtering mechanisms.
Option D: Amazon Bedrock custom models Custom models on Amazon Bedrock allow users to fine-tune models, but they do not inherently include safeguards for filtering harmful content unless explicitly implemented.
Reference: AWS Bedrock User Guide: Guardrails for Responsible AI (https://docs.aws.amazon.com/bedrock/latest/userguide/guardrails.html) AWS AI Practitioner Learning Path: Module on Responsible AI and Model Safety
Amazon Bedrock Developer Guide: Building Safe AI Applications (https://aws.amazon.com/bedrock/)