This put up is cowritten with David Kim, and Premjit Singh from Ring.
Scaling self-service assist globally presents challenges past translation. On this put up, we present you the way Ring, Amazon’s dwelling safety subsidiary, constructed a production-ready, multi-locale Retrieval-Augmented Era (RAG)-based assist chatbot utilizing Amazon Bedrock Information Bases. By eliminating per-Area infrastructure deployments, Ring lowered the price of scaling to every further locale by 21%. On the identical time, Ring maintained constant buyer experiences throughout 10 worldwide Areas.
On this put up, you’ll find out how Ring carried out metadata-driven filtering for Area-specific content material, separated content material administration into ingestion, analysis and promotion workflows, and achieved price financial savings whereas scaling up. The structure described on this put up makes use of Amazon Bedrock Information Bases, Amazon Bedrock, AWS Lambda, AWS Step Features, and Amazon Easy Storage Service (Amazon S3). Whether or not you’re increasing assist operations internationally or trying to optimize your current RAG structure, this implementation gives sensible patterns you may apply to your personal multi-locale assist programs.
The assist evolution journey for Ring
Buyer assist at Ring initially relied on a rule-based chatbot constructed with Amazon Lex. Whereas purposeful, the system had limitations with predefined dialog patterns that couldn’t deal with the varied vary of buyer inquiries. Throughout peak intervals, 16% of interactions escalated to human brokers, and assist engineers spent 10% of their time sustaining the rule-based system. As Ring expanded throughout worldwide locales, this strategy grew to become unsustainable.
Necessities for a RAG-based assist system
Ring confronted a problem: easy methods to present correct, contextually related assist throughout a number of worldwide locales with out creating separate infrastructure for every Area. The workforce recognized 4 necessities that will inform their architectural strategy.
- International content material localization
The worldwide presence of Ring required greater than translation. Every territory wanted Area-specific product info, from voltage specs to regulatory compliance particulars, supplied by way of a unified system. Throughout the UK, Germany, and eight different locales, Ring wanted to deal with distinct product configurations and assist eventualities for every Area.
- Serverless, managed structure
Ring wished their engineering workforce targeted on enhancing buyer expertise, not managing infrastructure. The workforce wanted a completely managed, serverless resolution.
- Scalable information administration
With a whole bunch of product guides, troubleshooting paperwork, and assist articles continually being up to date, Ring wanted vector search know-how that would retrieve exact info from a unified repository. The system needed to assist automated content material ingestion pipelines in order that the Ring content material workforce might publish updates that will develop into obtainable throughout a number of locales with out guide intervention.
- Efficiency and value optimization
The common end-to-end latency requirement for Ring was 7–8 seconds and efficiency evaluation revealed that cross-Area latency accounted for lower than 10% of complete response time. This discovering allowed Ring to undertake a centralized structure reasonably than deploying separate infrastructure in every Area, which lowered operational complexity and prices.
To deal with these necessities, Ring carried out metadata-driven filtering with content material locale tags. This strategy serves Area-specific content material from a single centralized system. For his or her serverless necessities, Ring selected Amazon Bedrock Information Bases and Lambda, which eliminated the necessity for infrastructure administration whereas offering automated scaling.
Overview of resolution
Ring designed their RAG-based chatbot structure to separate content material administration into two core processes: Ingestion & Analysis and Promotion. This two-phase strategy permits Ring to take care of steady content material enchancment whereas retaining manufacturing programs steady.
Ingestion and analysis workflow
Determine 1: Structure diagram exhibiting the Ring ingestion and analysis workflow with Step Features orchestrating each day information base creation, analysis, and high quality validation utilizing Information Bases and S3 storage.
- Content material add – The Ring content material workforce uploads assist documentation, troubleshooting guides, and product info to Amazon S3. The workforce structured the S3 objects with content material in encoded format and metadata attributes. For instance, a file for the content material “Steps to Exchange the doorbell battery” has the next construction:
{
“properties”: {
“slug”: “abcde”,
“contentLocale”: “en-GB”, # distinctive identifier
“sourceFormat”: “md”, # locale info
“metadataAttributes”: {
“group”: “Service”,
“slug”: “abcde”,
“contentLocale”: “en-GB”
},
“content material”: “U3RlcHMgdG8gUmVwbGFjZSB0aGUgZG9vcmJlbGwgYmF0dGVyeTo=
VXNlIHRoZSBpbmNsdWRlZCBzZWN1cml0eSBzY3Jld2RyaXZlciB0byByZW1vdmUgd
GhlIHNlY3VyaXR5IHNjcmV3IGxvY2F0ZWQgb24gdGhlIGJvdHRvbSBvZiB0aGUgZm
FjZXBsYXRlCgpSZW1vdmUgdGhlIGZhY2VwbGF0ZSBieSBwcmVzc2luZyBpbiBvbiB
0aGUgc2lkZXMgYW5kIGNhcmVmdWxseSBwdWxsaW5nIGl0IG91dCBhbmQgb2ZmCgpS
ZW1vdmUgdGhlIGJhdHRlcnkgZnJvbSB0aGUgZG9vcmJlbGwKCkNvbm5lY3QgdGhlI
GNoYXJnaW5nIGNhYmxlIHRvIHRoZSBiYXR0ZXJ5J3MgY2hhcmdpbmcgcG9ydAoKQ2h
hcmdlIHVudGlsIG9ubHkgdGhlIGdyZWVuIGxpZ2h0IHJlbWFpbnMgbGl0ICh3aGlsZ
SBjaGFyZ2luZywgeW91J2xsIHNlZSBib3RoIGEgc29saWQgZ3JlZW4gYW5kIGFtYmV
yIGxpZ2h0KQoKUmUtaW5zZXJ0IHRoZSBjaGFyZ2VkIGJhdHRlcnkgaW50byB0aGUgZ
G9vcmJlbGwKCkRlLWF0dGFjaCB0aGUgZmFjZXBsYXRlCgpTZWN1cmUgd2l0aCB0aGU
gc2VjdXJpdHkgc2NyZXc= # base64 encoded
}
}
- Content material processing – Ring configured Amazon S3 bucket occasion notifications with Lambda because the goal to mechanically course of uploaded content material.
- Uncooked and processed content material storage
The Lambda perform performs two key operations:- Copies the uncooked information to the Information Base Archive Bucket
- Extracts metadata and content material from uncooked information, storing them as separate recordsdata within the Information Base Supply Bucket with contentLocale classification (for instance, {locale}/Service.Ring.{Upsert/Delete}.{unique_identifier}.json)
For the doorbell battery instance, the Ring metadata and content material recordsdata have the next construction:
{locale}/Service.Ring.{Upsert/Delete}.{unique_identifier}.metadata.json
{
“metadataAttributes” : {
“group”: “Service”,
“slug”: “abcde”,
“contentLocale”: “en-GB”
}
}
{locale}/Service.Ring.{Upsert/Delete}.{unique_identifier}.json
{
“content material”: “Steps to Exchange the doorbell battery:
Use the included safety screwdriver to take away the safety screw situated on the underside of the faceplate
Take away the faceplate by urgent in on the edges and thoroughly pulling it out and off
Take away the battery from the doorbell
Join the charging cable to the battery’s charging port
Cost till solely the inexperienced gentle stays lit (whereas charging, you may see each a strong inexperienced and amber gentle)
Re-insert the charged battery into the doorbell
Re-attach the faceplate
Safe with the safety screw
}
- Each day Knowledge Copy and Information Base Creation
Ring makes use of AWS Step Features to orchestrate their each day workflow that:
- Copies content material and metadata from the Information Base Supply Bucket to Knowledge Supply (Model)
- Creates a brand new Information Base (Model) by indexing the each day bucket as information supply for vector embedding
Every model maintains a separate Information Base, giving Ring unbiased analysis capabilities and simple rollback choices.
- Each day Analysis Course of
The AWS Step Features workflow continues utilizing analysis datasets to:
- Run queries throughout Information Base variations
- Check retrieval accuracy and response high quality to check efficiency between variations
- Publish efficiency metrics to Tableau dashboards with outcomes organized by contentLocale
- High quality Validation and Golden Dataset Creation
Ring makes use of the Anthropic Claude Sonnet 4 massive language mannequin (LLM)-as-a-judge to:
- Consider metrics throughout Information Base variations to establish the best-performing model
- Examine retrieval accuracy, response high quality, and efficiency metrics organized by contentLocale
- Promote the highest-performing model to Knowledge Supply (Golden) for manufacturing use
This structure helps rollbacks to earlier variations for as much as 30 days. As a result of content material is up to date roughly 200 occasions per week, Ring determined to not preserve variations past 30 days.
Promotion workflow: customer-facing
Determine 2: Structure diagram exhibiting the Ring manufacturing chatbot system the place buyer queries circulate by way of AWS Lambda to retrieve context from Information Bases and generate responses utilizing basis fashions
- Buyer interplay – Prospects provoke assist queries by way of the chatbot interface. For instance, a buyer question for the battery alternative state of affairs seems like this:
{
“textual content”: “How can I change the doorbell battery?”,
“market”: “en-GB”
}
- Question orchestration and information retrieval
Ring configured Lambda to course of buyer queries and retrieve related content material from Amazon Bedrock Information Bases. The perform:
- Transforms incoming queries for the RAG system
- Applies metadata filtering with contentLocale tags utilizing equals operator for exact Regional content material focusing on
- Queries the validated Golden Knowledge Supply to retrieve contextually related content material
Right here’s the pattern code Ring makes use of in AWS Lambda:
## Metadata Filtering for Regional Content material Focusing on
num_results = 10
market = “en-GB”
knowledge_base_id = “A2BCDEFGHI”
user_text = “How can I change the doorbell battery?”
# Configure Regional content material filtering
vector_search_config = {“numberOfResults”: num_results}
vector_search_config[“filter”] = {
“equals”: {
“key”: “contentLocale”,
“worth”: market
}
}
# Run Amazon Bedrock Information Base search
response = boto3.consumer(“bedrock-agent-runtime”).retrieve(
knowledgeBaseId=knowledge_base_id,
retrievalQuery={“textual content”: user_text},
retrievalConfiguration={
“vectorSearchConfiguration”: vector_search_config,
},
)
- Response technology
Within the Lambda perform, the system:
- Types the retrieved content material primarily based on relevance rating and selects the highest-scoring context
- Combines the top-ranked context with the unique buyer question to create an augmented immediate
- Sends the augmented immediate to LLM on Amazon Bedrock
- Configures locale-specific prompts for every contentLocale
- Generates contextually related responses returned by way of the chatbot interface
Different issues in your implementation
When constructing your personal RAG-based system at scale, think about these architectural approaches and operational necessities past the core implementation:
Vector retailer choice
The Ring implementation makes use of Amazon OpenSearch Serverless because the vector retailer for his or her information bases. Nevertheless, Amazon Bedrock Information Bases additionally helps Amazon S3 Vectors as a vector retailer possibility. When selecting between these choices, think about:
- Amazon OpenSearch Serverless: Offers superior search capabilities, real-time indexing, and versatile querying choices. Greatest fitted to purposes requiring advanced search patterns or while you want further OpenSearch options past vector search.
- Amazon S3 vectors: Affords a cheaper possibility for simple vector search use instances. S3 vector shops present automated scaling, built-in sturdiness, and may be extra economical for large-scale deployments with predictable entry patterns.
Along with these two choices, AWS helps integrations with different information retailer choices, together with Amazon Kendra, Amazon Neptune Analytics, and Amazon Aurora PostgreSQL. Consider your particular necessities round question complexity, price optimization, and operational wants when deciding on your vector retailer. The prescriptive steerage gives a very good place to begin to judge vector shops in your RAG use case.
Versioning structure issues
Whereas Ring carried out separate Information Bases for every model, you would possibly think about an alternate strategy involving separate information sources for every model inside a single information base. This methodology leverages the x-amz-bedrock-kb-data-source-id filter parameter to focus on particular information sources throughout retrieval:
vector_search_config[“filter”] = {
“equals”: {
“key”: “x-amz-bedrock-kb-data-source-id”,
“worth”: ”
}
}
# Execute Bedrock Information Base search
response = boto3.consumer(“bedrock-agent-runtime”).retrieve(
knowledgeBaseId=knowledge_base_id,
retrievalQuery={“textual content”: user_text},
retrievalConfiguration={
“vectorSearchConfiguration”: vector_search_config,
},
)
When selecting between these approaches, weigh these particular trade-offs:
- Separate information bases per model (the strategy that Ring makes use of): Offers information supply administration and cleaner rollback capabilities, however requires managing extra information base situations.
- Single information base with a number of information sources: Reduces the variety of information base situations to take care of, however introduces complexity in information supply routing logic and filtering mechanisms, plus requires sustaining separate information shops for every information supply ID.
Catastrophe restoration: Multi-Area deployment
Think about your catastrophe restoration necessities when designing your RAG structure. Amazon Bedrock Information Bases are Regional assets. To attain strong catastrophe restoration, deploy your full structure throughout a number of Areas:
- Information bases: Create Information Base situations in a number of Areas
- Amazon S3 buckets: Keep cross-Area copies of your Golden Knowledge Supply
- Lambda capabilities and Step Features workflows: Deploy your orchestration logic in every Area
- Knowledge synchronization: Implement processes to maintain content material synchronized throughout Areas
The centralized structure serves its site visitors from a single Area, prioritizing price optimization over multi-region deployment. Consider your personal Restoration Time Goal (RTO) and Restoration Level Goal (RPO) necessities to find out whether or not a multi-Area deployment is important in your use case.
Basis mannequin throughput: Cross-Area inference
Amazon Bedrock basis fashions are Regional assets with Regional quotas. To deal with site visitors bursts and scale past single-Area quotas, Amazon Bedrock helps cross-Area inference (CRIS). CRIS mechanically routes inference requests throughout a number of AWS Areas to extend throughput:
CRIS: Routes requests solely inside particular geographic boundaries (comparable to inside the US or inside the EU) to satisfy information residency necessities. This may present as much as double the default in-Area quotas.
International CRIS: Routes requests throughout a number of industrial Areas worldwide, optimizing obtainable assets and offering increased mannequin throughput past geographic CRIS capabilities. International CRIS mechanically selects the optimum Area to course of every request.
CRIS operates independently out of your Information Base deployment technique. Even with a single-Area Information Base deployment, you may configure CRIS to scale your basis mannequin throughput throughout site visitors bursts. Word that CRIS applies solely to the inference layer—your Information Bases, S3 buckets, and orchestration logic stay Regional assets that require separate multi-Area deployment for catastrophe restoration.
Embedding mannequin choice and chunking technique
Choosing the suitable embedding mannequin and chunking technique is vital for RAG system efficiency as a result of it straight impacts retrieval accuracy and response high quality. Ring makes use of the Amazon Titan Embeddings mannequin with the default chunking technique, which proved efficient for his or her assist documentation.
Amazon Bedrock provides flexibility with a number of choices:
Embedding fashions:
- Amazon Titan embeddings: Optimized for text-based content material
- Amazon Nova multimodal embeddings: Helps “Textual content”, “Picture”, “Audio”, and “Video” modalities
Chunking methods:
When ingesting information, Amazon Bedrock splits paperwork into manageable chunks for environment friendly retrieval utilizing 4 methods:
- Commonplace chunking: Mounted-size chunks for uniform paperwork
- Hierarchical chunking: For structured paperwork with clear part hierarchies
- Semantic chunking: Splits content material primarily based on subject boundaries
- Multimodal content material chunking: For paperwork with blended content material varieties (textual content, photographs, tables)
Consider your content material traits to pick the optimum mixture in your particular use case.
Conclusion
On this put up, we confirmed how Ring constructed a production-ready, multi-locale RAG-based assist chatbot utilizing Amazon Bedrock Information Bases. The structure combines automated content material ingestion, systematic each day analysis utilizing an LLM-as-judge strategy, and metadata-driven content material focusing on to realize a 21% discount in infrastructure and operational price per further locale, whereas sustaining constant buyer experiences throughout 10 worldwide Areas.
Past the core RAG structure, we coated key design issues for manufacturing deployments: vector retailer choice, versioning methods, multi-Area deployment for catastrophe restoration, Cross-Area Inference for scaling basis mannequin throughput, embedding mannequin choice and chunking methods. These patterns apply broadly to any workforce constructing multi-locale or high-availability RAG programs on AWS.Ring continues to evolve their chatbot structure towards an agentic mannequin with dynamic agent choice and integration of a number of specialised brokers. This agentic strategy will permit Ring to route buyer inquiries to specialised brokers for machine troubleshooting, order administration, and product suggestions, demonstrating the extensibility of RAG-based assist programs constructed on Amazon Bedrock.
To be taught extra about Amazon Bedrock Information Bases, go to the Amazon Bedrock documentation.
Concerning the authors
Gopinath Jagadesan
Gopinath Jagadesan is a Senior Answer Architect at AWS, the place he works with Amazon to design, construct, and deploy well-architected options on AWS. He holds a grasp’s diploma in electrical and laptop engineering from the College of Illinois at Chicago. Gopinath is captivated with generative AI and its real-world purposes, serving to clients harness its potential to drive innovation and effectivity. Outdoors of labor, he enjoys enjoying soccer and spending time along with his household and pals.
David Kim
David Kim is a Software program Improvement Engineer at Ring, the place he designs and builds AI brokers to automate customer support experiences. He’s captivated with conversational AI and multi-agent programs, leveraging AWS Bedrock to create clever, scalable options. David additionally has a deep curiosity in quantum mechanics, exploring its potential intersections with computing. Outdoors of labor, he enjoys gaming, bouldering, watching TV reveals, and touring along with his household.
Premjit Singh
Premjit Singh is a Software program Improvement Supervisor with the Ring eCommerce platform at Ring. She focuses on enabling Ring clients to find and buy Ring merchandise on ring.com. She is captivated with leveraging AWS AI service choices, together with Amazon Bedrock, to construct brokers and exploring Kiro’s spec-driven growth paradigm. In her spare time, she enjoys watching TV reveals.

