How Ring scales global customer support with Amazon Bedrock Knowledge Bases

In this post, you'll learn how Ring implemented metadata-driven filtering for Region-specific content, separated content management into ingestion, evaluation and promotion workflows, and achieved cost savings while scaling up.

Mar 30, 2026 - 18:00
How Ring scales global customer support with Amazon Bedrock Knowledge Bases

This post is cowritten with David Kim, and Premjit Singh from Ring.

Scaling self-service support globally presents challenges beyond translation. In this post, we show you how Ring, Amazon’s home security subsidiary, built a production-ready, multi-locale Retrieval-Augmented Generation (RAG)-based support chatbot using Amazon Bedrock Knowledge Bases. By eliminating per-Region infrastructure deployments, Ring reduced the cost of scaling to each additional locale by 21%. At the same time, Ring maintained consistent customer experiences across 10 international Regions.

In this post, you’ll learn how Ring implemented metadata-driven filtering for Region-specific content, separated content management into ingestion, evaluation and promotion workflows, and achieved cost savings while scaling up. The architecture described in this post uses Amazon Bedrock Knowledge Bases, Amazon Bedrock, AWS Lambda, AWS Step Functions, and Amazon Simple Storage Service (Amazon S3). Whether you’re expanding support operations internationally or looking to optimize your existing RAG architecture, this implementation provides practical patterns you can apply to your own multi-locale support systems.

The support evolution journey for Ring

Customer support at Ring initially relied on a rule-based chatbot built with Amazon Lex. While functional, the system had limitations with predefined conversation patterns that couldn’t handle the diverse range of customer inquiries. During peak periods, 16% of interactions escalated to human agents, and support engineers spent 10% of their time maintaining the rule-based system. As Ring expanded across international locales, this approach became unsustainable.

Requirements for a RAG-based support system

Ring faced a challenge: how to provide accurate, contextually relevant support across multiple international locales without creating separate infrastructure for each Region. The team identified four requirements that would inform their architectural approach.

  1. Global content localization

The international presence of Ring required more than translation. Each territory needed Region-specific product information, from voltage specifications to regulatory compliance details, provided through a unified system. Across the UK, Germany, and eight other locales, Ring needed to handle distinct product configurations and support scenarios for each Region.

  1. Serverless, managed architecture

Ring wanted their engineering team focused on improving customer experience, not managing infrastructure. The team needed a fully managed, serverless solution.

  1. Scalable knowledge management

With hundreds of product guides, troubleshooting documents, and support articles constantly being updated, Ring needed vector search technology that could retrieve precise information from a unified repository. The system had to support automated content ingestion pipelines so that the Ring content team could publish updates that would become available across multiple locales without manual intervention.

  1. Performance and cost optimization

The average end-to-end latency requirement for Ring was 7–8 seconds and performance analysis revealed that cross-Region latency accounted for less than 10% of total response time. This finding allowed Ring to adopt a centralized architecture rather than deploying separate infrastructure in each Region, which reduced operational complexity and costs.

To address these requirements, Ring implemented metadata-driven filtering with content locale tags. This approach serves Region-specific content from a single centralized system. For their serverless requirements, Ring chose Amazon Bedrock Knowledge Bases and Lambda, which removed the need for infrastructure management while providing automatic scaling.

Overview of solution

Ring designed their RAG-based chatbot architecture to separate content management into two core processes: Ingestion & Evaluation and Promotion. This two-phase approach allows Ring to maintain continuous content improvement while keeping production systems stable.

Ingestion and evaluation workflow

Architecture diagram showing the Ring data pipeline from content team through S3 buckets, Lambda processing, and Amazon Bedrock Knowledge Bases. The numbered flow (1-8) shows content ingestion, metadata extraction, daily evaluation through Step Functions, and quality validation before promoting to Golden Data Source.

Figure 1: Architecture diagram showing the Ring ingestion and evaluation workflow with Step Functions orchestrating daily knowledge base creation, evaluation, and quality validation using Knowledge Bases and S3 storage.

  1. Content upload – The Ring content team uploads support documentation, troubleshooting guides, and product information to Amazon S3. The team structured the S3 objects with content in encoded format and metadata attributes. For example, a file for the content “Steps to Replace the doorbell battery” has the following structure:
{
	"properties": {
		"slug": "abcde",
		"contentLocale": "en-GB",  # unique identifier
		"sourceFormat": "md",      # locale information
		"metadataAttributes": {
			"group": "Service",
			"slug": "abcde",
			"contentLocale": "en-GB"
		},
		"content": "U3RlcHMgdG8gUmVwbGFjZSB0aGUgZG9vcmJlbGwgYmF0dGVyeTo= 
                VXNlIHRoZSBpbmNsdWRlZCBzZWN1cml0eSBzY3Jld2RyaXZlciB0byByZW1vdmUgd
                GhlIHNlY3VyaXR5IHNjcmV3IGxvY2F0ZWQgb24gdGhlIGJvdHRvbSBvZiB0aGUgZm
                FjZXBsYXRlCgpSZW1vdmUgdGhlIGZhY2VwbGF0ZSBieSBwcmVzc2luZyBpbiBvbiB
                0aGUgc2lkZXMgYW5kIGNhcmVmdWxseSBwdWxsaW5nIGl0IG91dCBhbmQgb2ZmCgpS
                ZW1vdmUgdGhlIGJhdHRlcnkgZnJvbSB0aGUgZG9vcmJlbGwKCkNvbm5lY3QgdGhlI
                GNoYXJnaW5nIGNhYmxlIHRvIHRoZSBiYXR0ZXJ5J3MgY2hhcmdpbmcgcG9ydAoKQ2h
                hcmdlIHVudGlsIG9ubHkgdGhlIGdyZWVuIGxpZ2h0IHJlbWFpbnMgbGl0ICh3aGlsZ
                SBjaGFyZ2luZywgeW91J2xsIHNlZSBib3RoIGEgc29saWQgZ3JlZW4gYW5kIGFtYmV
                yIGxpZ2h0KQoKUmUtaW5zZXJ0IHRoZSBjaGFyZ2VkIGJhdHRlcnkgaW50byB0aGUgZ
                G9vcmJlbGwKCkRlLWF0dGFjaCB0aGUgZmFjZXBsYXRlCgpTZWN1cmUgd2l0aCB0aGU
                gc2VjdXJpdHkgc2NyZXc=    # base64 encoded 
	}
}
  1. Content processing – Ring configured Amazon S3 bucket event notifications with Lambda as the target to automatically process uploaded content.
  1. Raw and processed content storage
    The Lambda function performs two key operations:

    • Copies the raw data to the Knowledge Base Archive Bucket
    • Extracts metadata and content from raw data, storing them as separate files in the Knowledge Base Source Bucket with contentLocale classification (for example, {locale}/Service.Ring.{Upsert/Delete}.{unique_identifier}.json)

    For the doorbell battery example, the Ring metadata and content files have the following structure:

    {locale}/Service.Ring.{Upsert/Delete}.{unique_identifier}.metadata.json

{
	"metadataAttributes" : {
		"group": "Service",
		"slug": "abcde",
		"contentLocale": "en-GB"
	}
}

{locale}/Service.Ring.{Upsert/Delete}.{unique_identifier}.json

{
	"content": "Steps to Replace the doorbell battery:
	Use the included security screwdriver to remove the security screw located on the bottom of the faceplate
	Remove the faceplate by pressing in on the sides and carefully pulling it out and off
	Remove the battery from the doorbell
	Connect the charging cable to the battery's charging port
	Charge until only the green light remains lit (while charging, you'll see both a solid green and amber light)
	Re-insert the charged battery into the doorbell
	Re-attach the faceplate
	Secure with the security screw
}
  1. Daily Data Copy and Knowledge Base Creation

Ring uses AWS Step Functions to orchestrate their daily workflow that:

  • Copies content and metadata from the Knowledge Base Source Bucket to Data Source (Version)
  • Creates a new Knowledge Base (Version) by indexing the daily bucket as data source for vector embedding

Each version maintains a separate Knowledge Base, giving Ring independent evaluation capabilities and straightforward rollback options.

  1. Daily Evaluation Process

The AWS Step Functions workflow continues using evaluation datasets to:

  • Run queries across Knowledge Base versions
  • Test retrieval accuracy and response quality to compare performance between versions
  • Publish performance metrics to Tableau dashboards with results organized by contentLocale
  1. Quality Validation and Golden Dataset Creation

Ring uses the Anthropic Claude Sonnet 4 large language model (LLM)-as-a-judge to:

  • Evaluate metrics across Knowledge Base versions to identify the best-performing version
  • Compare retrieval accuracy, response quality, and performance metrics organized by contentLocale
  • Promote the highest-performing version to Data Source (Golden) for production use

This architecture supports rollbacks to previous versions for up to 30 days. Because content is updated approximately 200 times per week, Ring decided not to maintain versions beyond 30 days.

Promotion workflow: customer-facing

Architecture diagram showing the Ring promotion pipeline with four-step customer interaction flow (1-4) from chatbot through AWS Lambda to Knowledge Bases retrieval and response generation using foundation models.

Figure 2: Architecture diagram showing the Ring production chatbot system where customer queries flow through AWS Lambda to retrieve context from Knowledge Bases and generate responses using foundation models

  1. Customer interaction – Customers initiate support queries through the chatbot interface. For example, a customer query for the battery replacement scenario looks like this:
{
	"text": "How can I replace the doorbell battery?",
	"market": "en-GB"
}
  1. Query orchestration and knowledge retrieval

Ring configured Lambda to process customer queries and retrieve relevant content from Amazon Bedrock Knowledge Bases. The function:

  • Transforms incoming queries for the RAG system
  • Applies metadata filtering with contentLocale tags using equals operator for precise Regional content targeting
  • Queries the validated Golden Data Source to retrieve contextually relevant content

Here’s the sample code Ring uses in AWS Lambda:

## Metadata Filtering for Regional Content Targeting

num_results = 10
market = "en-GB"
knowledge_base_id = "A2BCDEFGHI"
user_text = "How can I replace the doorbell battery?"

# Configure Regional content filtering
vector_search_config = {"numberOfResults": num_results}
vector_search_config["filter"] = {
	"equals": {
		"key": "contentLocale",
		"value": market
	}
}

# Run Amazon Bedrock Knowledge Base search
response = boto3.client("bedrock-agent-runtime").retrieve(
	knowledgeBaseId=knowledge_base_id,
	retrievalQuery={"text": user_text},
	retrievalConfiguration={
		"vectorSearchConfiguration": vector_search_config,
	},
)
  1. Response generation

In the Lambda function, the system:

  • Sorts the retrieved content based on relevance score and selects the highest-scoring context
  • Combines the top-ranked context with the original customer query to create an augmented prompt
  • Sends the augmented prompt to LLM on Amazon Bedrock
  • Configures locale-specific prompts for each contentLocale
  • Generates contextually relevant responses returned through the chatbot interface

Other considerations for your implementation

When building your own RAG-based system at scale, consider these architectural approaches and operational requirements beyond the core implementation:

Vector store selection

The Ring implementation uses Amazon OpenSearch Serverless as the vector store for their knowledge bases. However, Amazon Bedrock Knowledge Bases also supports Amazon S3 Vectors as a vector store option. When choosing between these options, consider:

  • Amazon OpenSearch Serverless: Provides advanced search capabilities, real-time indexing, and flexible querying options. Best suited for applications requiring complex search patterns or when you need additional OpenSearch features beyond vector search.
  • Amazon S3 vectors: Offers a more cost-effective option for straightforward vector search use cases. S3 vector stores provide automatic scaling, built-in durability, and can be more economical for large-scale deployments with predictable access patterns.

In addition to these two options, AWS supports integrations with other data store options, including Amazon Kendra, Amazon Neptune Analytics, and Amazon Aurora PostgreSQL. Evaluate your specific requirements around query complexity, cost optimization, and operational needs when selecting your vector store. The prescriptive guidance provides a good starting point to evaluate vector stores for your RAG use case.

Versioning architecture considerations

While Ring implemented separate Knowledge Bases for each version, you might consider an alternative approach involving separate data sources for each version within a single knowledge base. This method leverages the x-amz-bedrock-kb-data-source-id filter parameter to target specific data sources during retrieval:

vector_search_config["filter"] = {
	"equals": {
		"key": "x-amz-bedrock-kb-data-source-id",
		"value": ''
		}
	}
# Execute Bedrock Knowledge Base search
response = boto3.client("bedrock-agent-runtime").retrieve(
	knowledgeBaseId=knowledge_base_id,
	retrievalQuery={"text": user_text},
	retrievalConfiguration={
		"vectorSearchConfiguration": vector_search_config,
	},
)

When choosing between these approaches, weigh these specific trade-offs:

  • Separate knowledge bases per version (the approach that Ring uses): Provides data source management and cleaner rollback capabilities, but requires managing more knowledge base instances.
  • Single knowledge base with multiple data sources: Reduces the number of knowledge base instances to maintain, but introduces complexity in data source routing logic and filtering mechanisms, plus requires maintaining separate data stores for each data source ID.

Disaster recovery: Multi-Region deployment

Consider your disaster recovery requirements when designing your RAG architecture. Amazon Bedrock Knowledge Bases are Regional resources. To achieve robust disaster recovery, deploy your complete architecture across multiple Regions:

  • Knowledge bases: Create Knowledge Base instances in multiple Regions
  • Amazon S3 buckets: Maintain cross-Region copies of your Golden Data Source
  • Lambda functions and Step Functions workflows: Deploy your orchestration logic in each Region
  • Data synchronization: Implement processes to keep content synchronized across Regions

The centralized architecture serves its traffic from a single Region, prioritizing cost optimization over multi-region deployment. Evaluate your own Recovery Time Objective (RTO) and Recovery Point Objective (RPO) requirements to determine whether a multi-Region deployment is necessary for your use case.

Foundation model throughput: Cross-Region inference

Amazon Bedrock foundation models are Regional resources with Regional quotas. To handle traffic bursts and scale beyond single-Region quotas, Amazon Bedrock supports cross-Region inference (CRIS). CRIS automatically routes inference requests across multiple AWS Regions to increase throughput:

CRIS: Routes requests only within specific geographic boundaries (such as within the US or within the EU) to meet data residency requirements. This can provide up to double the default in-Region quotas.

Global CRIS: Routes requests across multiple commercial Regions worldwide, optimizing available resources and providing higher model throughput beyond geographic CRIS capabilities. Global CRIS automatically selects the optimal Region to process each request.

CRIS operates independently from your Knowledge Base deployment strategy. Even with a single-Region Knowledge Base deployment, you can configure CRIS to scale your foundation model throughput during traffic bursts. Note that CRIS applies only to the inference layer—your Knowledge Bases, S3 buckets, and orchestration logic remain Regional resources that require separate multi-Region deployment for disaster recovery.

Embedding model selection and chunking strategy

Selecting the appropriate embedding model and chunking strategy is important for RAG system performance because it directly affects retrieval accuracy and response quality. Ring uses the Amazon Titan Embeddings model with the default chunking strategy, which proved effective for their support documentation.

Amazon Bedrock offers flexibility with multiple options:

Embedding models:

  • Amazon Titan embeddings: Optimized for text-based content
  • Amazon Nova multimodal embeddings: Supports “Text”, “Image”, “Audio”, and “Video” modalities

Chunking strategies:

When ingesting data, Amazon Bedrock splits documents into manageable chunks for efficient retrieval using four strategies:

  • Standard chunking: Fixed-size chunks for uniform documents
  • Hierarchical chunking: For structured documents with clear section hierarchies
  • Semantic chunking: Splits content based on topic boundaries
  • Multimodal content chunking: For documents with mixed content types (text, images, tables)

Evaluate your content characteristics to select the optimal combination for your specific use case.

Conclusion

In this post, we showed how Ring built a production-ready, multi-locale RAG-based support chatbot using Amazon Bedrock Knowledge Bases. The architecture combines automated content ingestion, systematic daily evaluation using an LLM-as-judge approach, and metadata-driven content targeting to achieve a 21% reduction in infrastructure and operational cost per additional locale, while maintaining consistent customer experiences across 10 international Regions.

Beyond the core RAG architecture, we covered key design considerations for production deployments: vector store selection, versioning strategies, multi-Region deployment for disaster recovery, Cross-Region Inference for scaling foundation model throughput, embedding model selection and chunking strategies. These patterns apply broadly to any team building multi-locale or high-availability RAG systems on AWS.Ring continues to evolve their chatbot architecture toward an agentic model with dynamic agent selection and integration of multiple specialized agents. This agentic approach will allow Ring to route customer inquiries to specialized agents for device troubleshooting, order management, and product recommendations, demonstrating the extensibility of RAG-based support systems built on Amazon Bedrock.

To learn more about Amazon Bedrock Knowledge Bases, visit the Amazon Bedrock documentation.


About the authors

Gopinath Jagadesan

Gopinath Jagadesan

Gopinath Jagadesan is a Senior Solution Architect at AWS, where he works with Amazon to design, build, and deploy well-architected solutions on AWS. He holds a master’s degree in electrical and computer engineering from the University of Illinois at Chicago. Gopinath is passionate about generative AI and its real-world applications, helping customers harness its potential to drive innovation and efficiency. Outside of work, he enjoys playing soccer and spending time with his family and friends.

David Kim

David Kim

David Kim is a Software Development Engineer at Ring, where he designs and builds AI agents to automate customer service experiences. He is passionate about conversational AI and multi-agent systems, leveraging AWS Bedrock to create intelligent, scalable solutions. David also has a deep interest in quantum mechanics, exploring its potential intersections with computing. Outside of work, he enjoys gaming, bouldering, watching TV shows, and traveling with his family.

Premjit Singh

Premjit Singh

Premjit Singh is a Software Development Manager with the Ring eCommerce platform at Ring. She focuses on enabling Ring customers to discover and purchase Ring products on ring.com. She is passionate about leveraging AWS AI service offerings, including Amazon Bedrock, to build agents and exploring Kiro’s spec-driven development paradigm. In her spare time, she enjoys watching TV shows.

Jat AI Stay informed with the latest in artificial intelligence. Jat AI News Portal is your go-to source for AI trends, breakthroughs, and industry analysis. Connect with the community of technologists and business professionals shaping the future.