This put up is cowritten by Shawn Tsai from TrendMicro.
Delivering related, context-aware responses is essential for buyer satisfaction. For enterprise-grade AI chatbots, understanding not solely the present question but additionally the organizational context behind it’s key. Firm-wise reminiscence in Amazon Bedrock, powered by Amazon Neptune and Mem0, gives AI brokers with persistent, company-specific context—enabling them to be taught, adapt, and reply intelligently throughout a number of interactions. TrendMicro, one of many largest antivirus software program firms on the planet, developed the Pattern’s Companion chatbot, so their clients can discover data via pure, conversational interactions (be taught extra).
TrendMicro aimed to reinforce its AI chatbot service to ship personalised, context-aware assist for enterprise clients. The chatbot wanted to retain dialog historical past for continuity, reference company-specific data at scale, and make sure that reminiscence remained correct, safe, and updated. The problem is in integrating long-term reminiscence for organizational data with short-term reminiscence for ongoing conversations, whereas supporting company-wide data sharing. In collaboration with the AWS crew, together with AWS’s Generative AI Innovation Heart, TrendMicro addressed this problem utilizing Amazon Neptune, Amazon OpenSearch, and Amazon Bedrock, as we elaborate on this weblog.
Answer overview
TrendMicro applied company-wise reminiscence in Amazon Bedrock by combining a number of AWS providers. Amazon Neptune shops a company-specific data graph, representing organizational relationships, processes, and knowledge to allow exact and structured retrieval. Mem0 manages short-term conversational reminiscence for fast context and long-term reminiscence for persistent data throughout classes. Amazon Bedrock orchestrates the AI agent workflows, integrating with each Neptune and Mem0 to retrieve and apply contextual data throughout inference. This structure permits the chatbot to recall related historical past, retrieve structured firm data, and reply with tailor-made, context-rich solutions—serving to considerably enhance person expertise.
Reminiscence creation and replace
The structure begins with capturing person messages and extracting entities, relationships, and potential recollections via the Claude mannequin on Amazon Bedrock. These are then embedded with Amazon Bedrock Titan Textual content Embed and searched towards each Amazon OpenSearch Service and Amazon Neptune. Related entities and recollections are retrieved, and up to date via the mannequin earlier than being re-embedded and listed again into OpenSearch and Neptune. This closed-loop course of makes certain that entity-related recollections may be constantly refreshed and the data graph in Neptune stays in step with conversational insights.
Reminiscence retrieval
When dealing with person queries, the system applies the same embedding pipeline with Bedrock Titan to go looking throughout each vector embeddings in OpenSearch Service and entity triples in Neptune. The related recollections are then reranked utilizing Amazon Bedrock Rerank or Cohere Rerank fashions to guarantee that essentially the most contextually correct data is delivered. This twin retrieval technique gives each semantic flexibility from OpenSearch and structured precision from Neptune, enabling the chatbot to ship extremely related, context-aware solutions.
Response-memory mapping and human-in-the-loop suggestions
For every AI response, the system maps sentences to the particular recollections referenced, producing a reminiscence evaluation report. Customers are then offered with the chance to approve or reject these mappings. Accepted recollections stay a part of the data base, whereas rejected ones are faraway from each OpenSearch Service and Neptune. This makes certain that solely validated and trusted data persists. This human-in-the-loop mechanism strengthens belief and helps constantly enhance reminiscence accuracy and offers enterprise clients direct affect over the refinement of their AI’s data.
Amazon Neptune in motion
As an instance how Amazon Neptune enriches chatbot reminiscence, contemplate a buyer asking, “Who acknowledged Kublai as ruler?” With out the data graph, the AI would possibly return a imprecise response akin to: “Kublai was a Mongol ruler who gained recognition from totally different teams.” This sort of reply is generic and lacks precision.
When the identical query is requested however the Neptune entity graph is queried and positioned into the massive language mannequin’s (LLM) context window, the mannequin can floor its reasoning in structured triples like (Ilkhans, acknowledged, Kublai). The chatbot can then reply extra precisely: “In accordance with the organizational data base, Kublai was acknowledged by the Ilkhans as ruler.” This before-and-after instance demonstrates how structured entity relationships in Neptune permit the mannequin to supply solutions which can be each contextually related and verifiable.
Conclusion and subsequent step
As described within the AWS Pattern Micro case research, Pattern Micro makes use of AWS to assist ship safer, scalable, and clever buyer experiences. Constructing on this basis, Pattern Micro combines Amazon Bedrock, Amazon Neptune, Amazon OpenSearch Service, and Mem0 to create an AI chatbot with persistent, organization-specific reminiscence that delivers clever, context-aware conversations at scale. By integrating graph-based data with generative AI, Pattern Micro is predicted to enhance reply high quality, delivering clearer and extra correct responses whereas establishing a basis for AI methods that constantly adapt to evolving organizational data; This work stays underneath analysis and tuning to additional improve the end-user expertise.
Wanting forward, TrendMicro is exploring future enhancements akin to broader graph protection, extra replace pipelines, and multilingual assist. For readers who wish to dive deeper, we suggest exploring the GitHub pattern implementation, which incorporates the supply code we applied, and the Amazon Neptune Documentation for additional technical particulars and inspiration.
Concerning the authors
Shawn Tsai
Shawn Tsai is a senior architect at Pattern Micro, specializing in massive language mannequin utility improvement and safety practices, cloud structure design, large-scale software program structure design, and DevOps practices. He’s at the moment primarily answerable for Pattern Micro’s massive language mannequin utility improvement and safety framework.
Ray Wang
Ray Wang is a Senior Options Architect at AWS. With 12+ years of expertise within the backend and advisor, Ray is devoted to constructing fashionable options within the cloud, particularly in particularly in NoSQL, massive knowledge, machine studying, and Generative AI. As a hungry go-getter, he handed all 12 AWS certificates to extend the breadth and depth of his technical data. He likes to learn and watch sci-fi motion pictures in his spare time.
Zhihao Lin
Zhihao Lin is an Utilized Scientist on the AWS Generative AI Innovation Heart. With a Grasp’s diploma from Peking College and publications in high conferences akin to CVPR and IJCAI, he brings in depth AI/ML analysis expertise to his function. At AWS, he focuses on creating generative AI options, leveraging cutting-edge expertise for revolutionary functions. He focuses on fixing complicated laptop imaginative and prescient and pure language processing challenges and advancing the sensible use of generative AI in enterprise.

