How Amazon Finance constructed an AI assistant utilizing Amazon Bedrock and Amazon Kendra to help analysts for information discovery and enterprise insights
Finance analysts throughout Amazon Finance face mounting complexity in monetary planning and evaluation processes. When working with huge datasets spanning a number of programs, information lakes, and enterprise models, analysts encounter a number of vital challenges. First, they spend important time manually searching information catalogs and reconciling information from disparate sources, leaving much less time for precious evaluation and perception technology. Second, historic information and former enterprise selections usually reside in numerous paperwork and legacy programs, making it troublesome to make use of previous learnings throughout planning cycles. Third, as enterprise contexts quickly evolve, analysts want fast entry to related metrics, planning assumptions, and monetary insights to help data-driven decision-making.
Conventional instruments and processes fall brief in addressing these challenges. Key phrase-based searches usually miss contextual relationships in monetary information, and inflexible question constructions restrict analysts’ means to discover information dynamically. Moreover, the shortage of institutional information preservation means precious insights and determination rationales usually stay siloed or get misplaced over time, resulting in redundant evaluation and inconsistent planning assumptions throughout groups. These challenges considerably affect monetary planning effectivity, decision-making agility, and the general high quality of enterprise insights. Analysts wanted a extra intuitive strategy to entry, perceive, and use their group’s collective monetary information and information property.
The Amazon Finance technical group develops and manages complete expertise options that energy monetary decision-making and operational effectivity whereas standardizing throughout Amazon’s world operations. On this publish, we clarify how the group conceptualized and applied an answer to those enterprise challenges by harnessing the ability of generative AI utilizing Amazon Bedrock and clever search with Amazon Kendra.
Answer overview
To handle these enterprise challenges, Amazon Finance developed an AI-powered assistant resolution that makes use of generative AI and enterprise search capabilities. This resolution helps analysts work together with monetary information sources and documentation by way of pure language queries, minimizing the necessity for complicated guide searches throughout a number of programs. The assistant accesses a complete information base of monetary paperwork, historic information, and enterprise context, offering related and correct responses whereas sustaining enterprise safety requirements. This method not solely streamlines information discovery but in addition preserves institutional information and permits extra constant decision-making throughout the group.
The AI assistant’s methodology consists of two key resolution parts: clever retrieval and augmented technology. The retrieval system makes use of vector shops, that are specialised databases that effectively retailer and search high-dimensional representations of textual content meanings. In contrast to conventional databases that depend on key phrase matching, vector shops allow semantic search by changing person queries into vector representations and discovering related vectors within the database. Constructing on this retrieval basis, the system employs augmented technology to create correct and contextual responses. This method enhances conventional language fashions by incorporating exterior information sources throughout response technology, considerably decreasing hallucinations and enhancing factual accuracy. The method follows three steps: retrieving related data from information sources utilizing semantic search, conditioning the language mannequin with this context, and producing refined responses that incorporate the retrieved data. By combining these applied sciences, the assistant delivers responses which can be each contextually acceptable and grounded in verified organizational information, making it significantly efficient for knowledge-intensive purposes like monetary operations and planning.
We applied this Retrieval Augmented Technology (RAG) system by way of a mixture of enormous language fashions (LLMs) on Amazon Bedrock and clever search utilizing Amazon Kendra.
Within the following sections, we focus on the important thing architectural parts that we used within the resolution and describe how the general resolution works.
Amazon Bedrock
We selected Anthropic’s Claude 3 Sonnet, a robust language mannequin, for its distinctive language technology capabilities and talent to grasp and motive complicated matters. By integrating Anthropic’s Claude into the RAG module by way of Amazon Bedrock, the AI assistant can generate contextual and informative responses that seamlessly mix the retrieved information from the vector retailer with the mannequin’s pure language processing and technology talents, leading to a extra human-like and fascinating conversational expertise.
Amazon Kendra (Enterprise Version Index)
Amazon Kendra gives highly effective pure language processing for AI assistant purposes. It excels at understanding person questions and discovering related solutions by way of semantic search. The service works easily with generative AI fashions, significantly in RAG options. The enterprise safety features in Amazon Kendra help information safety and compliance. Its means to grasp person intent and join straight with Amazon Bedrock makes it perfect for enterprise assistants. This helps create significant conversations utilizing enterprise paperwork and information catalogs.
We selected Amazon Kendra Enterprise Version Index over Amazon OpenSearch Service, primarily on account of its subtle built-in capabilities and lowered want for guide configuration. Whereas OpenSearch Service requires intensive customization and technical experience, Amazon Kendra supplies out-of-the-box pure language understanding, automated doc processing for over 40 file codecs, pre-built enterprise connectors, and clever question dealing with together with synonym recognition and refinement strategies. The service combines key phrase, semantic, and vector search approaches routinely, whereas OpenSearch Service requires guide implementation of those options. These options of Amazon Kendra had been appropriate for our finance area use case, the place accuracy is crucial for usability.
We additionally selected Amazon Kendra Enterprise Version Index over Amazon Q Business for data retrieval, as a result of it stands out as a extra strong and versatile resolution. Though each instruments purpose to streamline entry to firm data, Amazon Kendra gives superior retrieval accuracy and better management over search parameters. With Amazon Kendra, you’ll be able to fine-tune relevance tuning, customise doc attributes, and implement customized synonyms to boost search precision. This stage of customization helped us tailor the search expertise to our particular wants within the Amazon Finance area and monitor the search outcomes previous to the augmented technology step inside person conversations.
Streamlit
We chosen Streamlit, a Python-based framework for creating interactive internet purposes, for constructing the AI assistant’s UI on account of its fast improvement capabilities, seamless integration with Python and the assistant’s backend parts, interactive and responsive UI parts, potential for information visualization, and easy deployment choices. With the Streamlit UI, the assistant supplies a user-friendly and fascinating interface that facilitates pure language interactions whereas permitting for environment friendly iteration and deployment of the applying.
Immediate template
Immediate templates enable for formatting person queries, integrating retrieved information, and offering directions or constraints for response technology, that are important for producing contextual and informative responses that mix the language technology talents of Anthropic’s Claude with the related information retrieved from the search powered by Amazon Kendra. The next is an instance immediate:
Answer structure
The next resolution structure diagram depicts how the important thing architectural parts work with one another to energy the answer.

The workflow consists of the next steps:
- The person asks the query in a chat field after authentication.
- The Streamlit software sends the question to an Amazon Kendra retriever for related doc retrieval.
- Amazon Kendra sends the related paragraph and doc references to the RAG resolution.
- The RAG resolution makes use of Anthropic’s Claude in Amazon Bedrock together with the immediate template and related paragraph as context.
- The LLM response is shipped again to the Streamlit UI.
- The response is proven to the person together with the suggestions function and session historical past.
- The person suggestions on responses is saved individually in Amazon Simple Storage Service (Amazon S3)
- Amazon Kendra indexes related paperwork saved in S3 buckets for doc search and retrieval.
Frontend structure
We designed the next frontend structure to permit for fast modifications and deployment, holding in thoughts the scalability and safety of the answer.

This workflow consists of the next steps:
- The person navigates to the applying URL of their browser.
- Amazon Route 53 resolves their request to the Amazon CloudFront distribution, which then selects the server closest to the person (to attenuate latency).
- CloudFront runs an AWS Lambda operate that makes positive the person has been authenticated. If not, the person is redirected to sign up. After they efficiently sign up, they’re redirected again to the applying web site. The stream repeats, and CloudFront triggers the Lambda operate once more. This time, the person is now in a position to entry the web site.
- Now authenticated, CloudFront returns the property of the online software.
- AWS Fargate makes it attainable to run containers with out having to handle the underlying Amazon Elastic Compute Cloud (Amazon EC2) cases. This permits operating containers as a real serverless service. Amazon Elastic Container Service (Amazon ECS) is configured with automated scaling (goal monitoring automated scaling, which scales based mostly on the Application Load Balancer (ALB) requests per goal).
Analysis of the answer’s efficiency
We applied a complete analysis framework to scrupulously assess the AI assistant’s efficiency and ensure it meets the excessive requirements required for monetary purposes. Our framework was designed to seize each quantitative metrics for measurable efficiency and qualitative indicators for person expertise and response high quality. Throughout our benchmarking assessments with analysts, we discovered that this resolution dramatically lowered search time by 30% as a result of analysts can now carry out pure language search, and it improved the accuracy of search outcomes by 80%.
Quantitative evaluation
We targeted totally on precision and recall testing, creating a various check set of over 50 enterprise queries that represented typical use circumstances our analysts encounter. Utilizing human-labeled solutions as our floor fact, we evaluated the system’s efficiency throughout two major classes: information discovery and information search. In information discovery situations, the place the system helps analysts find particular information sources and metrics, we achieved an preliminary precision price of 65% and a recall price of 60% with out performing metadata enrichment on the information sources. Though these charges would possibly seem average, they symbolize a big enchancment over the earlier guide search course of, which had an estimated success price of solely 35% and sometimes required a number of iterations throughout totally different programs. The first causes for the present charges of the brand new system had been attributed to the shortage of wealthy metadata about information sources and was a very good indicator for groups to facilitate higher metadata assortment of knowledge property, which is at the moment underway.
The information search functionality demonstrated preliminary charges of 83% precision and 74% recall with out performing metadata enrichment on information sources. This marked a considerable enchancment over conventional keyword-based search strategies, which generally achieved solely 45–50% precision in our inner testing. This enchancment is especially significant as a result of it interprets to analysts discovering the best data of their first search try roughly 8 out of 10 instances, in comparison with the earlier common of three–4 makes an attempt wanted to find the identical data.
Qualitative metrics
The qualitative analysis centered across the idea of faithfulness—a vital metric for monetary purposes the place accuracy and reliability are paramount. We employed an modern LLM-as-a-judge methodology to guage how nicely the AI assistant’s responses aligned with supply documentation and averted hallucinations or unsupported assertions. The outcomes confirmed a marked distinction between use circumstances: information discovery achieved a faithfulness rating of 70%, and enterprise information search demonstrated a formidable 88% faithfulness. These scores considerably outperform our earlier documentation search system, which had no built-in verification mechanism and sometimes led to analysts working with outdated or incorrect data.
Most significantly, the brand new system lowered the common time to seek out related data from 45–60 minutes to simply 5–10 minutes—an 85% enchancment in effectivity. Person satisfaction surveys point out that 92% of analysts favor the brand new system over conventional search strategies, citing improved accuracy and time financial savings as key advantages.
These analysis outcomes haven’t solely validated our method but in addition highlighted particular areas for future enhancement. We proceed to refine our analysis framework because the system evolves, ensuring it maintains excessive requirements of accuracy and reliability whereas assembly the dynamic wants of our monetary analysts. The analysis framework was instrumental in constructing confidence inside our enterprise person neighborhood, offering clear metrics that exhibit the system’s functionality to deal with complicated monetary queries whereas sustaining the accuracy requirements important for monetary operations.
Use circumstances
Our resolution transforms how finance customers work together with complicated monetary and operational information by way of pure language queries. On this part, we focus on some key examples demonstrating how the system simplifies information discovery.
Seamless information discovery
The answer permits customers to seek out information sources by way of pure language queries reasonably than requiring technical information of database constructions. It makes use of a complicated mixture of vector shops and enterprise search capabilities to match person questions with related information sources, although cautious consideration have to be paid to context administration and stopping over-reliance on earlier interactions. Previous to the AI assistant resolution, finance analysts wanted deep technical information to navigate complicated database constructions, usually spending hours looking out by way of a number of documentation sources simply to find particular information tables. Understanding system workflows required intensive assessment of technical documentation or reaching out to subject material consultants, creating bottlenecks and decreasing productiveness. Even skilled customers struggled to piece collectively full details about enterprise processes from fragmented sources throughout totally different programs. Now, analysts can merely ask questions in pure language, similar to “The place can I discover productiveness metrics?”, “How do I entry facility data?”, or “Which dashboard exhibits operational information?” and obtain exact, contextual solutions. The answer combines enterprise search capabilities with LLMs to grasp person intent and ship related data from each structured and unstructured information sources. Analysts now obtain correct instructions to particular consolidated reporting tables, clear explanations of enterprise processes, and related technical particulars when wanted. In our benchmark assessments, for information discovery duties alone, the system achieved 70% faithfulness and 65% precision, and doc search demonstrated even stronger outcomes with 83% precision and 88% faithfulness, with out metadata enrichments.
Helping understanding of inner enterprise processes from information documentation
Monetary analysts beforehand confronted a steep studying curve when working with enterprise planning instruments. The complexity of those programs meant that even primary duties required intensive documentation assessment or ready for help from overwhelmed subject material consultants. New group members might take weeks or months to turn into proficient, whereas even skilled customers struggled to maintain up with system updates and adjustments. This created a persistent bottleneck in monetary operations and planning processes. The introduction of the AI-powered assistant has essentially modified how analysts study and work together with these planning instruments. Moderately than looking out by way of a whole bunch of pages of technical documentation, analysts can now ask simple questions like “How do I forecast depreciation for brand new property?”, “How does the quarterly planning course of work?” or “What inputs are wanted for the quarterly planning cycle?” The system supplies clear, contextualized explanations drawn from verified documentation and system specs. Our benchmark assessments revealed that it achieved 83% precision and 88% faithfulness in retrieving and explaining technical and enterprise data. New analysts can turn into productive in a matter of weeks, skilled customers can rapidly confirm procedures, and subject material consultants can concentrate on extra complicated challenges reasonably than routine questions. This represents a big development in making enterprise programs extra accessible and environment friendly, whereas sustaining the accuracy and reliability required for monetary operations.Whereas the expertise continues to evolve, significantly in dealing with nuanced queries and sustaining complete protection of system updates, it has already reworked the way in which groups work together with planning instruments independently.
Conclusion
The AI-powered assistant resolution mentioned on this publish has demonstrated important enhancements in information discovery and enterprise insights technology, delivering a number of key advantages throughout Amazon Finance. Analysts can now rapidly discover related data by way of pure language queries, dramatically decreasing search time. The system’s means to synthesize insights from disparate information sources has notably enhanced data-driven decision-making, and its conversational interface and contextual responses promote self-service information exploration, successfully decreasing the burden on centralized information groups.
This modern AI assistant resolution showcases the sensible energy of AWS generative AI in reworking enterprise information discovery and doc search. By combining Amazon Kendra Enterprise Version Index, Amazon Bedrock, and superior LLMs, the implementation achieves spectacular precision charges, proving that subtle AI-powered search is each achievable and efficient. This success demonstrates how AWS generative AI providers can meet present enterprise wants whereas selling future improvements in enterprise search. These providers present a robust basis for organizations seeking to improve information discovery processes utilizing pure language to help clever enterprise purposes. To study extra about implementing AI-powered search options, see Build and scale the next wave of AI innovation on AWS and explore AWS AI use cases.
In regards to the authors
Saikat Gomes is a part of the Buyer Options group in Amazon Internet Companies. He’s enthusiastic about serving to enterprises succeed and notice advantages from cloud adoption. He’s a strategic advisor to his prospects for large-scale cloud transformations involving folks, course of, and expertise. Previous to becoming a member of AWS, he held a number of consulting management positions and led large- scale transformation packages within the retail trade for over 20 years. He’s based mostly out of Los Angeles, California.
Amit Dhanda serves as a Senior Scientist at Amazon’s Worldwide Operations Finance group, the place he makes use of AI/ML applied sciences to resolve complicated ecommerce challenges. Previous to Amazon, he was Director of Knowledge Science at Adore Me (now a part of Victoria’s Secret), the place he enhanced digital retail experiences by way of recommender programs. He held science management roles at EXL and Thomson Reuters, the place he developed ML fashions for buyer engagement/progress and textual content classification.