As global commerce accelerates, Adyen, a publicly-traded financial technology platform, is helping large companies like Meta, Uber, H&M, and Microsoft achieve their ambitions faster by providing end-to-end payments capabilities, data-driven insights, and financial products in a single global solution.
With more merchants signing on and with increased transaction volume comes increased pressure on support teams and a team at Adyen that immediately sought out leveraged solutions. "We are engineers so we are always looking at ways to scale our business using technology without having to increase the team size,” said Andreu Mora, SVP of Engineering - Data. “We want to understand, harness, and advance technology like LLMs to make our teams and customers more efficient and more satisfied.”
Adyen spun up a new and lean team of Data Scientists and Machine Learning Engineers based out of their new Tech Hub in Madrid, Spain to take on a range of high-impact projects, with the initial goal of accelerating support teams.
Focusing on support team efficiency and satisfaction led the team to an insight that passing tickets between teams was a major factor influencing response times. This challenge was particularly well-suited to be solved by two initial LLM applications:
- A smart ticket routing system designed to get a ticket to the right support person as quickly as possible based on content
- A support agent copilot designed to help agents answer tickets faster and more accurately with an approach they call Question Answering Suggestions.
Adyen decided to use LangChain to build it because they could rely on a single, easy-to-customize, framework to get from prototype to production, and avoid getting locked into a single model as they experimented. They also relied on LangSmith, LangChain’s developer platform to evaluate performance of their applications and compare how different underlying models affected response quality and costs.
LangChain’s flexibility allowed Adyen to switch the core part of the chain among various LLMs with ease. To ensure seamless interactions with their internal LLM API endpoint, they introduced a custom class extending from LangChain’s base LLM class, and integrated it with an event-driven microservice architecture hosted in a Kubernetes cluster. For a more in-depth exploration of the technical aspects behind Adyen's smart ticket routing and support agent copilot, check out the technical deep dive by Andreu Mora and Rafael Hernandez (Team Lead of Operations AI) on the Adyen blog here.
More efficient and accurate ticket routing
The foundation of Adyen’s smart ticket router is an internal tool that analyzes the theme and sentiment of each ticket, and dynamically updates its priority based on the user. With Adyen’s wide array of products, features & services, this LLM-driven approach enables merchants to receive support from the technical experts most suited to respond quickly.
Quicker support response times
In just 4 months the Adyen team was able to build a comprehensive collection of relevant documents (combining public and private company documents) and store them in a vector database with an embedding model that optimized for effective retrieval. The team’s first milestone on its way to generating proposed ticket responses was finding the most relevant and up-to-date document from a collection of public and private documents. This approach far outperformed traditional keyword-based search and, just as importantly, immediately established the team’s trust in the new system.
The next step was to connect to an LLM to produce a suggested response for support agents through their proprietary copilot. “With the right set of tickets in their queues and easily-modifiable potential answers to customer inquiries at their fingertips, support agents are more efficient and more satisfied.” said Andreu.