ANNEE GARTON

Document Search and ML Training Tool

Product Design Case Study
Note: details omitted to protect company IP.
Problem Statement

A large corporation undergoes hundreds of lawsuits per year and has to sort through, in some cases, millions of digital files to provide legally-required material for each court case. How can we help reduce the number of documents that are sent out to external lawyers for legal review? 

Large lawsuits can be extremely lengthy and costly due to the large number of documents requiring legal review and the high cost of legal labor. The corporation wants to reduce the time and cost associated with these large lawsuits. 

​
Approach
​

1. I worked with the customer to understand their existing process and tools and their major problems and frustrations. 
Existing Process
  • Take a statistically significant random sample of the document set and have lawyers review this sample, making Relevant / Not Relevant determinations. Use the resulting ratios to estimate the percentage of Relevant documents in the entire set, also known as the "prevalence" of the set. 
  • ​Then, they use search terms and boolean operators to find potentially relevant documents and prioritize documents for legal review based on search relevance. 
  • Documents are iteratively sent for review (prioritized by search relevance) until lawyers find a number of Relevant documents that is close enough to the estimated prevalence.

Hypothesis 

If we train a machine learning model using Relevant and Not Relevant designations, we can predict the relevance of documents and prioritize documents for legal review, resulting in a superior prioritization of documents, reducing the number documents sent for review, as well as potentially catching more documents that would've been missed by using search terms.

Proposed Revised Process
  • Take a statistically significant random sample of the document set and have lawyers review this sample, making Relevant / Not Relevant determinations. Use the resulting ratio to estimate the percentage of Relevant documents in the entire set, also known as the "prevalence" of the set. 
  • Train the ML model on the Relevant and Not Relevant documents and prioritize documents for review based on predicated relevance.
  • Documents are iteratively sent for review (prioritized by predicted relevance), and the model is iteratively updated as more documents are reviewed until lawyers find a number of Relevant documents that is close enough to the estimated prevalence. ​
Picture
Picture
Picture
2. I worked with our designer to create Design mock-ups. In our designs, we focused on simplicity, ease-of-use, and interpretability of the machine learning components. 
  • Explore: We designed an "Explore" mode that was similar to the customer's existing search tools. This enabled them to search and interact with documents in a way they were familiar with. The user could also make Relevant and Not Relevant designations in this search and explore mode.  
  • Train: We designed a very simple and focused "Train" mode that displayed one document at a time for the user to make a Relevant or Not Relevant decision on. The backend served up documents for the user (first the random sample to estimate prevalence, then in order of priority based on the model's predictions of relevance). The design was as simple and focused as possible to remove any distractions from the decision-making process and expedite the training process. 
  • Review for Export: As this was for a legal use case and our users wanted to ensure the process was defensible, it was important to them that the machine wasn't making any decisions and, instead, providing rankings.  We designed a "Review" screen so that once the model had trained, users could interact with the model's predictions. The "Review" screen displayed a table with each document's relevance score and, if already available, a Relevant and Not Relevant designation. The user could sort and filter on the table to hone in on a specific document set. Clicking the "Export" button exported the current set of documents in the table. For this design, the goal was to make it simple to interact with the model's predictions and make it easy for users to use the results in their current process. 
  • Analytics: We ensured the user had quick access to high-level analytics, e.g. the number and percentage of documents that had been reviewed, throughout the whole experience, as well as access to more in-depth, detailed analytics. 
Picture
Picture
Picture
Picture
3. I worked with 7 product engineers to develop the UI, the search backend, and the training system that automatically trains and updates document scores as new information is provided. Simultaneously, I worked with an ML engineer to refine and test the training algorithm on various sample datasets, brainstormed different experiments to run and features to tweak, and reviewed the outputs. 
Results 
​
The customer ran a test using the new process and compared it to their old process and found they could reduce the documents needed for review by 50-70% which would save them a comparable amount of money and time. Additionally, they loved the design of the product, citing its simplicity and ease-of-use, which would enable quick and easy onboarding of new users. After seeing a demo of the tool, many other teams within the client organization became interested in using it for their own purposes. 
Picture
Picture
Though our first customers were in the legal space and the tool was initially intended for use in the legal space, we knew there were many other potential applications for this product. Many industries are presented with the problem of having a set of documents in which they want to find important information. We have since expanded to work with other customers in different industries to solve this same problem. 
Proudly powered by Weebly