Evaluate Classifier: Elusion Test

Prev Next

Overview

An elusion test is a statistical validation method used in eDiscovery to assess a predictive coding model’s ability to distinguish relevant from non-relevant documents. In principle, an elusion test takes a sample of documents coded “non-relevant” by a model and sets them aside for manual, human review. The percentage of the model’s “non-relevant” documents manually coded by humans as “relevant” estimates the prevalence of relevant documents missed (eluded) by the model during assessment. This percentage, called the elusion rate, helps judge the accuracy of a model’s assessment capabilities.

Classifiers that have a high elusion rate (lower accuracy) should be trained until their rate is at an acceptable level — usually a value that has been defined by the user, or agreed upon by multiple parties.

Perform an Elusion Test

I. Create a New Tag for Your Elusion Test

See Supervised Learning Overview for instructions on how to create a new AI tag used to train a classifier.

II. Confirm Tag and Score Syncing

  1. Navigate to Supervised Learning > Tagging & Saving

  2. Confirm steps 1 through 3 (Tag Sync, Classification, and Score Display) are listed “Ready,” with a green checkmark.

Steps showing readiness status for tag sync, classification, and score display processes.

III. Create Saved Search That’s Below Score Threshold

  1. Under Filters in the Sidebar, expand Document Status and select Has Tags.

  2. Open search building for an Advanced Search.

  3. Toggle the Tag Paths: Has Any Value pill to NOT.

  4. Click Add.

  5. Expand Model Scores.

  6. Select your Responsiveness AI Tag.

  7. Set the range of scores between zero and your Score Threshold (here, 70).

    Usually, Case Owners decide the score threshold. Documents below a score threshold are considered Not Relevant / Not Responsive while documents at or above a score threshold are considered Relevant / Responsive. Including your score threshold as part of your elusion test yields more conservative results.

  8. Click Add To Search.

  9. Run the Search.

  10. Save the Search as “Elusion Test” with the current date.

IV. Sample Your Documents

  1. From the Grid bulk action menu, create a Sample using values for Confidence and Margin of Error as appropriate and consistent with your review.

  2. In the Sample Documents modal…Options for sample documents including confidence level and folder selection settings.

    1. Choose Confidence as your Sample Type.

    2. Choose a Confidence level (90% recommended) and Margin of Error (5% recommended). You can type decimal values into either of these fields.

    3. Select work folder to copy samples, and Select Search Folder to create saved search. For both of these fields, you can either choose from an already existing folder or press the ellipses [ … ] icon to create a new folder.

  3. Click SUBMIT.

V. Review Documents and Calculate Scores

  1. Have a subject matter expert review the above samples as either “Responsive” or “Not Responsive.”

  2. Calculate the elusion rate (percentage) using the below formula:

  1. Decide whether or not your rate is acceptable. Continuing to train a classifier will improve its accuracy and lower its elusion rate.

  2. Use the elusion rate to calculate how many non-relevant documents may have been coded incorrectly by your classifier.

Formula for estimating relevant documents based on elusion rate and non-responsive documents.

Final Considerations

Case Owners and/or Project Managers may be responsible for choosing appropriate Score Thresholds, Confidence Levels, Margin of Errors, and determining an appropriate elusion rate for your classifiers. These values can vary depending on risk assessment and project characteristics.

If new training is required, the tags used for your elusion test can be applied to the actual classifier tag, and the following workflow can be followed:

  1. Transfer review tag used for elusion test to the actual review tag.

  2. After tags are synced, kick off “Run Full Process” to force the classifier to rescore documents.

  3. Re-run elusion testing steps above.

  4. Re-assess whether new elusion rate is acceptable.

Other validation methods, like creating control sets or assessing for precision and recall, are also recommended for validating a classifier’s accuracy rating.