RelativityOne - Review Center Guide
RelativityOne - Review Center Guide
For the most recent version of this document, visit our documentation website.
Table of Contents
1 Review Center 5
1.1 Review Center overview 5
1.2 Review Center workflow 5
1.3 Understanding the integrative learning classifier 6
1.3.1 Language support in Review Center 6
1.4 Using Review Center versus batching 6
2 Creating a Review Center queue 8
2.1 Installing Review Center 8
2.2 Choosing a queue type 8
2.2.1 Saved search queues 8
2.2.2 Prioritized review queues 8
2.3 How document assignment works 9
2.3.1 Keeping document families together 10
2.4 Setting up the reviewer group 10
2.5 Creating required queue fields 10
2.6 Creating a queue template 11
2.7 Creating a new queue from a template 13
3 Monitoring a Review Center queue 15
3.1 Review Center dashboard 15
3.1.1 Queue tab strip 15
3.1.2 Queue Summary section 16
3.1.3 Review Progress section 20
3.2 Charts and tables 22
3.2.1 General charts and tables 22
3.2.2 Prioritized review charts 24
3.2.3 Reviewed Documents table 24
3.3 Deleting a queue 25
3.4 Fixing a misconfigured queue 25
3.5 Understanding document ranks 25
3.6 Tracking reviewer decisions 26
Note: Whenever you turn Coverage Mode on or off, manually refresh the queue. This updates the
document sorting for reviewers. For more information, see Turning Coverage Mode on and off on
page 19.
Note: If you set a relational field on a template or queue, set the same field in the Related Items drop-
down menu of the saved search Conditions tab. Only relational group members returned by the saved
search will be included in the queue. For more information, see Creating or editing a saved search in the
Searching guide.
1. Decide which user group or groups should contain the reviewers for the queue. For information on
creating and editing groups, see Groups in the Admin guide.
2. Add each group to the workspace.
3. Assign each reviewer group the following permissions:
Note: If a field has two colons (::) in the name, this is called a reflected field. Reflected fields
typically link two objects, and they cannot be used as the Rank Output field.
If you are creating a saved search queue, you do not need a Rank Output field, and the review field is
optional.
For more information about creating new fields, see Fields in the Admin guide.
Note: This field exists for all queues. If you toggle the Is Template setting to On for a regular
queue, it disappears from the dashboard and becomes usable as a template for other
queues. Toggling it off again returns the queue to the dashboard. The queue keeps all of its
statistics and coding decisions, but the queue state resets to Not Started.
3. Template Description—enter notes about the template such as its intended use, comments
about field settings, etc.
4. Queue Label—create and choose organizational labels that will apply to queues created from
this template. Some label ideas include First Level Review, Second Level Review, or Quality
Control. For more information, see Filtering the queue tab strip on page 16.
5. Reviewer Groups—this is not recommended for templates.
6. Queue Type—choose either Saved Search or Prioritized Review.
7. Data Source—select the saved search that contains the documents for your queue.
8. Rank Output (Prioritized Review only)—select the decimal field you created to hold the
document rank scores.
9. Review Field—select the single choice field you created for review. This field must have two
or more choices.
10. Positive Cutoff—on a scale of 0 to 100, enter the document rank that will be the dividing line
between positive and negative documents. All documents ranked at or above this number will
be predicted positive, and all documents ranked below it will be predicted negative. By default,
the cutoff is set at 50.
11. Relational Field—select a relational field for grouping documents in the queue. This makes
reviewers receive related documents together, such as members of the same document
family.
Note: If you set a relational field on a template or queue, set the same field in the Related
Items drop-down of the saved search Conditions tab. Only relational group members
returned by the saved search will be included in the queue. For more information, see Creat-
ing or editing a saved search in the Searching guide.
12. Allow Coded in Review (Saved Search only)—controls whether documents coded outside of
the queue will still be served up in the queue.
Note: Prioritized review queues use outside-coded documents to train their predictions, but
they only show them to reviewers if the Relational Field is set. For example, if the relational
field is set to Family Group and some members of a document family are already coded,
those will be served up to reviewers along with their family.
13. Queue Display Options—select which statistics you want reviewers to see on the queue card
in the Review Queues tab.
n Toggle this On to serve up mid-rank documents first. This is useful for training the model
as quickly as possible. You can turn this mode off at any time.
n Toggle this Off to serve up highest-rank documents first. This is the default setting for
finding relevant documents quickly.
16. Queue Refresh—controls whether the queue automatically refreshes after coding activity in
any queue. This refresh includes re-running the saved search and checking for outside-coded
documents. For prioritized review queues, this also re-trains the classifier with the latest coding
and re-ranks documents in order of predicted relevance. For more information, see Auto-
refreshing the queue on page 17.
n This panel shows reviewers a list of documents they previously reviewed in their queue.
For more information, see Finding previously viewed documents on page 28.
n If there are any conditions applied to the view, those conditions will also limit which
documents appear in the panel.
18. Reviewer Layout—select the coding layout that you want reviewers to see by default when
they enter the queue. If you do not choose a view, this defaults to the lowest ordered layout the
reviewer has permission to access.
19. Email Notification Recipients—enter email addresses to receive notifications about the
queue status. These emails tell users when a manually-triggered queue preparation
completes, a queue is empty, or a queue encounters an error while populating. To enter
multiple email addresses, separate them with a semicolon. Do not include a space.
4. Click Save.
The template now appears in the Review Library list.
Note: After a queue has been created from a template, the two of them are no longer connected. You can
edit the template without affecting the queue.
For information on starting, managing, and deleting queues, see Monitoring a Review Center queue on the
next page.
Below the queue name, each queue shows its status. The possible statuses are:
1. Click into the search bar above the queue tab strip.
A drop-down list of labels appears.
2. Select the labels you want to filter by. You can also type in the search bar to narrow the list, then
press Enter to select or deselect a label.
3. Close the list using the arrow at the right end of the bar.
The queue tab strip now only shows queues whose labels are listed in the search bar. If several
labels are listed, queues that match any one of them will appear.
The queue tab filters only apply to the tab strip. They do not affect any of the charts or statistics on the rest of
the page.
n Prepare Only—appears when the queue has not been started. This runs the saved search and trains
the classifier for the first time, but it does not start the queue. Alternately, you can click Prepare and
Start to perform both actions together.
n Refresh Queue—appears during a review that does not use auto-refresh. Clicking this refreshes the
queue.
n Auto Refresh—appears during a review that uses auto-refresh. Clicking this starts an immediate
refresh of the queue. For more information, see Auto-refreshing the queue below.
After you click Confirm, a Cancel option appears. For prioritized review queues, you may also see a
confirmation modal with the option to refresh the cache. For more information, see Caching extracted text in
prioritized review queues on the next page.
If you edit a queue's settings when the queue is partway through refreshing, the refresh will automatically
cancel. Any edits that affect the queue refresh will take effect during the next refresh.
If you need to trigger an immediate refresh, click on the words Auto Refresh to trigger an additional manual
refresh. For example, if new documents have been added to the saved search, you can click this to add
them to the queue quickly instead of waiting until the next auto-refresh.
While the queue is auto-refreshing, a Cancel option appears. If you cancel the current auto-refresh, the
queue will still try to auto-refresh again later.
Note: Canceling the queue preparation can take some time. If you need to remove reviewer access
immediately while canceling, edit the queue and remove the reviewer group.
Note: Whenever you turn Coverage Mode on or off, manually refresh the queue. This updates the
document sorting for reviewers. For more information, see Turning Coverage Mode on and off on the next
page.
n If the extracted text of documents in the queue's data source has not changed, leave the box
unchecked. This makes the refresh process much faster.
Note: You do not need to refresh the cache if you are simply adding or removing documents from
the queue.
n If the extracted text of documents in the queue's data source has changed, check the box. This tells
Review Center to re-cache the extracted text from all documents in the queue. Choosing to re-cache
the text may add significant time to the queue refresh.
Editing recommendations
Many edits are minor, and you can make them without pausing the queue. However, if you make a major
change such as changing the data source, we recommend:
1. On the right side of the Queue Summary section, click the three-dot menu.
2. Select Edit.
3. Click the Coverage Mode toggle to enable or disable it.
4. Click Save.
5. Manually refresh the queue. For more information, see Preparing or refreshing the queue on page 16.
If Coverage Mode is turned Off, the button will refer to the queue as "Prioritized Review."
On the Queue History table, you can also see if your queue was in Coverage Mode or not during each
queue refresh. For more information, see Queue History on page 23.
By default, the section shows a set of statistics that are calculated for all documents in the queue. By
clicking the triangle next to the section name, you can select another view.
n Total Docs—the total number of documents currently in the queue's data source. To be counted, the
queue must have been prepared or refreshed after the documents were added or removed. The
"100%" in smaller print underneath it indicates that this is the total document set.
n Docs Coded—the number of documents in the data source that either have a value in the review
field, or have been skipped. This includes documents coded outside the queue. The smaller
percentage underneath it shows the percentage of Docs Coded divided by Total Docs.
Note: The Predicted <Positive Choice> and Predicted <Negative Choice> fields only show their pre-
dictions after 50 or more documents have been coded.
n Docs Coded—the number of documents in the data source that were coded outside of the queue.
The smaller percentage underneath it shows the percentage of documents coded outside the queue
divided by all documents coded.
n <Positive Choice>—the number of documents that were coded positive outside of the queue. The
smaller percentage underneath it shows the percentage of documents coded positive outside the
queue divided by all documents coded.
n <Negative Choice>—the number of documents that were coded negative outside of the queue. The
smaller percentage underneath it shows the percentage of documents coded negative outside the
queue divided by all documents coded.
n To select a different visualization, click the blue arrow ( ) next to the visualization's name. This
opens a drop-down menu with all other visualizations.
n To switch from the chart view to the table view, click the Chart drop-down in the upper right corner
and select Table. This shows a table with the same information as the selected chart.
n To zoom in or out on a chart, hover the cursor over it and scroll. All charts reset to their default zoom
when you reload the page.
n
To download the panel contents, click the download symbol ( ) on the upper right. Charts download
as .png images, and tables download as .csv files.
Note: If any documents were coded by reviewers who are not part of this Relativity instance, those review-
ers will be listed as Unknown User 1, Unknown User 2, and so on. This can happen if a reviewer was
removed from the workspace or if the workspace has been archived and restored into a different instance.
n If you have more than one data point in a 15 minute increment, the chart shows them as two points on
a vertical line. This can happen if many reviewers are coding quickly.
n The Date field for a data point is the date the last document in the set of 100 was logged.
For prioritized review queues, the relevancy rate usually declines over time. However, the relevance rate
may spike if lots of new documents are added to the queue or if the definition of relevance changes during
review. For saved search queues, the shape of the relevancy rate graph varies depending on the saved
search being used.
If you zoom out on the Rank Distribution chart, you may see documents with ranks below zero. These are
documents that could not be classified. For more information, see Understanding document ranks on the
next page.
Note: If you delete a main queue that has a validation queue linked to it, it also deletes the validation
queue. For more information on validation queues, see Review validation on page 32.
If the classifier cannot classify a document, it will assign the document a value below zero. These values
are:
Negative
Document error
rank
-1 An error occurred while processing the data through the classifier.
-2 The extracted text field is empty. If you see this rank, consider making a saved search queue
to review these documents separately.
-3 The document's extracted text field is larger than the limit of 600 KB. If you see this rank, we
recommend filtering out large documents from your saved search to improve the per-
formance of the classifier.
n Review Center Coding::Reviewed On—the date of the original coding decision. Dates are based
on the UTC time zone.
n Review Center Coding::Reviewed By—the name of the reviewer who made the original coding
decision.
n Review Center Coding::Field Name—the name of the Review Field for the queue.
n Review Center Coding::Queue—the name of the Review Center queue that contains the
document.
n Review Center Coding::Value—the reviewer's coding decision.
For more information on creating views and saved searches, see Creating a view and Creating or editing a
saved search on the RelativityOne documentation site.
n <queue name> Validation <#>—this lists documents in an attached Validation queue. If the queue
has several attached Validation queues, each one will have its own tag.
n Removed—this lists any documents that were coded in the queue, but later removed from the data
source.
If you rename or delete a queue, this renames or deletes the matching Field Tree tags also.
1. Put your Reviewed On and Reviewed By fields into a saved search or view for monitoring.
2. Set your queue's review field as the Field to Monitor.
3. For most use cases, set Track Initial Change Only? to Yes. This sets it to track the first reviewer of
the document, instead of overwriting the Reviewed On and Reviewed By fields every time a user
edits the document.
If you set up the application after starting your queue, you can still see previous coding decisions by
following the steps under Populating Historical Records.
To return to your current document, click on the Navigate Last button in the upper right corner of the
document viewer.
Note: When you filter columns in the Documents panel, the filters only apply to documents on the current
page of the panel. For a comprehensive list of results, filter within the Documents tab of Relativity or run a
search from the saved search browser or field tree.
n Total docs in queue—the total number of documents in this queue, across all reviewers.
n Total remaining uncoded docs—the total number of uncoded documents in this queue, across all
reviewers.
n My docs reviewed total—how many documents you have reviewed total in this queue.
n Double check—always check the extracted text of a document to be sure it matches the content in
other views. Whenever possible, review from the Extracted Text viewer.
n Stay consistent—check with fellow reviewers to make sure your team has a consistent definition of
relevance. The AI classifier can handle occasional inconsistencies, but you’ll get the best results with
coordinated, consistent coding.
n When in doubt, ask—if something confuses you, don't guess. Ask a system admin or project man-
ager about the right course of action.
n Sufficient text—if there are only a few words or short phrases in a document, the engine will not
learn very much from it.
n Images—text contained only in images, such as a photograph of a contract, cannot be read by
Review Center. The system works only with the extracted text of a document.
n Numbers—numbers are not considered by Review Center.
Note: Review validation does not check for human error. We recommend that you conduct your own
quality checks to make sure reviewers are coding consistently.
n Discard pile—the set of documents that are uncoded, skipped, or coded as neutral. This also
includes documents that are being reviewed when validation starts, but their coding has not been
saved yet.
n Already-coded documents—documents that have already been coded as either positive or
negative. These are counted as part of the validation process, but they will not be served up to
reviewers a second time. Neutral-coded documents are considered part of the discard pile instead,
and those may be served up a second time.
1. Validation Reviewer Groups—the user groups you want reviewing the queue.
2. Cutoff—enable this to set a cutoff for the validation queue. For more information, see How
setting a cutoff affects validation statistics on page 41.
3. Positive Cutoff—if Cutoff is enabled, enter a custom cutoff value between 0 and 100 in this
field. This rank will be used as the dividing line between documents predicted positive and
documents predicted negative. Setting this value also adds a Precision statistic to the
validation results.
4. Set the sample size using three interconnected fields:
1. Sample Size—this sets a fixed number of documents for the sample size. By default,
this field is set to 1000 documents. The sample size must be larger than 5 and smaller
than the size of the discard pile.
2. Margin of Error Estimate (Elusion)—this calculates a size for the sample based on
how accurate the Elusion statistic will be.
3. Margin of Error Estimate (Recall)—this calculates a size for the sample based on how
accurate the Recall statistic will be.
Note: Each of these fields affects the others. For an explanation of how they work,
see Choosing the validation settings below.
5. Click Save.
1. You can specify exactly how many documents you want to sample. Review Center automatically
calculates the estimated margins of error for both Elusion and Recall based on the sample size you
select.
2. You can specify the desired margin of error for the elusion estimate and let Review Center calculate
an appropriate sample size. Review Center also automatically calculates the corresponding recall
margin of error.
Note: This is equivalent to the “statistical” option when configuring an Elusion with Recall test in
Active Learning.
3. You can specify the desired margin of error for the recall estimate and let Review Center calculate an
appropriate sample size. Review Center also automatically calculates the corresponding elusion
margin of error.
The final margin of error estimates may be slightly different from the ones chosen at setup, depending on
the documents found during validation. All validation statistics aim for a 95% confidence interval alongside
the margin of error.
The estimated elusion margin of error depends only on the sample size, and vice versa. Their relationship to
the estimated recall margin of error depends on the number of relevant documents that have already been
coded and the current size of the discard pile. It may vary among different validation samples, even within
the same review.
For more information on how validation statistics are calculated, see Review validation statistics on
page 40.
For more information on the validation statistics, see Reviewing validation results on page 37.
1. On the right side of the Queue Summary section, click on the three-dot menu and select Edit.
2. Edit any of the following settings:
n Reviewer Groups
n Queue Display Options
n Reviewer Document View
n Reviewer Layout
n Email Notification Recipients
3. Click Save.
For descriptions of the queue settings, see Creating a Review Center queue on page 8.
n The validation queue status changes to Rejected, and the main review queue changes to Paused.
n The main review queue re-opens for normal coding, and you can build the model again at any time.
Any documents coded since validation began, including those from the validation queue itself, will be
included in the model build.
n The Coding Progress strip on the dashboard displays the main queue's statistics.
You can run validation on the queue again at any later time, and you can reject validation rounds as many
times as needed. Even if you reject the results, Review Center keeps a record of them. For more
information, see Viewing results for previous validation queues on the next page.
1. On the right side of the Queue Summary section, click on the three-dot menu and select Reject
Validation.
2. Click Reject.
After you have rejected the validation results, you can resume normal reviews in the main queue.
n Relevance Rate—percentage of sampled documents that were coded relevant by reviewers, out of
all coded documents in the sample. If any documents were coded as neutral, this statistic also counts
them as relevant.
n Elusion Rate—the percentage of unreviewed documents that are predicted as non-relevant, but that
are actually relevant.. The range listed below it applies the margin of error to the sample elusion rate,
which is an estimate of the discard pile elusion rate.
Notes:
o If you do not set a cutoff for your validation queue, this is calculated as the percentage of all
unreviewed documents that are actually relevant.
o Documents that are skipped or coded neutral in the validation queue are treated as relevant
documents when calculating Elusion Rate. Therefore, coding all documents in the elusion
sample as positive or negative guarantees the statistical validity of the calculated elusion rate
as an estimate of the entire discard-pile elusion rate.
1. On the right side of the Queue Summary section, click on the three-dot menu and select Recalculate
Validation.
2. Click Recalculate.
1. Click the triangle symbol near the right side of the Queue Summary section.
n Elusion rate—the percentage of unreviewed documents that are predicted as non-relevant, but that
are actually relevant. The rate is rounded to the nearest single digit (tenth of a percent).
n Recall—the percentage of truly positive documents that were found by the review.
n Richness—the percentage of relevant documents across the entire review.
n Precision—the percentage of found documents that were truly positive. This statistic is only cal-
culated if you set a cutoff when creating the validation queue.
The calculations for elusion and recall change depending on whether the validation queue uses a cutoff. For
more information, see How setting a cutoff affects validation statistics on the next page.
For each of these metrics, the validation queue assumes that you trust the human coding decisions over
machine predictions, and that the prior coding decisions are correct. It does not second-guess human
decisions.
Note: Validation does not check for human error. We recommend that you conduct your own quality
checks to make sure reviewers are coding consistently.
n Coded, not relevant—documents that have been coded but are not relevant.
n Coded, relevant—documents that have been coded and are relevant.
n Uncoded, predicted not relevant—documents that have not been coded and are predicted not rel-
evant.
n Uncoded, predicted relevant—documents that have not been coded and are predicted relevant.
n Uncoded documents—have not received a positive or negative coding label. This includes any doc-
uments that:
o have not been reviewed yet.
o are being reviewed at the moment the validation starts, but their coding has not been saved
yet.
o were skipped.
o received a neutral coding label.
The system also knows how many documents are in buckets 3 and 4 altogether, but not the precise
breakdown between the two buckets.
You could find out by exhaustively coding the uncoded documents, but that’s time-consuming. Instead,
review validation uses statistical estimation to find out approximately how many are in each bucket. This
means that any statistics involving bucket 3 or 4 will include a confidence interval that indicates the degree
of uncertainty about how close the estimate might be to the true value.
Notes:
n These buckets are determined by a document's status at the start of Project Validation. For the
purpose of these calculations, documents do not "switch buckets" during the course of validation.
n If you choose not to set a cutoff for your queue, buckets 3 and 4 are combined, and all uncoded
documents are predicted not relevant. For more information, see How setting a cutoff affects
validation statistics below.
Recall = (Bucket 2 + relevant documents in bucket 4) / (Bucket 2 + relevant documents in buckets 3 and 4)
Recall measures the percentage of truly positive documents that were found by the review. Recall shares a
numerator with the precision metric, but the denominators are different. In recall, the denominator is "what is
truly relevant;" in precision, the denominator is "what we are producing." Documents coded outside of the
queue during validation count against recall.
If you do not set a cutoff for your validation queue, this is calculated as the number of documents that were
previously coded relevant, divided by the total number of documents coded relevant in any group.
6.4.3 Richness
This is the percentage of documents in the review that are relevant.
Richness = (Bucket 2 + any relevant documents found in buckets 3 and 4) / (All buckets)
6.4.4 Precision
Precision is the percentage of truly positive documents out of all documents that were expected to be
positive. Documents that were predicted positive, but coded negative, lower the precision percentage.
Note: The Workspace - Edit Security permission is only required to edit the assigned reviewer group.
If you want a user group to only see specific queues on the dashboard, you can restrict a queue using item-
level security on the Review Library tab. For more information, see Security and permissions in the Admin
guide.
Users with access to the Review Center dashboard can also track reviewer decisions using the Reviewed
Documents table. For more information, see Tracking reviewer decisions on page 26.
n Start time—the time the user clicked the button to start the job.
n End time—the time the last document became available in Review Center.
Documents in Pre-coded doc- Total run time (h:m- GB per Documents per
queue uments m:ss) hour hour
1,000,000 10 1:05:50 13.57 920,000
Test Num- Pre-coded doc- Time for initial run (h:m- Time for subsequent run (h:m-
ber uments m:ss) m:ss)
Test 1 10 1:05:50 0:17:18
Test 2 100,000 1:08:50 0:20:15
Test 3 200,000 1:06:05 0:19:28
Test 4 300,000 1:02:19 0:21:19
Test 5 400,000 0:59:42 0:21:58
Test 6 500,000 1:01:23 0:17:42
Test 7 600,000 0:58:06 0:23:51
Test 8 700,000 0:52:38 0:16:17
Test 9 800,000 0:56:36 0:17:05
Test 10 900,000 1:07:43 0:20:07
Test 11 1,000,000 0:58:43 0:21:07
Test Num- Pre-coded doc- Time for initial run (h:m- Time for subsequent run (h:m-
ber uments m:ss) m:ss)
Test 1 10 4:14:49 0:48:11
Test 2 300,000 4:27:43 0:49:11
Test 3 400,000 4:35:43 0:50:02
Test 4 700,000 4:53:09 0:47:46
Test 5 800,000 4:02:12 0:58:08
Test 6 900,000 4:34:30 0:52:27
Test 7 1,000,000 4:04:45 0:54:06