Support Centre

Model Training & Maintenance

Guides on how to create, improve and maintain Models in Re:infer, using platform features such as Discover, Explore and Validation

Reviewing label predictions

User permissions required: 'View Sources' AND 'Review and label'

What's in this article?




The second step in the Explore phase is called ‘Review predictions’. After the Discover phase and some training in shuffle mode, the model will have started making predictions for many of the labels in your taxonomy. 

The purpose of this step is to review these for each label, confirming whether they are correct and correcting them where they aren't, and thereby providing many more training examples for the model. 

There are therefore two key actions in this step when reviewing label predictions:


  • Where the predictions are correct, you should confirm/accept them 
  • Where they are incorrect, you should either dismiss them or alternatively add the correct label(s) that does apply. 


The images below show how predictions look in Re:infer for data with and without sentiment. Hovering your mouse over the label will also show the confidence the model has that the specific label applies. 

The transparency of the predicted label provides a visual indicator of Re:infer’s confidence. The darker the colour, the higher Re:infer’s confidence is in that label applying, and vice versa:



Verbatim with predictions without sentiment enabled



Verbatim with predictions with sentiment enabled


Remember: telling a model that a label does not apply is just as important as telling it what does

Key steps



  1. Select the unreviewed filter in Explore (this shows verbatims not yet reviewed by a human) *
  2. Select the label mode from the dropdown menu at the top of the page, and then the label from the taxonomy dropdown that you wish to train
    • Re:infer will now present you with unreviewed Verbatims with predictions for the label you have selected. These will be presented in descending order of confidence that the selected label applies, i.e. with the most confident first and least confident at the bottom
  3. To confirm a label applies simply click on it - e.g. 'Facilities > Spa and pool' shown above
  4. To add a different or additional label to one you have already selected, click the ‘+’ button and type it in. This is the way to correct wrong predictions, by adding the correct one and not clicking on any incorrectly predicted labels


Please Remember: at all times to also add in any other labels that apply to the verbatim you're reviewing

* Please Note: If you filter to unreviewed verbatims in Explore, the counts next to the labels will update to show you the predicted count of how many times that label occurs in the dataset. The more you train the label, the more this predicted count should accurately reflect the true number in the dataset.

To delete a label you applied in error you can hover over it and an ‘X’ will appear. Click this to remove the label.


Which predictions should you review?


  • From a training perspective, it's only useful to review predictions that are not high-confidence (90%+)
  • This is because when the model is very confident (i.e. above 90%), then by confirming the prediction you are not telling the model a lot of new information, it's already confident that the label applies, such as in the example below
  • If you select a label and see lots of high confidence predictions already, even if there's very few training examples, it's likely you should move straight on to using 'Teach label' instead (see here)
  • 'Teach label' focuses on training examples that Re:infer is most unsure about, meaning that each verbatim reviewed in this mode sends a powerful training signal to Re:infer
  • If predictions have high confidences and are wrong, however, then it's important to apply the correct label(s), thereby rejecting the incorrect prediction(s)
  • Predictions are shown in descending order of confidence, with the highest confidence predictions for a label shown first


How many predictions should you review per label?


As outlined in 'Intro to Explore', labels with less than 25 pinned examples will have a red training dial next to them (see below). Re:infer requires 25 pinned examples to be able to accurately evaluate how well it is able to predict that label. The fuller and brighter the circle, the fewer the training examples there are and the more training is required.

It's important, however, that these 25 or more examples are provided using a variety of training methods. It is not good practice to review and accept lots of predictions (particularly high confidence ones) in just 'label' mode. You should make sure to use the 'Shuffle', 'Teach' and 'Low Confidence' modes throughout the training process to provide examples for each label.

Label training dials

What are the red and amber warnings that appear next to some labels?


Once you reach 25 pinned examples you may see one of the below label performance indicators in place of the training health circle:

  • The grey circle is an indicator that the platform is calculating the performance of that label - it will update to either disappear, or an amber or red circle once calculated
  • Amber is an indicator that the label has slightly less than satisfactory performance and could be improved
  • Red is an indicator that the label is performing poorly and needs additional training / corrective actions to improve it
  • If there is no circle, then this means that the label is performing at a satisfactory level (though still may need improving depending on the use case and desired accuracy levels)
  • To understand more about label performance and how to improve it, you can start here

Label performance indicators

Previous: Training using 'Shuffle'    |     Next: Training using 'Teach label' (Explore)

Did you find it helpful? Yes No

Send feedback
Sorry we couldn't be helpful. Help us improve this article with your feedback.


View all