After launched the first version, we collect valuable feedback from clients. Through analyzed their feedback and behavior data, we made an iteration.
1. Reduce repetitive work
The system selected samples from conversation log randomly, leading to a plenty of similar questions in the task. It’s not productive to let annotators deal with these similar questions for many times. So I reduced the repetitive by clustering questions with high similarity, and only select a few from the cluster.
2. Simplify annotation flow
If the questions are not random distributed, then it can’t evaluate accurately. So I separated annotation into two types tasks — improvement, evaluation.
3. Add missing corpus with efficiency
Cluster the similar questions, and add it once in an Excel sheet. It saves a plenty of time to edit while annotating. And it also reduce repetitive work.
3. Refer to context of the conversation
Sometimes annotators need to know the context of the question. So we provide an access refer to the whole conversation of this session.
The annotation system helps our clients improve the AI performance successfully and rise the productivity of annotating.