My problem is that the bot predict several OOV or Out of scope intents with high probability (0.8-0.9). This make my bot not stable. To sup up, there are several False Positive intents.
are there a way to ameliorate the bot quality please ?
Yes, there is technically an infinite amount of messages that need to fall under the out-of-scope intent. But don’t forget that your bot is still in development. A lot of out-of-scope examples need to be added, and each time, your bot will be better at detecting real intents.
Also Tensorboard really helps a lot. Aim for a high test/validation accuracy, not training accuracy.
@Asmacats in general, if you’re seeing low accuracy and low confidence (i.e. less than 0.9 as a very rough guide) it usually means you either need more training data or to reduce your number of intents/the degree of overlap between them. This video talks about it a bit more: Conversational AI with Rasa: Training Data and Rules - YouTube
I believe that’s expected behaviour; tokens oov for another model can still be learned/given weights by the countVectorsFeaturizer: spaCy and OOV-Tokens
The high confidence is probably due to specific tokens being strongly associated with specific intents.