Hello,
I have an intent goodbye with some german examples to say goodbye:
##intent:goodbye
- bis später
- tschüss
- bye
and here is my config.yml for the spacy model
language: "de_core_news_sm"
pipeline:
- name: "nlp_spacy"
- name: "tokenizer_spacy"
- name: "intent_entity_featurizer_regex"
- name: "ner_crf"
features: [
["low", "title", "upper"],
["bias", "low", "prefix5", "prefix2", "suffix5", "suffix3", "suffix2", "upper", "title", "digit", "pattern", "pos", "pos2"],
["low", "title", "upper"]]
- name: "intent_featurizer_spacy"
- name: "intent_classifier_sklearn"
after training the nlu and the core I try to communicate with the bot but with some major problems. If I ask him total nonsence like syedxgvhkmijgedsfhdrwysa it falls in the intent goodbye with a confidence about 47%.
UserUttered(text: syedxgvhkmijgedsfhdrwysa, intent: {'name': 'goodbye', 'confidence': 0.4790237086878828}, entities: [])
So one method to prevent such problems is to create a fallback policy and configure it with a high nlu confidence. But are there any other methods to prevent these problems? Maybe configurations directly in spacy?
Another question: The string above is a very weird combination of many letters, absolutely with no analogy to my goodbye-examples. Why do I have such a high confidence?