Local llm error with enviromental variable OPENAI key

Hi all,

looks like we have the same problem using a Mistral model. First we got an error that says the command generator needs an OPENAI_API_KEY and after disabling the flow retrieval like in this post Rasa Pro Tutorial with Azure OpenAI Service not working - #2 by emilymoore04 we can train successfully but when chatting with the bot, another tool asks for the OPENAI_API_KEY

What we are trying is just get some experience with CALM with our Mistral model. And we are trying to get the default tutorial running with this model (Tutorial)

Our config looks like this:

recipe: default.v1
language: en
pipeline:
  - name: SingleStepLLMCommandGenerator
    llm:
      provider: mistral
      model: mistral-large-latest
    flow_retrieval:
      active: false

policies:
  - name: FlowPolicy
#  - name: EnterpriseSearchPolicy
#  - name: RulePolicy

And our .env file looks like:

RASA_PRO_LICENSE='rasa_pro_license_key'
MISTRAL_API_KEY='mistral_api_key'

Thanks in advance for your help!