I have a use case only for NLU part, where I have to use several models at the same time. So, I’m using the below code.
But the response time is very high around 4-6 sec. Is there any way where I can use several models at the same time and get the response very quickly (only NLU).
I took a clone of git repo version 1.10.x
Why do you have multiple models? It might be faster to include them all in one pipeline. But I guess that also depends on what kind of models you have. Can you share some more details?