Hi, i’ve watched this video
and i’m confused with hows positional encoding work with CLS vector
Since there’s a CLS token in DIET that summarize the entire utterance, and the example of positional encoding on 08:17 is word index vector. So, how positional enc works with CLS vector ?
I’m writing a thesis on how DIET works in intent and entities prediction, any answer would be appriciated
Hi @setopaisen That’s a good question. The CLS token is always added to end of the utterance, so the position index for it is basically - length of utterance + 1 (Assuming indices start at 1)