The tokens returned by Spacy should still be converted into distributed representations to be easily usable by a DL model. For the baseline, do we encode the question with the Universal Sentence Encoder and use a SOTA word embedding for the context? Which word embedding is still very popular? (Fastext, Glove, ?) Also, some researchers use Pair Byte Encoding (PBE) to encode information at the sub-word level.
The tokens returned by Spacy should still be converted into distributed representations to be easily usable by a DL model. For the baseline, do we encode the question with the Universal Sentence Encoder and use a SOTA word embedding for the context? Which word embedding is still very popular? (Fastext, Glove, ?) Also, some researchers use Pair Byte Encoding (PBE) to encode information at the sub-word level.