Application
Sentence embedding is used by the deep learning software libraries PyTorch and TensorFlow. Popular embeddings are based on the hidden layer outputs of transformer models like BERT, see SBERT. An alternative direction is to aggregate word embeddings, such those returned by Word2vec, into sentence embeddings. The most straightforward approach is to simply compute the average of word vectors, known as continuous bag-of-words (CBOW). However, more elaborate solutions based on word vector quantization have also been proposed. One such approach is the vector of locally aggregated word embeddings (VLAWE), which demonstrated performance improvements in downstream text classification tasks.Evaluation
A way of testing sentence encodings is to apply them on Sentences Involving Compositional Knowledge (SICK) corpus for both entailment (SICK-E) and relatedness (SICK-R). In the best results are obtained using a BiLSTM network trained on thSee also
* Distributional semantics * Word embeddingExternal links
References
{{Reflist Language modeling Artificial neural networks Natural language processing Computational linguistics