Sentence Embedding Algorithm
		
		
		
		
		
		Jump to navigation
		Jump to search
		
		
	
A Sentence Embedding Algorithm is a text-item embedding algorithm that can be implemented by a sentence embedding system to solve a sentence embedding task.
- AKA: Sentence Encoding Method.
 - Context:
- It can (typically) transform individual sentences into fixed-length vectors in a high-dimensional space.
 - It can (typically) leverage Neural Network Architectures, such as Recurrent Neural Networks (RNNs), Convolutional Neural Networks (CNNs), or Transformer models, for generating embeddings.
 - It can (often) be trained on large corpora of text data to capture a wide range of linguistic properties and nuances.
 - ...
 
 - Example(s):
- Sentence-BERT (SBERT) Algorithm, which adapts the BERT model using Siamese and triplet network structures for efficient sentence embedding.
 - Universal Sentence Encoder, which uses deep averaging network (DAN) and transformer architecture to generate sentence embeddings.
 - InferSent, a model trained on natural language inference data to derive universal sentence representations.
 - ...
 
 - Counter-Example(s):
- a Word Embedding Algorithm, which focuses on mapping individual words to vectors, without considering the broader sentence context.
 - a Document Embedding Algorithm, which generates document embeddings.
 - ...
 
 - Counter-Example(s):
 - See: Text Item, SentenceTransformers.
 
References
2017
- (Lin et al., 2017) ⇒ Zhouhan Lin, Minwei Feng, Cicero Nogueira dos Santos, Mo Yu, Bing Xiang, Bowen Zhou, and Yoshua Bengio. (2017). “A Structured Self-attentive Sentence Embedding.” In: Proceedings of the 5th International Conference on Learning Representations (ICRL-2017).