File size: 824 Bytes
f999714
 
 
37a4eb5
 
 
 
 
 
1
2
3
4
5
6
7
8
9
---
license: afl-3.0
---

This model is actually very accurate for this task, intuitively inspired by information retrieval techniques. In 2019, Nils Reimers and Iryna Gurevych introduced a new transformers model called Sentence-BERT, Sentence Embeddings using Siamese BERT-Networks. This model was introduce in this paper: https://doi.org/10.48550/arxiv.1908.10084}

This new Sentence-BERT model is modified on the BERT model by adding a pooling operation to the output of BERT model. In such a way, it can output a fixed size of the sentence embedding to calculate cosine similarity, and so on. To obtain a meaningful sentence embedding in a sentence vector space where similar or pairwise sentence embedding are close, they created a triplet network to modify the BERT model as the architecture below figure.

![](1.png)