Update README.md
Browse files
README.md
CHANGED
@@ -29,5 +29,5 @@ outputs = vision_tower(**inputs)
|
|
29 |
logits_per_image = outputs.pooler_output # this is the image-text similarity score
|
30 |
```
|
31 |
|
32 |
-
There's still a slight
|
33 |
|
|
|
29 |
logits_per_image = outputs.pooler_output # this is the image-text similarity score
|
30 |
```
|
31 |
|
32 |
+
There's still a slight difference where hf's CLIPVision model uses a [CLS] embedding as pool embedding while SigLIP uses global attention pooler to get the final latent feature.
|
33 |
|