multi stage tuning
Browse files- README.md +406 -567
- config.json +1 -1
- model.safetensors +1 -1
README.md
CHANGED
@@ -6,10 +6,8 @@ tags:
|
|
6 |
- sentence-similarity
|
7 |
- feature-extraction
|
8 |
- generated_from_trainer
|
9 |
-
- dataset_size:
|
10 |
-
- loss:
|
11 |
-
- loss:AnglELoss
|
12 |
-
base_model: nomic-ai/nomic-embed-text-v1.5
|
13 |
datasets: []
|
14 |
metrics:
|
15 |
- cosine_accuracy
|
@@ -27,58 +25,68 @@ metrics:
|
|
27 |
- spearman_dot
|
28 |
- pearson_max
|
29 |
- spearman_max
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
widget:
|
31 |
-
- source_sentence:
|
32 |
sentences:
|
33 |
-
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
- Coolibar UPF 50+ Men's Women's Gannett UV Gloves - Sun Protective (Medium- Light
|
40 |
-
Blue), Coolibar, Light Blue
|
41 |
-
- source_sentence: flame decal stickers
|
42 |
sentences:
|
43 |
-
-
|
44 |
-
|
45 |
-
-
|
46 |
-
|
47 |
-
|
48 |
-
-
|
49 |
-
|
50 |
-
|
51 |
-
- source_sentence: 'search_query: softies women''s ultra soft marshmallow hooded lounger'
|
52 |
sentences:
|
53 |
-
- 'search_document:
|
54 |
-
|
55 |
-
|
56 |
-
- 'search_document:
|
57 |
-
|
58 |
-
- 'search_document:
|
59 |
-
|
60 |
-
|
|
|
61 |
sentences:
|
62 |
-
- 'search_document:
|
63 |
-
|
64 |
-
|
65 |
-
|
66 |
-
|
67 |
-
|
68 |
-
|
69 |
-
|
|
|
|
|
|
|
70 |
sentences:
|
71 |
-
- 'search_document:
|
72 |
-
|
73 |
-
|
74 |
-
|
75 |
-
|
76 |
-
|
77 |
-
60x80 - Soft Heavy Blanket with Breathable TPE Insert No Glass Beads, Bedsure,
|
78 |
-
Navy'
|
79 |
pipeline_tag: sentence-similarity
|
80 |
model-index:
|
81 |
-
- name: SentenceTransformer
|
82 |
results:
|
83 |
- task:
|
84 |
type: triplet
|
@@ -88,19 +96,19 @@ model-index:
|
|
88 |
type: unknown
|
89 |
metrics:
|
90 |
- type: cosine_accuracy
|
91 |
-
value: 0.
|
92 |
name: Cosine Accuracy
|
93 |
- type: dot_accuracy
|
94 |
-
value: 0.
|
95 |
name: Dot Accuracy
|
96 |
- type: manhattan_accuracy
|
97 |
-
value: 0.
|
98 |
name: Manhattan Accuracy
|
99 |
- type: euclidean_accuracy
|
100 |
-
value: 0.
|
101 |
name: Euclidean Accuracy
|
102 |
- type: max_accuracy
|
103 |
-
value: 0.
|
104 |
name: Max Accuracy
|
105 |
- task:
|
106 |
type: semantic-similarity
|
@@ -110,52 +118,94 @@ model-index:
|
|
110 |
type: unknown
|
111 |
metrics:
|
112 |
- type: pearson_cosine
|
113 |
-
value: 0.
|
114 |
name: Pearson Cosine
|
115 |
- type: spearman_cosine
|
116 |
-
value: 0.
|
117 |
name: Spearman Cosine
|
118 |
- type: pearson_manhattan
|
119 |
-
value: 0.
|
120 |
name: Pearson Manhattan
|
121 |
- type: spearman_manhattan
|
122 |
-
value: 0.
|
123 |
name: Spearman Manhattan
|
124 |
- type: pearson_euclidean
|
125 |
-
value: 0.
|
126 |
name: Pearson Euclidean
|
127 |
- type: spearman_euclidean
|
128 |
-
value: 0.
|
129 |
name: Spearman Euclidean
|
130 |
- type: pearson_dot
|
131 |
-
value: 0.
|
132 |
name: Pearson Dot
|
133 |
- type: spearman_dot
|
134 |
-
value: 0.
|
135 |
name: Spearman Dot
|
136 |
- type: pearson_max
|
137 |
-
value: 0.
|
138 |
name: Pearson Max
|
139 |
- type: spearman_max
|
140 |
-
value: 0.
|
141 |
name: Spearman Max
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
142 |
---
|
143 |
|
144 |
-
# SentenceTransformer
|
145 |
|
146 |
-
This is a [sentence-transformers](https://www.SBERT.net) model
|
147 |
|
148 |
## Model Details
|
149 |
|
150 |
### Model Description
|
151 |
- **Model Type:** Sentence Transformer
|
152 |
-
- **Base model:** [
|
153 |
- **Maximum Sequence Length:** 8192 tokens
|
154 |
- **Output Dimensionality:** 768 tokens
|
155 |
- **Similarity Function:** Cosine Similarity
|
156 |
-
- **Training
|
157 |
- triplets
|
158 |
-
- pairs
|
159 |
<!-- - **Language:** Unknown -->
|
160 |
<!-- - **License:** Unknown -->
|
161 |
|
@@ -192,9 +242,9 @@ from sentence_transformers import SentenceTransformer
|
|
192 |
model = SentenceTransformer("lv12/esci-nomic-embed-text-v1_5_4")
|
193 |
# Run inference
|
194 |
sentences = [
|
195 |
-
'search_query:
|
196 |
-
'search_document:
|
197 |
-
'search_document:
|
198 |
]
|
199 |
embeddings = model.encode(sentences)
|
200 |
print(embeddings.shape)
|
@@ -240,11 +290,11 @@ You can finetune this model on your own dataset.
|
|
240 |
|
241 |
| Metric | Value |
|
242 |
|:--------------------|:-----------|
|
243 |
-
| **cosine_accuracy** | **0.
|
244 |
-
| dot_accuracy | 0.
|
245 |
-
| manhattan_accuracy | 0.
|
246 |
-
| euclidean_accuracy | 0.
|
247 |
-
| max_accuracy | 0.
|
248 |
|
249 |
#### Semantic Similarity
|
250 |
|
@@ -252,16 +302,35 @@ You can finetune this model on your own dataset.
|
|
252 |
|
253 |
| Metric | Value |
|
254 |
|:--------------------|:-----------|
|
255 |
-
| pearson_cosine | 0.
|
256 |
-
| **spearman_cosine** | **0.
|
257 |
-
| pearson_manhattan | 0.
|
258 |
-
| spearman_manhattan | 0.
|
259 |
-
| pearson_euclidean | 0.
|
260 |
-
| spearman_euclidean | 0.
|
261 |
-
| pearson_dot | 0.
|
262 |
-
| spearman_dot | 0.
|
263 |
-
| pearson_max | 0.
|
264 |
-
| spearman_max | 0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
265 |
|
266 |
<!--
|
267 |
## Bias, Risks and Limitations
|
@@ -277,121 +346,73 @@ You can finetune this model on your own dataset.
|
|
277 |
|
278 |
## Training Details
|
279 |
|
280 |
-
### Training
|
281 |
|
282 |
#### triplets
|
283 |
|
284 |
* Dataset: triplets
|
285 |
-
* Size:
|
286 |
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
287 |
* Approximate statistics based on the first 1000 samples:
|
288 |
-
| | anchor | positive
|
289 |
-
|
290 |
-
| type | string | string
|
291 |
-
| details | <ul><li>min: 7 tokens</li><li>mean: 11.
|
292 |
-
* Samples:
|
293 |
-
| anchor | positive | negative |
|
294 |
-
|:----------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
295 |
-
| <code>search_query: tarps heavy duty waterproof 8x10</code> | <code>search_document: 8' x 10' Super Heavy Duty 16 Mil Brown Poly Tarp Cover - Thick Waterproof, UV Resistant, Rip and Tear Proof Tarpaulin with Grommets and Reinforced Edges - by Xpose Safety, Xpose Safety, Brown</code> | <code>search_document: Grillkid 6'X8' 4.5 Mil Thick General Purpose Waterproof Poly Tarp, Grillkid, All Purpose</code> |
|
296 |
-
| <code>search_query: wireless keyboard without number pad</code> | <code>search_document: Macally 2.4G Small Wireless Keyboard - Ergonomic & Comfortable Computer Keyboard - Compact Keyboard for Laptop or Windows PC Desktop, Tablet, Smart TV - Plug & Play Mini Keyboard with 12 Hot Keys, Macally, Black</code> | <code>search_document: Wireless Keyboard - iClever GKA22S Rechargeable Keyboard with Number Pad, Full-Size Stainless Steel Ultra Slim Keyboard, 2.4G Stable Connection Wireless Keyboard for iMac, Mackbook, PC, Laptop, iClever, Silver</code> |
|
297 |
-
| <code>search_query: geometry earrings</code> | <code>search_document: Simple Stud Earrings for Women, Geometric Minimalist Stud Earring Set Tiny Circle Triangle Square Bar Stud Earrings Mini Cartilage Tragus Earrings, choice of all, B:Circle Sliver</code> | <code>search_document: BONALUNA Bohemian Wood And Marble Effect Oblong Shaped Drop Statement Earrings (VIVID TURQUOISE), BONALUNA, VIVID TURQUOISE</code> |
|
298 |
-
* Loss: [<code>CachedMultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativesrankingloss) with these parameters:
|
299 |
-
```json
|
300 |
-
{
|
301 |
-
"scale": 20.0,
|
302 |
-
"similarity_fct": "cos_sim"
|
303 |
-
}
|
304 |
-
```
|
305 |
-
|
306 |
-
#### pairs
|
307 |
-
|
308 |
-
* Dataset: pairs
|
309 |
-
* Size: 498,114 training samples
|
310 |
-
* Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>score</code>
|
311 |
-
* Approximate statistics based on the first 1000 samples:
|
312 |
-
| | sentence1 | sentence2 | score |
|
313 |
-
|:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:---------------------------------------------------------------|
|
314 |
-
| type | string | string | float |
|
315 |
-
| details | <ul><li>min: 3 tokens</li><li>mean: 6.73 tokens</li><li>max: 33 tokens</li></ul> | <ul><li>min: 10 tokens</li><li>mean: 40.14 tokens</li><li>max: 98 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.81</li><li>max: 1.0</li></ul> |
|
316 |
* Samples:
|
317 |
-
|
|
318 |
-
|
319 |
-
| <code>
|
320 |
-
| <code>
|
321 |
-
| <code>
|
322 |
-
* Loss: [<code>
|
323 |
```json
|
324 |
{
|
325 |
-
"
|
326 |
-
"
|
327 |
}
|
328 |
```
|
329 |
|
330 |
-
### Evaluation
|
331 |
|
332 |
#### triplets
|
333 |
|
334 |
* Dataset: triplets
|
335 |
-
* Size:
|
336 |
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
337 |
* Approximate statistics based on the first 1000 samples:
|
338 |
-
| | anchor | positive
|
339 |
-
|
340 |
-
| type | string | string
|
341 |
-
| details | <ul><li>min: 7 tokens</li><li>mean: 11.
|
342 |
-
* Samples:
|
343 |
-
| anchor | positive | negative |
|
344 |
-
|:-------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
345 |
-
| <code>search_query: hitch fifth wheel</code> | <code>search_document: ENIXWILL 5th Wheel Trailer Hitch Lifting Device Bracket Pin Fit for Hitch Companion and Patriot Series Hitch, ENIXWILL, Black</code> | <code>search_document: ECOTRIC Fifth 5th Wheel Trailer Hitch Mount Rails and Installation Kits for Full-Size Trucks, ECOTRIC, black</code> |
|
346 |
-
| <code>search_query: dek pro</code> | <code>search_document: Cubiker Computer Desk 47 inch Home Office Writing Study Desk, Modern Simple Style Laptop Table with Storage Bag, Brown, Cubiker, Brown</code> | <code>search_document: FEZIBO Dual Motor L Shaped Electric Standing Desk, 48 Inches Stand Up Corner Desk, Home Office Sit Stand Desk with Rustic Brown Top and Black Frame, FEZIBO, Rustic Brown</code> |
|
347 |
-
| <code>search_query: 1 year baby mouth without teeth cleaner</code> | <code>search_document: Baby Toothbrush,Infant Toothbrush,Baby Tongue Cleaner,Infant Toothbrush,Baby Tongue Cleaner Newborn,Toothbrush Tongue Cleaner Dental Care for 0-36 Month Baby,36 Pcs + Free 4 Pcs, Babycolor, Blue</code> | <code>search_document: Slotic Baby Toothbrush for 0-2 Years, Safe and Sturdy, Toddler Oral Care Teether Brush, Extra Soft Bristle for Baby Teeth and Infant Gums, Dentist Recommended (4-Pack), Slotic, 4 Pack</code> |
|
348 |
-
* Loss: [<code>CachedMultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativesrankingloss) with these parameters:
|
349 |
-
```json
|
350 |
-
{
|
351 |
-
"scale": 20.0,
|
352 |
-
"similarity_fct": "cos_sim"
|
353 |
-
}
|
354 |
-
```
|
355 |
-
|
356 |
-
#### pairs
|
357 |
-
|
358 |
-
* Dataset: pairs
|
359 |
-
* Size: 10,000 evaluation samples
|
360 |
-
* Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>score</code>
|
361 |
-
* Approximate statistics based on the first 1000 samples:
|
362 |
-
| | sentence1 | sentence2 | score |
|
363 |
-
|:--------|:--------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------|
|
364 |
-
| type | string | string | float |
|
365 |
-
| details | <ul><li>min: 3 tokens</li><li>mean: 6.8 tokens</li><li>max: 34 tokens</li></ul> | <ul><li>min: 9 tokens</li><li>mean: 39.7 tokens</li><li>max: 101 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.77</li><li>max: 1.0</li></ul> |
|
366 |
* Samples:
|
367 |
-
|
|
368 |
-
|
369 |
-
| <code>
|
370 |
-
| <code>
|
371 |
-
| <code>
|
372 |
-
* Loss: [<code>
|
373 |
```json
|
374 |
{
|
375 |
-
"
|
376 |
-
"
|
377 |
}
|
378 |
```
|
379 |
|
380 |
### Training Hyperparameters
|
381 |
#### Non-Default Hyperparameters
|
382 |
|
383 |
-
- `per_device_train_batch_size`:
|
384 |
-
- `per_device_eval_batch_size`:
|
385 |
- `gradient_accumulation_steps`: 2
|
386 |
-
- `learning_rate`: 1e-
|
387 |
-
- `lr_scheduler_type`:
|
388 |
-
- `lr_scheduler_kwargs`: {'
|
389 |
-
- `warmup_ratio`: 0.
|
390 |
-
- `dataloader_drop_last`: True
|
391 |
- `dataloader_num_workers`: 4
|
392 |
- `dataloader_prefetch_factor`: 4
|
393 |
- `load_best_model_at_end`: True
|
394 |
- `gradient_checkpointing`: True
|
|
|
395 |
- `batch_sampler`: no_duplicates
|
396 |
|
397 |
#### All Hyperparameters
|
@@ -400,13 +421,13 @@ You can finetune this model on your own dataset.
|
|
400 |
- `overwrite_output_dir`: False
|
401 |
- `do_predict`: False
|
402 |
- `prediction_loss_only`: True
|
403 |
-
- `per_device_train_batch_size`:
|
404 |
-
- `per_device_eval_batch_size`:
|
405 |
- `per_gpu_train_batch_size`: None
|
406 |
- `per_gpu_eval_batch_size`: None
|
407 |
- `gradient_accumulation_steps`: 2
|
408 |
- `eval_accumulation_steps`: None
|
409 |
-
- `learning_rate`: 1e-
|
410 |
- `weight_decay`: 0.0
|
411 |
- `adam_beta1`: 0.9
|
412 |
- `adam_beta2`: 0.999
|
@@ -414,9 +435,9 @@ You can finetune this model on your own dataset.
|
|
414 |
- `max_grad_norm`: 1.0
|
415 |
- `num_train_epochs`: 3
|
416 |
- `max_steps`: -1
|
417 |
-
- `lr_scheduler_type`:
|
418 |
-
- `lr_scheduler_kwargs`: {'
|
419 |
-
- `warmup_ratio`: 0.
|
420 |
- `warmup_steps`: 0
|
421 |
- `log_level`: passive
|
422 |
- `log_level_replica`: warning
|
@@ -485,7 +506,7 @@ You can finetune this model on your own dataset.
|
|
485 |
- `push_to_hub_model_id`: None
|
486 |
- `push_to_hub_organization`: None
|
487 |
- `mp_parameters`:
|
488 |
-
- `auto_find_batch_size`:
|
489 |
- `full_determinism`: False
|
490 |
- `torchdynamo`: None
|
491 |
- `ray_scope`: last
|
@@ -506,378 +527,208 @@ You can finetune this model on your own dataset.
|
|
506 |
### Training Logs
|
507 |
<details><summary>Click to expand</summary>
|
508 |
|
509 |
-
| Epoch | Step
|
510 |
-
|
511 |
-
| 0.
|
512 |
-
| 0.
|
513 |
-
| 0.
|
514 |
-
| 0.
|
515 |
-
| 0.
|
516 |
-
| 0.
|
517 |
-
| 0.
|
518 |
-
| 0.
|
519 |
-
| 0.
|
520 |
-
| 0.
|
521 |
-
| 0.
|
522 |
-
| 0.
|
523 |
-
| 0.
|
524 |
-
| 0.
|
525 |
-
| 0.
|
526 |
-
| 0.
|
527 |
-
| 0.
|
528 |
-
| 0.
|
529 |
-
| 0.
|
530 |
-
| 0.
|
531 |
-
| 0.
|
532 |
-
| 0.
|
533 |
-
| 0.
|
534 |
-
| 0.
|
535 |
-
| 0.
|
536 |
-
| 0.
|
537 |
-
| 0.
|
538 |
-
| 0.
|
539 |
-
| 0.
|
540 |
-
| 0.
|
541 |
-
| 0.
|
542 |
-
| 0.
|
543 |
-
| 0.
|
544 |
-
| 0.
|
545 |
-
| 0.
|
546 |
-
| 0.
|
547 |
-
| 0.
|
548 |
-
| 0.
|
549 |
-
| 0.
|
550 |
-
| 0.
|
551 |
-
| 0.
|
552 |
-
| 0.
|
553 |
-
| 0.
|
554 |
-
| 0.
|
555 |
-
| 0.
|
556 |
-
| 0.
|
557 |
-
| 0.
|
558 |
-
| 0.
|
559 |
-
| 0.
|
560 |
-
| 0.
|
561 |
-
| 0.
|
562 |
-
| 0.
|
563 |
-
| 0.
|
564 |
-
| 0.
|
565 |
-
| 0.
|
566 |
-
| 0.
|
567 |
-
| 0.
|
568 |
-
| 0.
|
569 |
-
| 0.
|
570 |
-
| 0.
|
571 |
-
| 0.
|
572 |
-
| 0.
|
573 |
-
| 0.
|
574 |
-
| 0.
|
575 |
-
| 0.
|
576 |
-
| 0.
|
577 |
-
| 0.
|
578 |
-
| 0.
|
579 |
-
| 0.
|
580 |
-
| 0.
|
581 |
-
| 0.
|
582 |
-
| 0.
|
583 |
-
| 0.
|
584 |
-
| 0.
|
585 |
-
| 0.
|
586 |
-
| 0.
|
587 |
-
| 0.
|
588 |
-
| 0.
|
589 |
-
| 0.
|
590 |
-
| 0.
|
591 |
-
| 0.
|
592 |
-
| 0.
|
593 |
-
| 0.
|
594 |
-
| 0.
|
595 |
-
| 0.
|
596 |
-
| 0.
|
597 |
-
| 0.
|
598 |
-
| 0.
|
599 |
-
| 0.
|
600 |
-
| 0.
|
601 |
-
| 0.
|
602 |
-
| 0.
|
603 |
-
| 0.
|
604 |
-
| 0.
|
605 |
-
| 0.
|
606 |
-
| 0.
|
607 |
-
| 0.
|
608 |
-
| 0.
|
609 |
-
| 0.
|
610 |
-
| 0.
|
611 |
-
| 0.
|
612 |
-
| 0.
|
613 |
-
| 0.
|
614 |
-
| 0.
|
615 |
-
| 0.
|
616 |
-
| 0.
|
617 |
-
| 0.
|
618 |
-
| 0.
|
619 |
-
| 0.
|
620 |
-
| 0.
|
621 |
-
| 0.
|
622 |
-
| 0.
|
623 |
-
| 0.
|
624 |
-
| 0.
|
625 |
-
| 0.
|
626 |
-
| 0.
|
627 |
-
| 0.
|
628 |
-
| 0.
|
629 |
-
| 0.
|
630 |
-
| 0.
|
631 |
-
| 0.
|
632 |
-
| 0.
|
633 |
-
| 0.
|
634 |
-
| 0.
|
635 |
-
| 0.
|
636 |
-
| 0.
|
637 |
-
| 0.
|
638 |
-
| 0.
|
639 |
-
| 0.
|
640 |
-
| 0.
|
641 |
-
| 0.
|
642 |
-
| 0.
|
643 |
-
| 0.
|
644 |
-
| 0.
|
645 |
-
| 0.
|
646 |
-
| 0.
|
647 |
-
| 0.
|
648 |
-
| 0.
|
649 |
-
| 0.
|
650 |
-
| 0.
|
651 |
-
| 0.
|
652 |
-
| 0.
|
653 |
-
| 0.
|
654 |
-
| 0.
|
655 |
-
| 0.
|
656 |
-
| 0.
|
657 |
-
| 0.
|
658 |
-
| 0.
|
659 |
-
| 0.
|
660 |
-
| 0.
|
661 |
-
| 0.
|
662 |
-
| 0.
|
663 |
-
| 0.
|
664 |
-
| 0.
|
665 |
-
| 0.
|
666 |
-
| 0.
|
667 |
-
| 0.
|
668 |
-
| 0.
|
669 |
-
| 0.
|
670 |
-
| 0.
|
671 |
-
| 0.
|
672 |
-
| 0.
|
673 |
-
| 0.
|
674 |
-
| 0.
|
675 |
-
| 0.
|
676 |
-
| 0.
|
677 |
-
| 0.
|
678 |
-
| 0.
|
679 |
-
| 0.
|
680 |
-
| 0.
|
681 |
-
| 0.
|
682 |
-
| 0.
|
683 |
-
| 0.
|
684 |
-
| 0.
|
685 |
-
| 0.
|
686 |
-
| 0.
|
687 |
-
| 0.
|
688 |
-
| 0.
|
689 |
-
| 0.
|
690 |
-
| 0.
|
691 |
-
| 0.
|
692 |
-
| 0.
|
693 |
-
| 0.
|
694 |
-
| 0.
|
695 |
-
| 0.
|
696 |
-
| 0.
|
697 |
-
| 0.
|
698 |
-
| 0.
|
699 |
-
| 0.
|
700 |
-
| 0.
|
701 |
-
| 0.
|
702 |
-
| 0.
|
703 |
-
| 0.
|
704 |
-
| 0.
|
705 |
-
| 0.
|
706 |
-
| 0.
|
707 |
-
| 0.
|
708 |
-
| 0.
|
709 |
-
| 0.
|
710 |
-
| 0.
|
711 |
-
| 0.5441 | 20100 | 1.634 | - | - | - | - |
|
712 |
-
| 0.5468 | 20200 | 1.7865 | - | - | - | - |
|
713 |
-
| 0.5495 | 20300 | 1.8573 | - | - | - | - |
|
714 |
-
| 0.5522 | 20400 | 1.5575 | - | - | - | - |
|
715 |
-
| 0.5549 | 20500 | 1.6594 | - | - | - | - |
|
716 |
-
| 0.5576 | 20600 | 1.8793 | - | - | - | - |
|
717 |
-
| 0.5603 | 20700 | 1.7643 | - | - | - | - |
|
718 |
-
| 0.5630 | 20800 | 1.538 | - | - | - | - |
|
719 |
-
| 0.5657 | 20900 | 1.8634 | - | - | - | - |
|
720 |
-
| 0.5684 | 21000 | 1.916 | 3.7223 | 0.6982 | 0.7258 | 0.4650 |
|
721 |
-
| 0.5711 | 21100 | 1.5947 | - | - | - | - |
|
722 |
-
| 0.5738 | 21200 | 1.5321 | - | - | - | - |
|
723 |
-
| 0.5766 | 21300 | 1.7004 | - | - | - | - |
|
724 |
-
| 0.5793 | 21400 | 1.6947 | - | - | - | - |
|
725 |
-
| 0.5820 | 21500 | 1.5228 | - | - | - | - |
|
726 |
-
| 0.5847 | 21600 | 1.7152 | - | - | - | - |
|
727 |
-
| 0.5874 | 21700 | 1.6883 | - | - | - | - |
|
728 |
-
| 0.5901 | 21800 | 1.6779 | - | - | - | - |
|
729 |
-
| 0.5928 | 21900 | 1.7323 | - | - | - | - |
|
730 |
-
| 0.5955 | 22000 | 1.9633 | 3.7266 | 0.6996 | 0.7288 | 0.4635 |
|
731 |
-
| 0.5982 | 22100 | 1.7498 | - | - | - | - |
|
732 |
-
| 0.6009 | 22200 | 1.7513 | - | - | - | - |
|
733 |
-
| 0.6036 | 22300 | 1.7078 | - | - | - | - |
|
734 |
-
| 0.6063 | 22400 | 1.6438 | - | - | - | - |
|
735 |
-
| 0.6090 | 22500 | 1.6743 | - | - | - | - |
|
736 |
-
| 0.6117 | 22600 | 1.6701 | - | - | - | - |
|
737 |
-
| 0.6145 | 22700 | 1.7871 | - | - | - | - |
|
738 |
-
| 0.6172 | 22800 | 1.6247 | - | - | - | - |
|
739 |
-
| 0.6199 | 22900 | 1.7817 | - | - | - | - |
|
740 |
-
| 0.6226 | 23000 | 1.6606 | 3.7321 | 0.6993 | 0.7286 | 0.4614 |
|
741 |
-
| 0.6253 | 23100 | 1.8987 | - | - | - | - |
|
742 |
-
| 0.6280 | 23200 | 1.6494 | - | - | - | - |
|
743 |
-
| 0.6307 | 23300 | 1.6776 | - | - | - | - |
|
744 |
-
| 0.6334 | 23400 | 1.75 | - | - | - | - |
|
745 |
-
| 0.6361 | 23500 | 1.5131 | - | - | - | - |
|
746 |
-
| 0.6388 | 23600 | 1.7946 | - | - | - | - |
|
747 |
-
| 0.6415 | 23700 | 1.665 | - | - | - | - |
|
748 |
-
| 0.6442 | 23800 | 1.6681 | - | - | - | - |
|
749 |
-
| 0.6469 | 23900 | 1.8255 | - | - | - | - |
|
750 |
-
| 0.6496 | 24000 | 1.6759 | 3.7227 | 0.7017 | 0.7281 | 0.4625 |
|
751 |
-
| 0.6523 | 24100 | 1.554 | - | - | - | - |
|
752 |
-
| 0.6551 | 24200 | 1.6435 | - | - | - | - |
|
753 |
-
| 0.6578 | 24300 | 1.8224 | - | - | - | - |
|
754 |
-
| 0.6605 | 24400 | 1.6186 | - | - | - | - |
|
755 |
-
| 0.6632 | 24500 | 1.7156 | - | - | - | - |
|
756 |
-
| 0.6659 | 24600 | 1.5247 | - | - | - | - |
|
757 |
-
| 0.6686 | 24700 | 1.6264 | - | - | - | - |
|
758 |
-
| 0.6713 | 24800 | 1.7673 | - | - | - | - |
|
759 |
-
| 0.6740 | 24900 | 1.8072 | - | - | - | - |
|
760 |
-
| 0.6767 | 25000 | 1.765 | 3.7407 | 0.7026 | 0.7283 | 0.4589 |
|
761 |
-
| 0.6794 | 25100 | 1.6422 | - | - | - | - |
|
762 |
-
| 0.6821 | 25200 | 1.7846 | - | - | - | - |
|
763 |
-
| 0.6848 | 25300 | 1.7366 | - | - | - | - |
|
764 |
-
| 0.6875 | 25400 | 1.7839 | - | - | - | - |
|
765 |
-
| 0.6902 | 25500 | 1.441 | - | - | - | - |
|
766 |
-
| 0.6930 | 25600 | 1.5533 | - | - | - | - |
|
767 |
-
| 0.6957 | 25700 | 1.6922 | - | - | - | - |
|
768 |
-
| 0.6984 | 25800 | 1.5544 | - | - | - | - |
|
769 |
-
| 0.7011 | 25900 | 1.456 | - | - | - | - |
|
770 |
-
| 0.7038 | 26000 | 1.6494 | 3.7274 | 0.7059 | 0.7268 | 0.4661 |
|
771 |
-
| 0.7065 | 26100 | 1.6963 | - | - | - | - |
|
772 |
-
| 0.7092 | 26200 | 1.7892 | - | - | - | - |
|
773 |
-
| 0.7119 | 26300 | 1.6669 | - | - | - | - |
|
774 |
-
| 0.7146 | 26400 | 1.6758 | - | - | - | - |
|
775 |
-
| 0.7173 | 26500 | 1.6322 | - | - | - | - |
|
776 |
-
| 0.7200 | 26600 | 1.5416 | - | - | - | - |
|
777 |
-
| 0.7227 | 26700 | 1.681 | - | - | - | - |
|
778 |
-
| 0.7254 | 26800 | 1.5159 | - | - | - | - |
|
779 |
-
| 0.7281 | 26900 | 1.715 | - | - | - | - |
|
780 |
-
| 0.7308 | 27000 | 1.6164 | 3.7456 | 0.7061 | 0.7257 | 0.4570 |
|
781 |
-
| 0.7336 | 27100 | 1.6784 | - | - | - | - |
|
782 |
-
| 0.7363 | 27200 | 1.5886 | - | - | - | - |
|
783 |
-
| 0.7390 | 27300 | 1.6736 | - | - | - | - |
|
784 |
-
| 0.7417 | 27400 | 1.5659 | - | - | - | - |
|
785 |
-
| 0.7444 | 27500 | 1.6552 | - | - | - | - |
|
786 |
-
| 0.7471 | 27600 | 1.5672 | - | - | - | - |
|
787 |
-
| 0.7498 | 27700 | 1.5873 | - | - | - | - |
|
788 |
-
| 0.7525 | 27800 | 1.6746 | - | - | - | - |
|
789 |
-
| 0.7552 | 27900 | 1.7503 | - | - | - | - |
|
790 |
-
| 0.7579 | 28000 | 1.7287 | 3.7390 | 0.7076 | 0.7244 | 0.4636 |
|
791 |
-
| 0.7606 | 28100 | 1.6216 | - | - | - | - |
|
792 |
-
| 0.7633 | 28200 | 1.6101 | - | - | - | - |
|
793 |
-
| 0.7660 | 28300 | 1.5651 | - | - | - | - |
|
794 |
-
| 0.7687 | 28400 | 1.5659 | - | - | - | - |
|
795 |
-
| 0.7714 | 28500 | 1.5248 | - | - | - | - |
|
796 |
-
| 0.7742 | 28600 | 1.3725 | - | - | - | - |
|
797 |
-
| 0.7769 | 28700 | 1.7881 | - | - | - | - |
|
798 |
-
| 0.7796 | 28800 | 1.739 | - | - | - | - |
|
799 |
-
| 0.7823 | 28900 | 1.6464 | - | - | - | - |
|
800 |
-
| 0.7850 | 29000 | 1.6841 | 3.7212 | 0.7073 | 0.7247 | 0.4626 |
|
801 |
-
| 0.7877 | 29100 | 1.6254 | - | - | - | - |
|
802 |
-
| 0.7904 | 29200 | 1.6728 | - | - | - | - |
|
803 |
-
| 0.7931 | 29300 | 1.5605 | - | - | - | - |
|
804 |
-
| 0.7958 | 29400 | 1.687 | - | - | - | - |
|
805 |
-
| 0.7985 | 29500 | 1.7799 | - | - | - | - |
|
806 |
-
| 0.8012 | 29600 | 1.6792 | - | - | - | - |
|
807 |
-
| 0.8039 | 29700 | 1.5241 | - | - | - | - |
|
808 |
-
| 0.8066 | 29800 | 1.6341 | - | - | - | - |
|
809 |
-
| 0.8093 | 29900 | 1.5571 | - | - | - | - |
|
810 |
-
| 0.8121 | 30000 | 1.5228 | 3.7397 | 0.7105 | 0.7234 | 0.4682 |
|
811 |
-
| 0.8148 | 30100 | 1.5988 | - | - | - | - |
|
812 |
-
| 0.8175 | 30200 | 1.4222 | - | - | - | - |
|
813 |
-
| 0.8202 | 30300 | 1.4629 | - | - | - | - |
|
814 |
-
| 0.8229 | 30400 | 1.6381 | - | - | - | - |
|
815 |
-
| 0.8256 | 30500 | 1.4585 | - | - | - | - |
|
816 |
-
| 0.8283 | 30600 | 1.6774 | - | - | - | - |
|
817 |
-
| 0.8310 | 30700 | 1.811 | - | - | - | - |
|
818 |
-
| 0.8337 | 30800 | 1.5872 | - | - | - | - |
|
819 |
-
| 0.8364 | 30900 | 1.4762 | - | - | - | - |
|
820 |
-
| 0.8391 | 31000 | 1.7079 | 3.7256 | 0.7128 | 0.7215 | 0.4645 |
|
821 |
-
| 0.8418 | 31100 | 1.4948 | - | - | - | - |
|
822 |
-
| 0.8445 | 31200 | 1.4556 | - | - | - | - |
|
823 |
-
| 0.8472 | 31300 | 1.5191 | - | - | - | - |
|
824 |
-
| 0.8499 | 31400 | 1.598 | - | - | - | - |
|
825 |
-
| 0.8527 | 31500 | 1.6586 | - | - | - | - |
|
826 |
-
| 0.8554 | 31600 | 1.6893 | - | - | - | - |
|
827 |
-
| 0.8581 | 31700 | 1.7764 | - | - | - | - |
|
828 |
-
| 0.8608 | 31800 | 1.3632 | - | - | - | - |
|
829 |
-
| 0.8635 | 31900 | 1.6681 | - | - | - | - |
|
830 |
-
| 0.8662 | 32000 | 1.6232 | 3.7358 | 0.7161 | 0.7232 | 0.4651 |
|
831 |
-
| 0.8689 | 32100 | 1.4556 | - | - | - | - |
|
832 |
-
| 0.8716 | 32200 | 1.8698 | - | - | - | - |
|
833 |
-
| 0.8743 | 32300 | 1.7566 | - | - | - | - |
|
834 |
-
| 0.8770 | 32400 | 1.6082 | - | - | - | - |
|
835 |
-
| 0.8797 | 32500 | 1.6465 | - | - | - | - |
|
836 |
-
| 0.8824 | 32600 | 1.5018 | - | - | - | - |
|
837 |
-
| 0.8851 | 32700 | 1.8482 | - | - | - | - |
|
838 |
-
| 0.8878 | 32800 | 1.5147 | - | - | - | - |
|
839 |
-
| 0.8905 | 32900 | 1.699 | - | - | - | - |
|
840 |
-
| 0.8933 | 33000 | 1.5738 | 3.7323 | 0.7176 | 0.7246 | 0.4657 |
|
841 |
-
| 0.8960 | 33100 | 1.635 | - | - | - | - |
|
842 |
-
| 0.8987 | 33200 | 1.7069 | - | - | - | - |
|
843 |
-
| 0.9014 | 33300 | 1.6272 | - | - | - | - |
|
844 |
-
| 0.9041 | 33400 | 1.7648 | - | - | - | - |
|
845 |
-
| 0.9068 | 33500 | 1.6683 | - | - | - | - |
|
846 |
-
| 0.9095 | 33600 | 1.4867 | - | - | - | - |
|
847 |
-
| 0.9122 | 33700 | 1.6677 | - | - | - | - |
|
848 |
-
| 0.9149 | 33800 | 1.5527 | - | - | - | - |
|
849 |
-
| 0.9176 | 33900 | 1.6804 | - | - | - | - |
|
850 |
-
| 0.9203 | 34000 | 1.425 | 3.7477 | 0.7172 | 0.7231 | 0.4596 |
|
851 |
-
| 0.9230 | 34100 | 1.771 | - | - | - | - |
|
852 |
-
| 0.9257 | 34200 | 1.5767 | - | - | - | - |
|
853 |
-
| 0.9284 | 34300 | 1.5424 | - | - | - | - |
|
854 |
-
| 0.9312 | 34400 | 1.5985 | - | - | - | - |
|
855 |
-
| 0.9339 | 34500 | 1.6763 | - | - | - | - |
|
856 |
-
| 0.9366 | 34600 | 1.6608 | - | - | - | - |
|
857 |
-
| 0.9393 | 34700 | 1.7736 | - | - | - | - |
|
858 |
-
| 0.9420 | 34800 | 1.8955 | - | - | - | - |
|
859 |
-
| 0.9447 | 34900 | 1.5688 | - | - | - | - |
|
860 |
-
| 0.9474 | 35000 | 1.6123 | 3.7410 | 0.7196 | 0.7226 | 0.4671 |
|
861 |
-
| 0.9501 | 35100 | 1.7264 | - | - | - | - |
|
862 |
-
| 0.9528 | 35200 | 1.5511 | - | - | - | - |
|
863 |
-
| 0.9555 | 35300 | 1.6409 | - | - | - | - |
|
864 |
-
| 0.9582 | 35400 | 1.47 | - | - | - | - |
|
865 |
-
| 0.9609 | 35500 | 1.8675 | - | - | - | - |
|
866 |
-
| 0.9636 | 35600 | 1.6868 | - | - | - | - |
|
867 |
-
| 0.9663 | 35700 | 1.744 | - | - | - | - |
|
868 |
-
| 0.9690 | 35800 | 1.6734 | - | - | - | - |
|
869 |
-
| 0.9718 | 35900 | 1.4154 | - | - | - | - |
|
870 |
-
| 0.9745 | 36000 | 1.4793 | 3.7393 | 0.7190 | 0.7223 | 0.4677 |
|
871 |
-
| 0.9772 | 36100 | 1.7126 | - | - | - | - |
|
872 |
-
| 0.9799 | 36200 | 1.7037 | - | - | - | - |
|
873 |
-
| 0.9826 | 36300 | 1.6306 | - | - | - | - |
|
874 |
-
| 0.9853 | 36400 | 1.7783 | - | - | - | - |
|
875 |
-
| 0.9880 | 36500 | 1.5751 | - | - | - | - |
|
876 |
-
| 0.9907 | 36600 | 1.6079 | - | - | - | - |
|
877 |
-
| 0.9934 | 36700 | 1.7162 | - | - | - | - |
|
878 |
-
| 0.9961 | 36800 | 1.447 | - | - | - | - |
|
879 |
-
| 0.9988 | 36900 | 1.6155 | - | - | - | - |
|
880 |
-
| 1.0015 | 37000 | 1.7294 | 3.7512 | 0.7177 | 0.7236 | 0.4659 |
|
881 |
|
882 |
</details>
|
883 |
|
@@ -907,27 +758,15 @@ You can finetune this model on your own dataset.
|
|
907 |
}
|
908 |
```
|
909 |
|
910 |
-
####
|
911 |
-
```bibtex
|
912 |
-
@misc{gao2021scaling,
|
913 |
-
title={Scaling Deep Contrastive Learning Batch Size under Memory Limited Setup},
|
914 |
-
author={Luyu Gao and Yunyi Zhang and Jiawei Han and Jamie Callan},
|
915 |
-
year={2021},
|
916 |
-
eprint={2101.06983},
|
917 |
-
archivePrefix={arXiv},
|
918 |
-
primaryClass={cs.LG}
|
919 |
-
}
|
920 |
-
```
|
921 |
-
|
922 |
-
#### AnglELoss
|
923 |
```bibtex
|
924 |
-
@misc{
|
925 |
-
title={
|
926 |
-
author={
|
927 |
-
year={
|
928 |
-
eprint={
|
929 |
archivePrefix={arXiv},
|
930 |
-
primaryClass={cs.
|
931 |
}
|
932 |
```
|
933 |
|
|
|
6 |
- sentence-similarity
|
7 |
- feature-extraction
|
8 |
- generated_from_trainer
|
9 |
+
- dataset_size:1500000
|
10 |
+
- loss:TripletLoss
|
|
|
|
|
11 |
datasets: []
|
12 |
metrics:
|
13 |
- cosine_accuracy
|
|
|
25 |
- spearman_dot
|
26 |
- pearson_max
|
27 |
- spearman_max
|
28 |
+
- cosine_accuracy@10
|
29 |
+
- cosine_precision@10
|
30 |
+
- cosine_recall@10
|
31 |
+
- cosine_ndcg@10
|
32 |
+
- cosine_mrr@10
|
33 |
+
- cosine_map@10
|
34 |
+
- dot_accuracy@10
|
35 |
+
- dot_precision@10
|
36 |
+
- dot_recall@10
|
37 |
+
- dot_ndcg@10
|
38 |
+
- dot_mrr@10
|
39 |
+
- dot_map@10
|
40 |
widget:
|
41 |
+
- source_sentence: 'search_query: speeder man slippers for boys size 12'
|
42 |
sentences:
|
43 |
+
- 'search_document: Marvel Boys Spider-Man Bootie Slippers (11-12 M US Little Kid),
|
44 |
+
BBC International, Red/Blue'
|
45 |
+
- 'search_document: MIXIN Little Kids Boys Spring Cozy Comfort Comfy Slippers Size
|
46 |
+
Dark Blue 13 13.5 M, MIXIN, Blue'
|
47 |
+
- 'search_document: Third Eye Blind, , '
|
48 |
+
- source_sentence: 'search_query: water filter refrigerator whirlpool'
|
|
|
|
|
|
|
49 |
sentences:
|
50 |
+
- 'search_document: Assorted Tootsie Frooties - 2 lbs of Delicious Assorted Bulk
|
51 |
+
Wrapped Candy with Refrigerator Magnet, Emporium Candy, '
|
52 |
+
- 'search_document: Whirlpool WHER25 Reverse Osmosis (RO) Filtration System With
|
53 |
+
Chrome Faucet | Extra Long Life | Easy To Replace UltraEase Filter Cartridges,
|
54 |
+
White, Whirlpool, White'
|
55 |
+
- 'search_document: Everydrop by Whirlpool Ice and Water Refrigerator Filter 3,
|
56 |
+
EDR3RXD1, Single-Pack, White, EveryDrop by Whirlpool, White'
|
57 |
+
- source_sentence: 'search_query: t towels kitchen'
|
|
|
58 |
sentences:
|
59 |
+
- 'search_document: Utopia Kitchen Flour Sack Tea Towels 24 Pack, 28" x 28" Ring
|
60 |
+
Spun 100% Cotton Dish Cloths - Machine Washable - for Cleaning & Drying - White,
|
61 |
+
Utopia Kitchen, White'
|
62 |
+
- 'search_document: Rael Certified Organic Cotton Panty Liners (Regular, 44 Count)
|
63 |
+
& Extra Long Overnight Pads (12 Count) Bundle, Rael, '
|
64 |
+
- 'search_document: LSK Bath Hand Towel Holder Standing,T-Shape Towel Rack Countertop
|
65 |
+
with Round Base for Kitchen Bathroom Toliet ,Rustproof Stainless Steel (Black),
|
66 |
+
LSK, Black'
|
67 |
+
- source_sentence: 'search_query: cozy lights bedroom'
|
68 |
sentences:
|
69 |
+
- 'search_document: Fluorescent Light Covers | Fluorescent Light Covers for Ceiling
|
70 |
+
Lights, Classroom, Office, or Blue Light Covers Fluorescent Filter- Eliminates
|
71 |
+
Flicker & Glare - 48" by 24" (4 Pack, Sky Blue Panel), Everyday Educate, Blue'
|
72 |
+
- 'search_document: 43ft Led Globe String Lights, 100LEDs Outdoor Indoor String
|
73 |
+
Lights Plug in, 8 Modes Twinkle Lights, Warm White, Fairy String Lights Christmas
|
74 |
+
Decoration, Yard, Party, Bedroom, Yinuo Candle, White'
|
75 |
+
- 'search_document: PARMIDA 6 inch Dimmable LED Square Recessed Retrofit Lighting,
|
76 |
+
Easy Downlight Installation, 12W (100W Eqv.), 950lm, Ceiling Can Lights, Energy
|
77 |
+
Star & ETL-Listed, 5 Year Warranty, 3000K - 4 Pack, Parmida LED Technologies,
|
78 |
+
3000k (Soft White)'
|
79 |
+
- source_sentence: 'search_query: punxhing bag'
|
80 |
sentences:
|
81 |
+
- 'search_document: Power Core Bag (EA), Everlast, Black/White'
|
82 |
+
- 'search_document: Nifeida High Mount 3rd Brake Light for 2004-2008 Ford F150,
|
83 |
+
2007-2010 Ford Explorer, 2006-2008 Lincoln Mark LT 22LEDs Waterproof Wedge Center
|
84 |
+
Tail Light Heavy Duty Third Stop Light Cargo Lamp, nifeida, '
|
85 |
+
- 'search_document: Everlast 70-Pound MMA Poly Canvas Heavy Bag (Black), Everlast,
|
86 |
+
Black'
|
|
|
|
|
87 |
pipeline_tag: sentence-similarity
|
88 |
model-index:
|
89 |
+
- name: SentenceTransformer
|
90 |
results:
|
91 |
- task:
|
92 |
type: triplet
|
|
|
96 |
type: unknown
|
97 |
metrics:
|
98 |
- type: cosine_accuracy
|
99 |
+
value: 0.7297333333333333
|
100 |
name: Cosine Accuracy
|
101 |
- type: dot_accuracy
|
102 |
+
value: 0.2886
|
103 |
name: Dot Accuracy
|
104 |
- type: manhattan_accuracy
|
105 |
+
value: 0.7276666666666667
|
106 |
name: Manhattan Accuracy
|
107 |
- type: euclidean_accuracy
|
108 |
+
value: 0.7302
|
109 |
name: Euclidean Accuracy
|
110 |
- type: max_accuracy
|
111 |
+
value: 0.7302
|
112 |
name: Max Accuracy
|
113 |
- task:
|
114 |
type: semantic-similarity
|
|
|
118 |
type: unknown
|
119 |
metrics:
|
120 |
- type: pearson_cosine
|
121 |
+
value: 0.4148970258272059
|
122 |
name: Pearson Cosine
|
123 |
- type: spearman_cosine
|
124 |
+
value: 0.398627797786331
|
125 |
name: Spearman Cosine
|
126 |
- type: pearson_manhattan
|
127 |
+
value: 0.37912086775286635
|
128 |
name: Pearson Manhattan
|
129 |
- type: spearman_manhattan
|
130 |
+
value: 0.3706070496698137
|
131 |
name: Spearman Manhattan
|
132 |
- type: pearson_euclidean
|
133 |
+
value: 0.3795867481634979
|
134 |
name: Pearson Euclidean
|
135 |
- type: spearman_euclidean
|
136 |
+
value: 0.3712739507251931
|
137 |
name: Spearman Euclidean
|
138 |
- type: pearson_dot
|
139 |
+
value: 0.37803301496759
|
140 |
name: Pearson Dot
|
141 |
- type: spearman_dot
|
142 |
+
value: 0.37716678508954316
|
143 |
name: Spearman Dot
|
144 |
- type: pearson_max
|
145 |
+
value: 0.4148970258272059
|
146 |
name: Pearson Max
|
147 |
- type: spearman_max
|
148 |
+
value: 0.398627797786331
|
149 |
name: Spearman Max
|
150 |
+
- task:
|
151 |
+
type: information-retrieval
|
152 |
+
name: Information Retrieval
|
153 |
+
dataset:
|
154 |
+
name: Unknown
|
155 |
+
type: unknown
|
156 |
+
metrics:
|
157 |
+
- type: cosine_accuracy@10
|
158 |
+
value: 0.97
|
159 |
+
name: Cosine Accuracy@10
|
160 |
+
- type: cosine_precision@10
|
161 |
+
value: 0.6959
|
162 |
+
name: Cosine Precision@10
|
163 |
+
- type: cosine_recall@10
|
164 |
+
value: 0.6232158089273498
|
165 |
+
name: Cosine Recall@10
|
166 |
+
- type: cosine_ndcg@10
|
167 |
+
value: 0.8306950477475292
|
168 |
+
name: Cosine Ndcg@10
|
169 |
+
- type: cosine_mrr@10
|
170 |
+
value: 0.91050753968254
|
171 |
+
name: Cosine Mrr@10
|
172 |
+
- type: cosine_map@10
|
173 |
+
value: 0.7765982865646258
|
174 |
+
name: Cosine Map@10
|
175 |
+
- type: dot_accuracy@10
|
176 |
+
value: 0.945
|
177 |
+
name: Dot Accuracy@10
|
178 |
+
- type: dot_precision@10
|
179 |
+
value: 0.6282000000000001
|
180 |
+
name: Dot Precision@10
|
181 |
+
- type: dot_recall@10
|
182 |
+
value: 0.5617500930766831
|
183 |
+
name: Dot Recall@10
|
184 |
+
- type: dot_ndcg@10
|
185 |
+
value: 0.7560665190218724
|
186 |
+
name: Dot Ndcg@10
|
187 |
+
- type: dot_mrr@10
|
188 |
+
value: 0.8675638888888889
|
189 |
+
name: Dot Mrr@10
|
190 |
+
- type: dot_map@10
|
191 |
+
value: 0.682285591931217
|
192 |
+
name: Dot Map@10
|
193 |
---
|
194 |
|
195 |
+
# SentenceTransformer
|
196 |
|
197 |
+
This is a [sentence-transformers](https://www.SBERT.net) model trained on the triplets dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
198 |
|
199 |
## Model Details
|
200 |
|
201 |
### Model Description
|
202 |
- **Model Type:** Sentence Transformer
|
203 |
+
<!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
|
204 |
- **Maximum Sequence Length:** 8192 tokens
|
205 |
- **Output Dimensionality:** 768 tokens
|
206 |
- **Similarity Function:** Cosine Similarity
|
207 |
+
- **Training Dataset:**
|
208 |
- triplets
|
|
|
209 |
<!-- - **Language:** Unknown -->
|
210 |
<!-- - **License:** Unknown -->
|
211 |
|
|
|
242 |
model = SentenceTransformer("lv12/esci-nomic-embed-text-v1_5_4")
|
243 |
# Run inference
|
244 |
sentences = [
|
245 |
+
'search_query: punxhing bag',
|
246 |
+
'search_document: Everlast 70-Pound MMA Poly Canvas Heavy Bag (Black), Everlast, Black',
|
247 |
+
'search_document: Power Core Bag (EA), Everlast, Black/White',
|
248 |
]
|
249 |
embeddings = model.encode(sentences)
|
250 |
print(embeddings.shape)
|
|
|
290 |
|
291 |
| Metric | Value |
|
292 |
|:--------------------|:-----------|
|
293 |
+
| **cosine_accuracy** | **0.7297** |
|
294 |
+
| dot_accuracy | 0.2886 |
|
295 |
+
| manhattan_accuracy | 0.7277 |
|
296 |
+
| euclidean_accuracy | 0.7302 |
|
297 |
+
| max_accuracy | 0.7302 |
|
298 |
|
299 |
#### Semantic Similarity
|
300 |
|
|
|
302 |
|
303 |
| Metric | Value |
|
304 |
|:--------------------|:-----------|
|
305 |
+
| pearson_cosine | 0.4149 |
|
306 |
+
| **spearman_cosine** | **0.3986** |
|
307 |
+
| pearson_manhattan | 0.3791 |
|
308 |
+
| spearman_manhattan | 0.3706 |
|
309 |
+
| pearson_euclidean | 0.3796 |
|
310 |
+
| spearman_euclidean | 0.3713 |
|
311 |
+
| pearson_dot | 0.378 |
|
312 |
+
| spearman_dot | 0.3772 |
|
313 |
+
| pearson_max | 0.4149 |
|
314 |
+
| spearman_max | 0.3986 |
|
315 |
+
|
316 |
+
#### Information Retrieval
|
317 |
+
|
318 |
+
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
319 |
+
|
320 |
+
| Metric | Value |
|
321 |
+
|:--------------------|:-----------|
|
322 |
+
| cosine_accuracy@10 | 0.97 |
|
323 |
+
| cosine_precision@10 | 0.6959 |
|
324 |
+
| cosine_recall@10 | 0.6232 |
|
325 |
+
| cosine_ndcg@10 | 0.8307 |
|
326 |
+
| cosine_mrr@10 | 0.9105 |
|
327 |
+
| **cosine_map@10** | **0.7766** |
|
328 |
+
| dot_accuracy@10 | 0.945 |
|
329 |
+
| dot_precision@10 | 0.6282 |
|
330 |
+
| dot_recall@10 | 0.5618 |
|
331 |
+
| dot_ndcg@10 | 0.7561 |
|
332 |
+
| dot_mrr@10 | 0.8676 |
|
333 |
+
| dot_map@10 | 0.6823 |
|
334 |
|
335 |
<!--
|
336 |
## Bias, Risks and Limitations
|
|
|
346 |
|
347 |
## Training Details
|
348 |
|
349 |
+
### Training Dataset
|
350 |
|
351 |
#### triplets
|
352 |
|
353 |
* Dataset: triplets
|
354 |
+
* Size: 1,500,000 training samples
|
355 |
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
356 |
* Approximate statistics based on the first 1000 samples:
|
357 |
+
| | anchor | positive | negative |
|
358 |
+
|:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
359 |
+
| type | string | string | string |
|
360 |
+
| details | <ul><li>min: 7 tokens</li><li>mean: 11.0 tokens</li><li>max: 27 tokens</li></ul> | <ul><li>min: 9 tokens</li><li>mean: 38.87 tokens</li><li>max: 91 tokens</li></ul> | <ul><li>min: 9 tokens</li><li>mean: 39.17 tokens</li><li>max: 96 tokens</li></ul> |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
361 |
* Samples:
|
362 |
+
| anchor | positive | negative |
|
363 |
+
|:--------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
364 |
+
| <code>search_query: chiffon ivory fabric</code> | <code>search_document: DJBM 59’’ Solid Color Sheer Chiffon Fabric Yards Continuous All Colors for DIY Decoration Valance Ivory/1 Yard, DJBM, Ivory</code> | <code>search_document: 58" Blush Solid Color Sheer Chiffon Fabric by The Bolt - 25 Yards, Stylish FABRIC, </code> |
|
365 |
+
| <code>search_query: tummy control lingerie for women for sex</code> | <code>search_document: Women's Plus Size Chemise Floral Lace Lingerie Sexy Bodysuit Mesh Babydoll Sleepwear (R007,White, XXXX-Large), Chic Lover, White</code> | <code>search_document: Avidlove Lingerie for Women Teddy One Piece Lace Babydoll Bodysuit Black Medium, Avidlove, Black</code> |
|
366 |
+
| <code>search_query: photo metal prints</code> | <code>search_document: Smile Art Design Personalized Photo Print Desktop Photo Panel Print with Kickstand Upload Your Tabletop Berlin Shape Frame Custom Photo Print Personalized Gift for Man Woman 5" x7", Smile Art Design, Berlin 5" X 7"</code> | <code>search_document: They Whispered to Her You Cannot Withstand The Storm - Positive Motivational Uplifting Encouragement Gifts for Women Teens - Inspirational Quote Wall Art - Boho Decoration Print - Dragonfly Wall Decor, Yellowbird Art & Design, </code> |
|
367 |
+
* Loss: [<code>TripletLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#tripletloss) with these parameters:
|
368 |
```json
|
369 |
{
|
370 |
+
"distance_metric": "TripletDistanceMetric.COSINE",
|
371 |
+
"triplet_margin": 0.7
|
372 |
}
|
373 |
```
|
374 |
|
375 |
+
### Evaluation Dataset
|
376 |
|
377 |
#### triplets
|
378 |
|
379 |
* Dataset: triplets
|
380 |
+
* Size: 15,000 evaluation samples
|
381 |
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
382 |
* Approximate statistics based on the first 1000 samples:
|
383 |
+
| | anchor | positive | negative |
|
384 |
+
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
|
385 |
+
| type | string | string | string |
|
386 |
+
| details | <ul><li>min: 7 tokens</li><li>mean: 11.14 tokens</li><li>max: 31 tokens</li></ul> | <ul><li>min: 9 tokens</li><li>mean: 38.83 tokens</li><li>max: 105 tokens</li></ul> | <ul><li>min: 9 tokens</li><li>mean: 39.12 tokens</li><li>max: 113 tokens</li></ul> |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
387 |
* Samples:
|
388 |
+
| anchor | positive | negative |
|
389 |
+
|:------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
390 |
+
| <code>search_query: a cold dark place</code> | <code>search_document: A Cold Dark Place (An Emily Kenyon Thriller Book 1), , </code> | <code>search_document: The Great Alone, , </code> |
|
391 |
+
| <code>search_query: quay sunglasses</code> | <code>search_document: Quay Women's Hindsight Sunglasses, Matte Black/Rainbow Mirror, Quay, Multicolor</code> | <code>search_document: SORVINO Aviator Sunglasses for Women Classic Oversized Sun Glasses UV400 Protection (A-Black Frame/Faded Lens, 60), SORVINO, Black Frame/Faded Lens</code> |
|
392 |
+
| <code>search_query: baby girl jean fress</code> | <code>search_document: Toddler Baby Girl Play Wear Floral Princess Summer Dresses with Bow Ruffle Denim Skirt Set for Summer 9-12months Light Blue, NZRVAWS, </code> | <code>search_document: Slowera Baby Toddler Girls Denim Ruffled Bodysuit Blue Soft One-Piece Romper (Blue, 6-12 Months), Slowera, Blue</code> |
|
393 |
+
* Loss: [<code>TripletLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#tripletloss) with these parameters:
|
394 |
```json
|
395 |
{
|
396 |
+
"distance_metric": "TripletDistanceMetric.COSINE",
|
397 |
+
"triplet_margin": 0.7
|
398 |
}
|
399 |
```
|
400 |
|
401 |
### Training Hyperparameters
|
402 |
#### Non-Default Hyperparameters
|
403 |
|
404 |
+
- `per_device_train_batch_size`: 32
|
405 |
+
- `per_device_eval_batch_size`: 16
|
406 |
- `gradient_accumulation_steps`: 2
|
407 |
+
- `learning_rate`: 1e-07
|
408 |
+
- `lr_scheduler_type`: polynomial
|
409 |
+
- `lr_scheduler_kwargs`: {'lr_end': 1e-08, 'power': 2.0}
|
410 |
+
- `warmup_ratio`: 0.05
|
|
|
411 |
- `dataloader_num_workers`: 4
|
412 |
- `dataloader_prefetch_factor`: 4
|
413 |
- `load_best_model_at_end`: True
|
414 |
- `gradient_checkpointing`: True
|
415 |
+
- `auto_find_batch_size`: True
|
416 |
- `batch_sampler`: no_duplicates
|
417 |
|
418 |
#### All Hyperparameters
|
|
|
421 |
- `overwrite_output_dir`: False
|
422 |
- `do_predict`: False
|
423 |
- `prediction_loss_only`: True
|
424 |
+
- `per_device_train_batch_size`: 32
|
425 |
+
- `per_device_eval_batch_size`: 16
|
426 |
- `per_gpu_train_batch_size`: None
|
427 |
- `per_gpu_eval_batch_size`: None
|
428 |
- `gradient_accumulation_steps`: 2
|
429 |
- `eval_accumulation_steps`: None
|
430 |
+
- `learning_rate`: 1e-07
|
431 |
- `weight_decay`: 0.0
|
432 |
- `adam_beta1`: 0.9
|
433 |
- `adam_beta2`: 0.999
|
|
|
435 |
- `max_grad_norm`: 1.0
|
436 |
- `num_train_epochs`: 3
|
437 |
- `max_steps`: -1
|
438 |
+
- `lr_scheduler_type`: polynomial
|
439 |
+
- `lr_scheduler_kwargs`: {'lr_end': 1e-08, 'power': 2.0}
|
440 |
+
- `warmup_ratio`: 0.05
|
441 |
- `warmup_steps`: 0
|
442 |
- `log_level`: passive
|
443 |
- `log_level_replica`: warning
|
|
|
506 |
- `push_to_hub_model_id`: None
|
507 |
- `push_to_hub_organization`: None
|
508 |
- `mp_parameters`:
|
509 |
+
- `auto_find_batch_size`: True
|
510 |
- `full_determinism`: False
|
511 |
- `torchdynamo`: None
|
512 |
- `ray_scope`: last
|
|
|
527 |
### Training Logs
|
528 |
<details><summary>Click to expand</summary>
|
529 |
|
530 |
+
| Epoch | Step | Training Loss | triplets loss | cosine_accuracy | cosine_map@10 | spearman_cosine |
|
531 |
+
|:------:|:----:|:-------------:|:-------------:|:---------------:|:-------------:|:---------------:|
|
532 |
+
| 0.0004 | 10 | 0.6484 | - | - | - | - |
|
533 |
+
| 0.0009 | 20 | 0.6538 | - | - | - | - |
|
534 |
+
| 0.0013 | 30 | 0.651 | - | - | - | - |
|
535 |
+
| 0.0017 | 40 | 0.6491 | - | - | - | - |
|
536 |
+
| 0.0021 | 50 | 0.6433 | - | - | - | - |
|
537 |
+
| 0.0026 | 60 | 0.6545 | - | - | - | - |
|
538 |
+
| 0.0030 | 70 | 0.6459 | - | - | - | - |
|
539 |
+
| 0.0034 | 80 | 0.6517 | - | - | - | - |
|
540 |
+
| 0.0038 | 90 | 0.6505 | - | - | - | - |
|
541 |
+
| 0.0043 | 100 | 0.6477 | 0.6560 | 0.7255 | 0.7684 | 0.3931 |
|
542 |
+
| 0.0047 | 110 | 0.6493 | - | - | - | - |
|
543 |
+
| 0.0051 | 120 | 0.6514 | - | - | - | - |
|
544 |
+
| 0.0055 | 130 | 0.6482 | - | - | - | - |
|
545 |
+
| 0.0060 | 140 | 0.6566 | - | - | - | - |
|
546 |
+
| 0.0064 | 150 | 0.6476 | - | - | - | - |
|
547 |
+
| 0.0068 | 160 | 0.6523 | - | - | - | - |
|
548 |
+
| 0.0073 | 170 | 0.6503 | - | - | - | - |
|
549 |
+
| 0.0077 | 180 | 0.6568 | - | - | - | - |
|
550 |
+
| 0.0081 | 190 | 0.6496 | - | - | - | - |
|
551 |
+
| 0.0085 | 200 | 0.6512 | 0.6558 | 0.725 | 0.7687 | 0.3932 |
|
552 |
+
| 0.0090 | 210 | 0.642 | - | - | - | - |
|
553 |
+
| 0.0094 | 220 | 0.644 | - | - | - | - |
|
554 |
+
| 0.0098 | 230 | 0.6512 | - | - | - | - |
|
555 |
+
| 0.0102 | 240 | 0.6512 | - | - | - | - |
|
556 |
+
| 0.0107 | 250 | 0.6477 | - | - | - | - |
|
557 |
+
| 0.0111 | 260 | 0.6521 | - | - | - | - |
|
558 |
+
| 0.0115 | 270 | 0.6487 | - | - | - | - |
|
559 |
+
| 0.0119 | 280 | 0.6541 | - | - | - | - |
|
560 |
+
| 0.0124 | 290 | 0.6481 | - | - | - | - |
|
561 |
+
| 0.0128 | 300 | 0.6495 | 0.6555 | 0.7247 | 0.7690 | 0.3933 |
|
562 |
+
| 0.0132 | 310 | 0.6536 | - | - | - | - |
|
563 |
+
| 0.0137 | 320 | 0.6515 | - | - | - | - |
|
564 |
+
| 0.0141 | 330 | 0.6526 | - | - | - | - |
|
565 |
+
| 0.0145 | 340 | 0.6509 | - | - | - | - |
|
566 |
+
| 0.0149 | 350 | 0.6506 | - | - | - | - |
|
567 |
+
| 0.0154 | 360 | 0.6451 | - | - | - | - |
|
568 |
+
| 0.0158 | 370 | 0.6505 | - | - | - | - |
|
569 |
+
| 0.0162 | 380 | 0.6529 | - | - | - | - |
|
570 |
+
| 0.0166 | 390 | 0.6553 | - | - | - | - |
|
571 |
+
| 0.0171 | 400 | 0.6512 | 0.6550 | 0.7245 | 0.7693 | 0.3934 |
|
572 |
+
| 0.0175 | 410 | 0.6478 | - | - | - | - |
|
573 |
+
| 0.0179 | 420 | 0.6497 | - | - | - | - |
|
574 |
+
| 0.0183 | 430 | 0.6498 | - | - | - | - |
|
575 |
+
| 0.0188 | 440 | 0.6499 | - | - | - | - |
|
576 |
+
| 0.0192 | 450 | 0.6462 | - | - | - | - |
|
577 |
+
| 0.0196 | 460 | 0.6497 | - | - | - | - |
|
578 |
+
| 0.0201 | 470 | 0.6408 | - | - | - | - |
|
579 |
+
| 0.0205 | 480 | 0.6507 | - | - | - | - |
|
580 |
+
| 0.0209 | 490 | 0.6505 | - | - | - | - |
|
581 |
+
| 0.0213 | 500 | 0.6481 | 0.6545 | 0.7249 | 0.7694 | 0.3936 |
|
582 |
+
| 0.0218 | 510 | 0.6436 | - | - | - | - |
|
583 |
+
| 0.0222 | 520 | 0.653 | - | - | - | - |
|
584 |
+
| 0.0226 | 530 | 0.652 | - | - | - | - |
|
585 |
+
| 0.0230 | 540 | 0.6509 | - | - | - | - |
|
586 |
+
| 0.0235 | 550 | 0.6469 | - | - | - | - |
|
587 |
+
| 0.0239 | 560 | 0.6515 | - | - | - | - |
|
588 |
+
| 0.0243 | 570 | 0.6485 | - | - | - | - |
|
589 |
+
| 0.0247 | 580 | 0.6499 | - | - | - | - |
|
590 |
+
| 0.0252 | 590 | 0.6472 | - | - | - | - |
|
591 |
+
| 0.0256 | 600 | 0.6569 | 0.6538 | 0.7255 | 0.7702 | 0.3939 |
|
592 |
+
| 0.0260 | 610 | 0.6457 | - | - | - | - |
|
593 |
+
| 0.0265 | 620 | 0.6485 | - | - | - | - |
|
594 |
+
| 0.0269 | 630 | 0.6467 | - | - | - | - |
|
595 |
+
| 0.0273 | 640 | 0.6494 | - | - | - | - |
|
596 |
+
| 0.0277 | 650 | 0.6492 | - | - | - | - |
|
597 |
+
| 0.0282 | 660 | 0.6439 | - | - | - | - |
|
598 |
+
| 0.0286 | 670 | 0.6468 | - | - | - | - |
|
599 |
+
| 0.0290 | 680 | 0.6493 | - | - | - | - |
|
600 |
+
| 0.0294 | 690 | 0.6483 | - | - | - | - |
|
601 |
+
| 0.0299 | 700 | 0.6474 | 0.6530 | 0.7259 | 0.7704 | 0.3942 |
|
602 |
+
| 0.0303 | 710 | 0.6512 | - | - | - | - |
|
603 |
+
| 0.0307 | 720 | 0.6431 | - | - | - | - |
|
604 |
+
| 0.0311 | 730 | 0.6484 | - | - | - | - |
|
605 |
+
| 0.0316 | 740 | 0.6488 | - | - | - | - |
|
606 |
+
| 0.032 | 750 | 0.6525 | - | - | - | - |
|
607 |
+
| 0.0324 | 760 | 0.6464 | - | - | - | - |
|
608 |
+
| 0.0329 | 770 | 0.6453 | - | - | - | - |
|
609 |
+
| 0.0333 | 780 | 0.6471 | - | - | - | - |
|
610 |
+
| 0.0337 | 790 | 0.6478 | - | - | - | - |
|
611 |
+
| 0.0341 | 800 | 0.6456 | 0.6520 | 0.7262 | 0.7704 | 0.3945 |
|
612 |
+
| 0.0346 | 810 | 0.6494 | - | - | - | - |
|
613 |
+
| 0.0350 | 820 | 0.6374 | - | - | - | - |
|
614 |
+
| 0.0354 | 830 | 0.6441 | - | - | - | - |
|
615 |
+
| 0.0358 | 840 | 0.643 | - | - | - | - |
|
616 |
+
| 0.0363 | 850 | 0.6471 | - | - | - | - |
|
617 |
+
| 0.0367 | 860 | 0.6401 | - | - | - | - |
|
618 |
+
| 0.0371 | 870 | 0.6381 | - | - | - | - |
|
619 |
+
| 0.0375 | 880 | 0.6491 | - | - | - | - |
|
620 |
+
| 0.0380 | 890 | 0.6428 | - | - | - | - |
|
621 |
+
| 0.0384 | 900 | 0.6393 | 0.6509 | 0.7262 | 0.7712 | 0.3948 |
|
622 |
+
| 0.0388 | 910 | 0.6453 | - | - | - | - |
|
623 |
+
| 0.0393 | 920 | 0.6426 | - | - | - | - |
|
624 |
+
| 0.0397 | 930 | 0.644 | - | - | - | - |
|
625 |
+
| 0.0401 | 940 | 0.6469 | - | - | - | - |
|
626 |
+
| 0.0405 | 950 | 0.6372 | - | - | - | - |
|
627 |
+
| 0.0410 | 960 | 0.6491 | - | - | - | - |
|
628 |
+
| 0.0414 | 970 | 0.6441 | - | - | - | - |
|
629 |
+
| 0.0418 | 980 | 0.6394 | - | - | - | - |
|
630 |
+
| 0.0422 | 990 | 0.6416 | - | - | - | - |
|
631 |
+
| 0.0427 | 1000 | 0.6465 | 0.6497 | 0.7263 | 0.7719 | 0.3951 |
|
632 |
+
| 0.0431 | 1010 | 0.6485 | - | - | - | - |
|
633 |
+
| 0.0435 | 1020 | 0.6391 | - | - | - | - |
|
634 |
+
| 0.0439 | 1030 | 0.6489 | - | - | - | - |
|
635 |
+
| 0.0444 | 1040 | 0.6417 | - | - | - | - |
|
636 |
+
| 0.0448 | 1050 | 0.6355 | - | - | - | - |
|
637 |
+
| 0.0452 | 1060 | 0.6445 | - | - | - | - |
|
638 |
+
| 0.0457 | 1070 | 0.6427 | - | - | - | - |
|
639 |
+
| 0.0461 | 1080 | 0.6443 | - | - | - | - |
|
640 |
+
| 0.0465 | 1090 | 0.6422 | - | - | - | - |
|
641 |
+
| 0.0469 | 1100 | 0.6406 | 0.6483 | 0.7265 | 0.7721 | 0.3955 |
|
642 |
+
| 0.0474 | 1110 | 0.6415 | - | - | - | - |
|
643 |
+
| 0.0478 | 1120 | 0.6389 | - | - | - | - |
|
644 |
+
| 0.0482 | 1130 | 0.6437 | - | - | - | - |
|
645 |
+
| 0.0486 | 1140 | 0.6412 | - | - | - | - |
|
646 |
+
| 0.0491 | 1150 | 0.6457 | - | - | - | - |
|
647 |
+
| 0.0495 | 1160 | 0.6364 | - | - | - | - |
|
648 |
+
| 0.0499 | 1170 | 0.6389 | - | - | - | - |
|
649 |
+
| 0.0503 | 1180 | 0.6394 | - | - | - | - |
|
650 |
+
| 0.0508 | 1190 | 0.6465 | - | - | - | - |
|
651 |
+
| 0.0512 | 1200 | 0.6453 | 0.6468 | 0.7269 | 0.7734 | 0.3959 |
|
652 |
+
| 0.0516 | 1210 | 0.6465 | - | - | - | - |
|
653 |
+
| 0.0521 | 1220 | 0.6389 | - | - | - | - |
|
654 |
+
| 0.0525 | 1230 | 0.6448 | - | - | - | - |
|
655 |
+
| 0.0529 | 1240 | 0.6325 | - | - | - | - |
|
656 |
+
| 0.0533 | 1250 | 0.6347 | - | - | - | - |
|
657 |
+
| 0.0538 | 1260 | 0.6363 | - | - | - | - |
|
658 |
+
| 0.0542 | 1270 | 0.6387 | - | - | - | - |
|
659 |
+
| 0.0546 | 1280 | 0.641 | - | - | - | - |
|
660 |
+
| 0.0550 | 1290 | 0.6381 | - | - | - | - |
|
661 |
+
| 0.0555 | 1300 | 0.6412 | 0.6452 | 0.7273 | 0.7747 | 0.3962 |
|
662 |
+
| 0.0559 | 1310 | 0.6384 | - | - | - | - |
|
663 |
+
| 0.0563 | 1320 | 0.6391 | - | - | - | - |
|
664 |
+
| 0.0567 | 1330 | 0.6327 | - | - | - | - |
|
665 |
+
| 0.0572 | 1340 | 0.6388 | - | - | - | - |
|
666 |
+
| 0.0576 | 1350 | 0.6328 | - | - | - | - |
|
667 |
+
| 0.0580 | 1360 | 0.6327 | - | - | - | - |
|
668 |
+
| 0.0585 | 1370 | 0.6333 | - | - | - | - |
|
669 |
+
| 0.0589 | 1380 | 0.6388 | - | - | - | - |
|
670 |
+
| 0.0593 | 1390 | 0.6373 | - | - | - | - |
|
671 |
+
| 0.0597 | 1400 | 0.636 | 0.6434 | 0.7278 | 0.7749 | 0.3966 |
|
672 |
+
| 0.0602 | 1410 | 0.6357 | - | - | - | - |
|
673 |
+
| 0.0606 | 1420 | 0.6394 | - | - | - | - |
|
674 |
+
| 0.0610 | 1430 | 0.6339 | - | - | - | - |
|
675 |
+
| 0.0614 | 1440 | 0.6385 | - | - | - | - |
|
676 |
+
| 0.0619 | 1450 | 0.6288 | - | - | - | - |
|
677 |
+
| 0.0623 | 1460 | 0.6393 | - | - | - | - |
|
678 |
+
| 0.0627 | 1470 | 0.638 | - | - | - | - |
|
679 |
+
| 0.0631 | 1480 | 0.6353 | - | - | - | - |
|
680 |
+
| 0.0636 | 1490 | 0.6335 | - | - | - | - |
|
681 |
+
| 0.064 | 1500 | 0.6329 | 0.6414 | 0.7281 | 0.7749 | 0.3970 |
|
682 |
+
| 0.0644 | 1510 | 0.6363 | - | - | - | - |
|
683 |
+
| 0.0649 | 1520 | 0.6311 | - | - | - | - |
|
684 |
+
| 0.0653 | 1530 | 0.6367 | - | - | - | - |
|
685 |
+
| 0.0657 | 1540 | 0.636 | - | - | - | - |
|
686 |
+
| 0.0661 | 1550 | 0.6351 | - | - | - | - |
|
687 |
+
| 0.0666 | 1560 | 0.637 | - | - | - | - |
|
688 |
+
| 0.0670 | 1570 | 0.6352 | - | - | - | - |
|
689 |
+
| 0.0674 | 1580 | 0.6285 | - | - | - | - |
|
690 |
+
| 0.0678 | 1590 | 0.6311 | - | - | - | - |
|
691 |
+
| 0.0683 | 1600 | 0.6347 | 0.6393 | 0.7283 | 0.7752 | 0.3974 |
|
692 |
+
| 0.0687 | 1610 | 0.6393 | - | - | - | - |
|
693 |
+
| 0.0691 | 1620 | 0.6368 | - | - | - | - |
|
694 |
+
| 0.0695 | 1630 | 0.6354 | - | - | - | - |
|
695 |
+
| 0.0700 | 1640 | 0.6283 | - | - | - | - |
|
696 |
+
| 0.0704 | 1650 | 0.6289 | - | - | - | - |
|
697 |
+
| 0.0708 | 1660 | 0.6291 | - | - | - | - |
|
698 |
+
| 0.0713 | 1670 | 0.6274 | - | - | - | - |
|
699 |
+
| 0.0717 | 1680 | 0.6217 | - | - | - | - |
|
700 |
+
| 0.0721 | 1690 | 0.6281 | - | - | - | - |
|
701 |
+
| 0.0725 | 1700 | 0.6341 | 0.6371 | 0.7289 | 0.7757 | 0.3978 |
|
702 |
+
| 0.0730 | 1710 | 0.6354 | - | - | - | - |
|
703 |
+
| 0.0734 | 1720 | 0.6265 | - | - | - | - |
|
704 |
+
| 0.0738 | 1730 | 0.6214 | - | - | - | - |
|
705 |
+
| 0.0742 | 1740 | 0.6301 | - | - | - | - |
|
706 |
+
| 0.0747 | 1750 | 0.621 | - | - | - | - |
|
707 |
+
| 0.0751 | 1760 | 0.6259 | - | - | - | - |
|
708 |
+
| 0.0755 | 1770 | 0.6261 | - | - | - | - |
|
709 |
+
| 0.0759 | 1780 | 0.6273 | - | - | - | - |
|
710 |
+
| 0.0764 | 1790 | 0.6311 | - | - | - | - |
|
711 |
+
| 0.0768 | 1800 | 0.6232 | 0.6347 | 0.7292 | 0.7755 | 0.3981 |
|
712 |
+
| 0.0772 | 1810 | 0.6293 | - | - | - | - |
|
713 |
+
| 0.0777 | 1820 | 0.617 | - | - | - | - |
|
714 |
+
| 0.0781 | 1830 | 0.6303 | - | - | - | - |
|
715 |
+
| 0.0785 | 1840 | 0.6225 | - | - | - | - |
|
716 |
+
| 0.0789 | 1850 | 0.6313 | - | - | - | - |
|
717 |
+
| 0.0794 | 1860 | 0.6229 | - | - | - | - |
|
718 |
+
| 0.0798 | 1870 | 0.6236 | - | - | - | - |
|
719 |
+
| 0.0802 | 1880 | 0.6265 | - | - | - | - |
|
720 |
+
| 0.0806 | 1890 | 0.6179 | - | - | - | - |
|
721 |
+
| 0.0811 | 1900 | 0.6277 | 0.6321 | 0.7292 | 0.7760 | 0.3984 |
|
722 |
+
| 0.0815 | 1910 | 0.6266 | - | - | - | - |
|
723 |
+
| 0.0819 | 1920 | 0.6209 | - | - | - | - |
|
724 |
+
| 0.0823 | 1930 | 0.6258 | - | - | - | - |
|
725 |
+
| 0.0828 | 1940 | 0.6143 | - | - | - | - |
|
726 |
+
| 0.0832 | 1950 | 0.6176 | - | - | - | - |
|
727 |
+
| 0.0836 | 1960 | 0.628 | - | - | - | - |
|
728 |
+
| 0.0841 | 1970 | 0.6147 | - | - | - | - |
|
729 |
+
| 0.0845 | 1980 | 0.6175 | - | - | - | - |
|
730 |
+
| 0.0849 | 1990 | 0.6135 | - | - | - | - |
|
731 |
+
| 0.0853 | 2000 | 0.6214 | 0.6293 | 0.7297 | 0.7766 | 0.3986 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
732 |
|
733 |
</details>
|
734 |
|
|
|
758 |
}
|
759 |
```
|
760 |
|
761 |
+
#### TripletLoss
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
762 |
```bibtex
|
763 |
+
@misc{hermans2017defense,
|
764 |
+
title={In Defense of the Triplet Loss for Person Re-Identification},
|
765 |
+
author={Alexander Hermans and Lucas Beyer and Bastian Leibe},
|
766 |
+
year={2017},
|
767 |
+
eprint={1703.07737},
|
768 |
archivePrefix={arXiv},
|
769 |
+
primaryClass={cs.CV}
|
770 |
}
|
771 |
```
|
772 |
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "models/nomic-embed-text-esci/checkpoint-
|
3 |
"activation_function": "swiglu",
|
4 |
"architectures": [
|
5 |
"NomicBertModel"
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "models/nomic-embed-text-train-esci/checkpoint-2000",
|
3 |
"activation_function": "swiglu",
|
4 |
"architectures": [
|
5 |
"NomicBertModel"
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 546938168
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3cf0ece0c783f88dbd3b658097e010b54fb1c12a8c464d21cac81fd4625675b8
|
3 |
size 546938168
|