Update zero-shot table to match results reported in paper.
Browse files
README.md
CHANGED
@@ -121,7 +121,7 @@ We also use a weight decay of 0.2 and a batch size of 32K.
|
|
121 |
### Testing Data
|
122 |
|
123 |
We tested BioCLIP on the following collection of 10 biologically-relevant tasks.
|
124 |
-
- [Meta-Album](https://paperswithcode.com/dataset/meta-album): Specifically, we used the Plankton, Insects, Insects 2, PlantNet, Fungi, PlantVillage, Medicinal Leaf, and PlantDoc datasets from Set-0 through Set-2 (Set-3 was still not released as of our publication/evaluation (Nov. 2023).
|
125 |
- [Birds 525](https://www.kaggle.com/datasets/gpiosenka/100-bird-species): We evaluated on the 2,625 test images provided with the dataset.
|
126 |
- [Rare Species](https://huggingface.co/datasets/imageomics/rare-species): A new dataset we curated for the purpose of testing this model and to contribute to the ML for Conservation community. It consists of 400 species labeled Near Threatened through Extinct in the Wild by the [IUCN Red List](https://www.iucnredlist.org/), with 30 images per species. For more information, see our dataset, [Rare Species](https://huggingface.co/datasets/imageomics/rare-species).
|
127 |
|
@@ -170,8 +170,8 @@ Please see [our paper](https://doi.org/10.48550/arXiv.2311.18803) for few-shot r
|
|
170 |
<td>5.4</td>
|
171 |
<td>15.9</td>
|
172 |
<td>26.1</td>
|
173 |
-
<td>
|
174 |
-
<td>21.
|
175 |
</tr>
|
176 |
<tr>
|
177 |
<td>OpenCLIP</td>
|
@@ -184,36 +184,36 @@ Please see [our paper](https://doi.org/10.48550/arXiv.2311.18803) for few-shot r
|
|
184 |
<td>8.0</td>
|
185 |
<td>12.4</td>
|
186 |
<td>25.8</td>
|
187 |
-
<td>
|
188 |
-
<td>20.
|
189 |
</tr>
|
190 |
<tr>
|
191 |
<td>BioCLIP</td>
|
192 |
-
<td><b>
|
193 |
-
<td><b>
|
194 |
-
<td><b>
|
195 |
-
<td><b>
|
196 |
-
<td><b>91.
|
197 |
-
<td
|
198 |
-
<td><b>24.
|
199 |
-
<td><b>
|
200 |
-
<td><b>
|
201 |
-
<td><b>
|
202 |
-
<td><b>
|
203 |
</tr>
|
204 |
<tr>
|
205 |
<td>iNat21 Only</td>
|
206 |
-
<td>
|
207 |
-
<td>2.
|
208 |
-
<td>
|
209 |
-
<td>
|
210 |
-
<td>
|
211 |
-
<td>
|
212 |
-
<td>
|
213 |
-
<td>
|
214 |
-
<td>
|
215 |
-
<td>
|
216 |
-
<td>
|
217 |
</tr>
|
218 |
</tbody>
|
219 |
</table>
|
@@ -221,7 +221,7 @@ Please see [our paper](https://doi.org/10.48550/arXiv.2311.18803) for few-shot r
|
|
221 |
|
222 |
### Summary
|
223 |
|
224 |
-
BioCLIP outperforms general-domain baselines by
|
225 |
|
226 |
### Model Examination
|
227 |
|
|
|
121 |
### Testing Data
|
122 |
|
123 |
We tested BioCLIP on the following collection of 10 biologically-relevant tasks.
|
124 |
+
- [Meta-Album](https://paperswithcode.com/dataset/meta-album): Specifically, we used the Plankton, Insects, Insects 2, PlantNet, Fungi, PlantVillage, Medicinal Leaf, and PlantDoc datasets from Set-0 through Set-2 (Set-3 was still not released as of our publication/evaluation (Nov. 2023)).
|
125 |
- [Birds 525](https://www.kaggle.com/datasets/gpiosenka/100-bird-species): We evaluated on the 2,625 test images provided with the dataset.
|
126 |
- [Rare Species](https://huggingface.co/datasets/imageomics/rare-species): A new dataset we curated for the purpose of testing this model and to contribute to the ML for Conservation community. It consists of 400 species labeled Near Threatened through Extinct in the Wild by the [IUCN Red List](https://www.iucnredlist.org/), with 30 images per species. For more information, see our dataset, [Rare Species](https://huggingface.co/datasets/imageomics/rare-species).
|
127 |
|
|
|
170 |
<td>5.4</td>
|
171 |
<td>15.9</td>
|
172 |
<td>26.1</td>
|
173 |
+
<td>31.8</td>
|
174 |
+
<td>21.9</td>
|
175 |
</tr>
|
176 |
<tr>
|
177 |
<td>OpenCLIP</td>
|
|
|
184 |
<td>8.0</td>
|
185 |
<td>12.4</td>
|
186 |
<td>25.8</td>
|
187 |
+
<td>29.8</td>
|
188 |
+
<td>20.4</td>
|
189 |
</tr>
|
190 |
<tr>
|
191 |
<td>BioCLIP</td>
|
192 |
+
<td><b>72.1</b></td>
|
193 |
+
<td><b>6.1</b></td>
|
194 |
+
<td><b>34.8</b></td>
|
195 |
+
<td><b>20.4</b></td>
|
196 |
+
<td><b>91.4</b></td>
|
197 |
+
<td>40.7</td>
|
198 |
+
<td><b>24.4</b></td>
|
199 |
+
<td><b>38.6</b></td>
|
200 |
+
<td><b>28.4</b></td>
|
201 |
+
<td><b>38.0</b></td>
|
202 |
+
<td><b>39.4</b></td>
|
203 |
</tr>
|
204 |
<tr>
|
205 |
<td>iNat21 Only</td>
|
206 |
+
<td>56.1</td>
|
207 |
+
<td>2.6</td>
|
208 |
+
<td>30.7</td>
|
209 |
+
<td>11.5</td>
|
210 |
+
<td>88.2</td>
|
211 |
+
<td><b>43.0</b></td>
|
212 |
+
<td>18.4</td>
|
213 |
+
<td>25.6</td>
|
214 |
+
<td>20.5</td>
|
215 |
+
<td>21.3</td>
|
216 |
+
<td>31.7</td>
|
217 |
</tr>
|
218 |
</tbody>
|
219 |
</table>
|
|
|
221 |
|
222 |
### Summary
|
223 |
|
224 |
+
BioCLIP outperforms general-domain baselines by 17% on average for zero-shot.
|
225 |
|
226 |
### Model Examination
|
227 |
|