Update README.md
Browse files
README.md
CHANGED
@@ -9,86 +9,92 @@ These models are:
|
|
9 |
- 100x faster than using LLMs for similar tasks
|
10 |
- Easy to fine tune
|
11 |
|
|
|
|
|
12 |
### To run the models:
|
13 |
Use https://github.com/nlmatics/nlm-model-service
|
|
|
14 |
### To acccess the models
|
15 |
Use https://github.com/nlmatics/nlm-utils
|
16 |
|
|
|
|
|
|
|
17 |
## List of Models
|
18 |
|
19 |
Click on each model to see details:
|
20 |
|
21 |
### roberta.large.boolq
|
22 |
|
23 |
-
Location
|
24 |
|
25 |
Trained with MNLI + Boolq
|
26 |
|
27 |
-
Trained by
|
28 |
|
29 |
-
Application
|
30 |
|
31 |
-
Training Process
|
32 |
|
33 |
### roberta.large.qa
|
34 |
See folder: [roberta.large.qa](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.qa)
|
35 |
|
36 |
Trained with SQuAD 2.0 + Custom Dataset preferring shorter spans better suited for data extraction
|
37 |
|
38 |
-
Trained by
|
39 |
|
40 |
-
Application
|
41 |
|
42 |
-
Training Process
|
43 |
|
44 |
### roberta.large.stsb
|
45 |
-
See folder
|
46 |
|
47 |
Trained with STSB dataset
|
48 |
|
49 |
-
Trained by
|
50 |
|
51 |
-
Application
|
52 |
|
53 |
-
Training Process
|
54 |
|
55 |
### roberta.large.phraseqa
|
56 |
-
See folder
|
57 |
|
58 |
Trained with Roberta 2.0 with the question words removed from the question
|
59 |
|
60 |
-
Trained By
|
61 |
|
62 |
-
Application
|
63 |
|
64 |
-
Training Process
|
65 |
|
66 |
### roberta.large.qasrl
|
67 |
|
68 |
-
See folder
|
69 |
|
70 |
Trained with QASRL dataset
|
71 |
|
72 |
-
Application
|
73 |
|
74 |
-
Trained By
|
75 |
|
76 |
### roberta.large.qatype.lower.RothWithQ
|
77 |
|
78 |
-
See folder
|
79 |
|
80 |
Trained with the Roth Question Type dataset.
|
81 |
|
82 |
-
Application
|
83 |
|
84 |
-
Trained By
|
85 |
|
86 |
### roberta.large.io_qa
|
87 |
|
88 |
See folder: [roberta.large.io_qa](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.io_qa)
|
89 |
Trained with SQuAD 2.0 dataset
|
90 |
|
91 |
-
Trained By
|
92 |
|
93 |
-
Training Process
|
94 |
|
|
|
9 |
- 100x faster than using LLMs for similar tasks
|
10 |
- Easy to fine tune
|
11 |
|
12 |
+
All the models below were trained at Nlmatics Corp. from 2019-2023 with base model from: https://github.com/facebookresearch/fairseq/blob/main/examples/roberta/README.md
|
13 |
+
|
14 |
### To run the models:
|
15 |
Use https://github.com/nlmatics/nlm-model-service
|
16 |
+
|
17 |
### To acccess the models
|
18 |
Use https://github.com/nlmatics/nlm-utils
|
19 |
|
20 |
+
### To train the models
|
21 |
+
TBD
|
22 |
+
|
23 |
## List of Models
|
24 |
|
25 |
Click on each model to see details:
|
26 |
|
27 |
### roberta.large.boolq
|
28 |
|
29 |
+
*Location:* [roberta.large.boolq](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.boolq)
|
30 |
|
31 |
Trained with MNLI + Boolq
|
32 |
|
33 |
+
*Trained by:* Evan Li
|
34 |
|
35 |
+
*Application:* Given a passage and a question, answer the question with yes, no or unsure.
|
36 |
|
37 |
+
*Training Process:* https://blogs.nlmatics.com/2020/03/12/Boolean-Question-Answering-with-Neutral-Labels.html
|
38 |
|
39 |
### roberta.large.qa
|
40 |
See folder: [roberta.large.qa](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.qa)
|
41 |
|
42 |
Trained with SQuAD 2.0 + Custom Dataset preferring shorter spans better suited for data extraction
|
43 |
|
44 |
+
*Trained by:* Ambika Sukla
|
45 |
|
46 |
+
*Application:* Given a passage and a question, pick the shortest span from the passage that answers the question
|
47 |
|
48 |
+
*Training Process:* start, end location head on the top of Roberta Base
|
49 |
|
50 |
### roberta.large.stsb
|
51 |
+
*See folder:* [roberta.large.stsb](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.stsb)
|
52 |
|
53 |
Trained with STSB dataset
|
54 |
|
55 |
+
*Trained by:* Meta/Fairseq
|
56 |
|
57 |
+
*Application:* Given two passages, return a score beteen 0 and 1 to evaluate their similarity
|
58 |
|
59 |
+
*Training Process:* regression head on top of Roberta Base
|
60 |
|
61 |
### roberta.large.phraseqa
|
62 |
+
*See folder:* [roberta.large.phraseqa](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.phraseqa)
|
63 |
|
64 |
Trained with Roberta 2.0 with the question words removed from the question
|
65 |
|
66 |
+
*Trained By:* Batya Stein
|
67 |
|
68 |
+
*Application:* Given a passage and phrase (key), extract a value from the passage
|
69 |
|
70 |
+
*Training Process:* https://blogs.nlmatics.com/2020/08/25/Optimizing-Transformer-Q&A-Models-for-Naturalistic-Search.html
|
71 |
|
72 |
### roberta.large.qasrl
|
73 |
|
74 |
+
*See folder:* [roberta.large.qasrl](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.qasrl)
|
75 |
|
76 |
Trained with QASRL dataset
|
77 |
|
78 |
+
*Application:* Given a passage, get back values for who, what, when, where etc.
|
79 |
|
80 |
+
*Trained By:* Nima Sheikholeslami
|
81 |
|
82 |
### roberta.large.qatype.lower.RothWithQ
|
83 |
|
84 |
+
*See folder:* [roberta.large.qatype.lower.RothWithQ](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.qatype.lower.RothWithQ)
|
85 |
|
86 |
Trained with the Roth Question Type dataset.
|
87 |
|
88 |
+
*Application:* Given a question, return one of the answer types e.g. number, location. See the Roth dataset for full list.
|
89 |
|
90 |
+
*Trained By:* Evan Li
|
91 |
|
92 |
### roberta.large.io_qa
|
93 |
|
94 |
See folder: [roberta.large.io_qa](https://huggingface.co/ansukla/roberta/tree/main/roberta.large.io_qa)
|
95 |
Trained with SQuAD 2.0 dataset
|
96 |
|
97 |
+
*Trained By:* Nima Sheikholeslami
|
98 |
|
99 |
+
*Training Process:* Use io head to support multiple spans.
|
100 |
|