Ozan Oktay commited on
Commit
9f0b167
1 Parent(s): 8cc04d6
Files changed (2) hide show
  1. config.json +3 -2
  2. configuration_cxrbert.py +5 -0
config.json CHANGED
@@ -1,12 +1,13 @@
1
  {
2
- "_name_or_path": "/tmp/hf_demo_may_11th",
3
  "architectures": [
4
  "CXRBertModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.25,
7
  "auto_map": {
8
  "AutoConfig": "configuration_cxrbert.CXRBertConfig",
9
- "AutoModel": "modeling_cxrbert.CXRBertModel"
 
10
  },
11
  "classifier_dropout": null,
12
  "gradient_checkpointing": false,
 
1
  {
2
+ "_name_or_path": "CXR-BERT-specialized-HD_c01b28b3-c621-4c74-8f70-a1cf88137dc9_0",
3
  "architectures": [
4
  "CXRBertModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.25,
7
  "auto_map": {
8
  "AutoConfig": "configuration_cxrbert.CXRBertConfig",
9
+ "AutoModel": "modeling_cxrbert.CXRBertModel",
10
+ "AutoModelForMaskedLM": "modeling_cxrbert.CXRBertModel"
11
  },
12
  "classifier_dropout": null,
13
  "gradient_checkpointing": false,
configuration_cxrbert.py CHANGED
@@ -9,6 +9,11 @@ from transformers import BertConfig
9
 
10
 
11
  class CXRBertConfig(BertConfig):
 
 
 
 
 
12
  model_type = "cxr-bert"
13
 
14
  def __init__(self, projection_size: int = 128, **kwargs: Any) -> None:
 
9
 
10
 
11
  class CXRBertConfig(BertConfig):
12
+ """
13
+ Config class for CXR-BERT model.
14
+ :param projection_size: Dimensionality of the joint latent space.
15
+ """
16
+
17
  model_type = "cxr-bert"
18
 
19
  def __init__(self, projection_size: int = 128, **kwargs: Any) -> None: