ares1123 commited on
Commit
c0e1fc7
·
verified ·
1 Parent(s): c152a60

First commit

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. EfficientNetB3_best_model.h5 +3 -0
  2. feature_extractor/preprocessor_config.json +20 -0
  3. pose_estimations.pkl +3 -0
  4. safety_checker/config.json +175 -0
  5. safety_checker/model.safetensors +3 -0
  6. safety_checker/pytorch_model.bin +3 -0
  7. sample_images/12021D006-Q11@11=person_whole_front.jpg +0 -0
  8. sample_images/1MY21D076-A11@12=person_whole_front.jpg +0 -0
  9. sample_images/1MY21D088-Q11@11=person_whole_front.jpg +0 -0
  10. sample_images/1MY21D08A-M11@13=person_whole_front.jpg +0 -0
  11. sample_images/2WF21E000-A11@2.1=person_whole_front.jpg +0 -0
  12. sample_images/4AP21D03K-M11@3.4=person_whole_front.jpg +0 -0
  13. sample_images/4AP21E056-A11@14=person_whole_front.jpg +0 -0
  14. sample_images/4BE21D0A3-C11@10=cloth_front.jpg +0 -0
  15. sample_images/4BE21D0A3-K11@10=cloth_front_mask.jpg +0 -0
  16. sample_images/4JO21E01T-J11@11=person_whole_front.jpg +0 -0
  17. sample_images/4JO21E01T-J11@9=person_whole_side.jpg +0 -0
  18. sample_images/4VI21D022-C11@11=person_whole_front.jpg +0 -0
  19. sample_images/4VI21D022-C11@9=person_whole_back.jpg +0 -0
  20. sample_images/4VI21D025-K11@11=person_whole_front.jpg +0 -0
  21. sample_images/4VI21D02C-K11@29.1=cloth_front.jpg +0 -0
  22. sample_images/7SI21D050-J11@9=person_whole_front.jpg +0 -0
  23. sample_images/A0Q21C02B-Q11@9=person_whole_front.jpg +0 -0
  24. sample_images/AA321D00O-Q11@9=person_whole_side.jpg +0 -0
  25. sample_images/AJ121D048-Q11@9=person_whole_front.jpg +0 -0
  26. sample_images/AM321D000-K11@10=cloth_front.jpg +0 -0
  27. sample_images/AN621CAK9-B11@9=person_whole_front.jpg +0 -0
  28. sample_images/AN621D0CY-G11@14.1=person_whole_side.jpg +0 -0
  29. sample_images/AN621D0CY-Q11@9=person_whole_front.jpg +0 -0
  30. sample_images/AN621DAA0-Q11@10=cloth_front.jpg +0 -0
  31. sample_images/AN621EA00-K13@11=person_whole_front.jpg +0 -0
  32. sample_images/AN621EA2Y-G11@15=person_whole_front.jpg +0 -0
  33. sample_images/ARC21D007-A12@7=person_half_back.jpg +0 -0
  34. sample_images/ARC21D007-A12@8=person_half_front.jpg +0 -0
  35. sample_images/AX821D000-G11@9=person_whole_front.jpg +0 -0
  36. sample_images/AX821D00G-Q11@9.1=person_whole_front.jpg +0 -0
  37. sample_images/CE321D0LN-I11@13=person_whole_front.jpg +0 -0
  38. sample_images/CE321D0LR-L11@14=person_whole_front.jpg +0 -0
  39. sample_images/CE321D0M2-A11@14=person_whole_front.jpg +0 -0
  40. sample_images/CU721E01G-C11@11.1=cloth_front.jpg +0 -0
  41. sample_images/DE121D0F5-K11@12=cloth_front.jpg +0 -0
  42. sample_images/DOA21E00H-Q11@9=person_whole_front.jpg +0 -0
  43. sample_images/DP521D0EL-J11@9=person_whole_side.jpg +0 -0
  44. sample_images/DP521E0PH-T11@13=person_whole_front.jpg +0 -0
  45. sample_images/DP521E0SD-G11@9=person_whole_front.jpg +0 -0
  46. sample_images/DP521E0SY-G11@33.1=person_whole_front.jpg +0 -0
  47. sample_images/DP621E05D-K11@9=person_whole_front.jpg +0 -0
  48. sample_images/DP621E05U-Q11@13=person_whole_front.jpg +0 -0
  49. sample_images/DP721J005-C11@10=cloth_front.jpg +0 -0
  50. sample_images/DP721J005-C11@6=person_whole_back.jpg +0 -0
EfficientNetB3_best_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c702c42ec094b3bbab82a00eb50e844c2e3ff19f25a812f716edab023d1ad760
3
+ size 134
feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": 224,
3
+ "do_center_crop": true,
4
+ "do_convert_rgb": true,
5
+ "do_normalize": true,
6
+ "do_resize": true,
7
+ "feature_extractor_type": "CLIPFeatureExtractor",
8
+ "image_mean": [
9
+ 0.48145466,
10
+ 0.4578275,
11
+ 0.40821073
12
+ ],
13
+ "image_std": [
14
+ 0.26862954,
15
+ 0.26130258,
16
+ 0.27577711
17
+ ],
18
+ "resample": 3,
19
+ "size": 224
20
+ }
pose_estimations.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dcfdfd15820f2ce60ac80179c067aabf85174b0e2c6dabf23203a76f2a0b1e1
3
+ size 132
safety_checker/config.json ADDED
@@ -0,0 +1,175 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_commit_hash": "4bb648a606ef040e7685bde262611766a5fdd67b",
3
+ "_name_or_path": "CompVis/stable-diffusion-safety-checker",
4
+ "architectures": [
5
+ "StableDiffusionSafetyChecker"
6
+ ],
7
+ "initializer_factor": 1.0,
8
+ "logit_scale_init_value": 2.6592,
9
+ "model_type": "clip",
10
+ "projection_dim": 768,
11
+ "text_config": {
12
+ "_name_or_path": "",
13
+ "add_cross_attention": false,
14
+ "architectures": null,
15
+ "attention_dropout": 0.0,
16
+ "bad_words_ids": null,
17
+ "bos_token_id": 0,
18
+ "chunk_size_feed_forward": 0,
19
+ "cross_attention_hidden_size": null,
20
+ "decoder_start_token_id": null,
21
+ "diversity_penalty": 0.0,
22
+ "do_sample": false,
23
+ "dropout": 0.0,
24
+ "early_stopping": false,
25
+ "encoder_no_repeat_ngram_size": 0,
26
+ "eos_token_id": 2,
27
+ "exponential_decay_length_penalty": null,
28
+ "finetuning_task": null,
29
+ "forced_bos_token_id": null,
30
+ "forced_eos_token_id": null,
31
+ "hidden_act": "quick_gelu",
32
+ "hidden_size": 768,
33
+ "id2label": {
34
+ "0": "LABEL_0",
35
+ "1": "LABEL_1"
36
+ },
37
+ "initializer_factor": 1.0,
38
+ "initializer_range": 0.02,
39
+ "intermediate_size": 3072,
40
+ "is_decoder": false,
41
+ "is_encoder_decoder": false,
42
+ "label2id": {
43
+ "LABEL_0": 0,
44
+ "LABEL_1": 1
45
+ },
46
+ "layer_norm_eps": 1e-05,
47
+ "length_penalty": 1.0,
48
+ "max_length": 20,
49
+ "max_position_embeddings": 77,
50
+ "min_length": 0,
51
+ "model_type": "clip_text_model",
52
+ "no_repeat_ngram_size": 0,
53
+ "num_attention_heads": 12,
54
+ "num_beam_groups": 1,
55
+ "num_beams": 1,
56
+ "num_hidden_layers": 12,
57
+ "num_return_sequences": 1,
58
+ "output_attentions": false,
59
+ "output_hidden_states": false,
60
+ "output_scores": false,
61
+ "pad_token_id": 1,
62
+ "prefix": null,
63
+ "problem_type": null,
64
+ "pruned_heads": {},
65
+ "remove_invalid_values": false,
66
+ "repetition_penalty": 1.0,
67
+ "return_dict": true,
68
+ "return_dict_in_generate": false,
69
+ "sep_token_id": null,
70
+ "task_specific_params": null,
71
+ "temperature": 1.0,
72
+ "tf_legacy_loss": false,
73
+ "tie_encoder_decoder": false,
74
+ "tie_word_embeddings": true,
75
+ "tokenizer_class": null,
76
+ "top_k": 50,
77
+ "top_p": 1.0,
78
+ "torch_dtype": null,
79
+ "torchscript": false,
80
+ "transformers_version": "4.22.0.dev0",
81
+ "typical_p": 1.0,
82
+ "use_bfloat16": false,
83
+ "vocab_size": 49408
84
+ },
85
+ "text_config_dict": {
86
+ "hidden_size": 768,
87
+ "intermediate_size": 3072,
88
+ "num_attention_heads": 12,
89
+ "num_hidden_layers": 12
90
+ },
91
+ "torch_dtype": "float32",
92
+ "transformers_version": null,
93
+ "vision_config": {
94
+ "_name_or_path": "",
95
+ "add_cross_attention": false,
96
+ "architectures": null,
97
+ "attention_dropout": 0.0,
98
+ "bad_words_ids": null,
99
+ "bos_token_id": null,
100
+ "chunk_size_feed_forward": 0,
101
+ "cross_attention_hidden_size": null,
102
+ "decoder_start_token_id": null,
103
+ "diversity_penalty": 0.0,
104
+ "do_sample": false,
105
+ "dropout": 0.0,
106
+ "early_stopping": false,
107
+ "encoder_no_repeat_ngram_size": 0,
108
+ "eos_token_id": null,
109
+ "exponential_decay_length_penalty": null,
110
+ "finetuning_task": null,
111
+ "forced_bos_token_id": null,
112
+ "forced_eos_token_id": null,
113
+ "hidden_act": "quick_gelu",
114
+ "hidden_size": 1024,
115
+ "id2label": {
116
+ "0": "LABEL_0",
117
+ "1": "LABEL_1"
118
+ },
119
+ "image_size": 224,
120
+ "initializer_factor": 1.0,
121
+ "initializer_range": 0.02,
122
+ "intermediate_size": 4096,
123
+ "is_decoder": false,
124
+ "is_encoder_decoder": false,
125
+ "label2id": {
126
+ "LABEL_0": 0,
127
+ "LABEL_1": 1
128
+ },
129
+ "layer_norm_eps": 1e-05,
130
+ "length_penalty": 1.0,
131
+ "max_length": 20,
132
+ "min_length": 0,
133
+ "model_type": "clip_vision_model",
134
+ "no_repeat_ngram_size": 0,
135
+ "num_attention_heads": 16,
136
+ "num_beam_groups": 1,
137
+ "num_beams": 1,
138
+ "num_channels": 3,
139
+ "num_hidden_layers": 24,
140
+ "num_return_sequences": 1,
141
+ "output_attentions": false,
142
+ "output_hidden_states": false,
143
+ "output_scores": false,
144
+ "pad_token_id": null,
145
+ "patch_size": 14,
146
+ "prefix": null,
147
+ "problem_type": null,
148
+ "pruned_heads": {},
149
+ "remove_invalid_values": false,
150
+ "repetition_penalty": 1.0,
151
+ "return_dict": true,
152
+ "return_dict_in_generate": false,
153
+ "sep_token_id": null,
154
+ "task_specific_params": null,
155
+ "temperature": 1.0,
156
+ "tf_legacy_loss": false,
157
+ "tie_encoder_decoder": false,
158
+ "tie_word_embeddings": true,
159
+ "tokenizer_class": null,
160
+ "top_k": 50,
161
+ "top_p": 1.0,
162
+ "torch_dtype": null,
163
+ "torchscript": false,
164
+ "transformers_version": "4.22.0.dev0",
165
+ "typical_p": 1.0,
166
+ "use_bfloat16": false
167
+ },
168
+ "vision_config_dict": {
169
+ "hidden_size": 1024,
170
+ "intermediate_size": 4096,
171
+ "num_attention_heads": 16,
172
+ "num_hidden_layers": 24,
173
+ "patch_size": 14
174
+ }
175
+ }
safety_checker/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:202da5e2d9477d6098113bd1603ba8983ab5ceb3522cd621f9abe461e0137b94
3
+ size 135
safety_checker/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c85428012ca5bc371c1b876e7b3dc12c6b7beedff663037bcb35895f096ae1f
3
+ size 135
sample_images/12021D006-Q11@11=person_whole_front.jpg ADDED
sample_images/1MY21D076-A11@12=person_whole_front.jpg ADDED
sample_images/1MY21D088-Q11@11=person_whole_front.jpg ADDED
sample_images/1MY21D08A-M11@13=person_whole_front.jpg ADDED
sample_images/2WF21E000-A11@2.1=person_whole_front.jpg ADDED
sample_images/4AP21D03K-M11@3.4=person_whole_front.jpg ADDED
sample_images/4AP21E056-A11@14=person_whole_front.jpg ADDED
sample_images/4BE21D0A3-C11@10=cloth_front.jpg ADDED
sample_images/4BE21D0A3-K11@10=cloth_front_mask.jpg ADDED
sample_images/4JO21E01T-J11@11=person_whole_front.jpg ADDED
sample_images/4JO21E01T-J11@9=person_whole_side.jpg ADDED
sample_images/4VI21D022-C11@11=person_whole_front.jpg ADDED
sample_images/4VI21D022-C11@9=person_whole_back.jpg ADDED
sample_images/4VI21D025-K11@11=person_whole_front.jpg ADDED
sample_images/4VI21D02C-K11@29.1=cloth_front.jpg ADDED
sample_images/7SI21D050-J11@9=person_whole_front.jpg ADDED
sample_images/A0Q21C02B-Q11@9=person_whole_front.jpg ADDED
sample_images/AA321D00O-Q11@9=person_whole_side.jpg ADDED
sample_images/AJ121D048-Q11@9=person_whole_front.jpg ADDED
sample_images/AM321D000-K11@10=cloth_front.jpg ADDED
sample_images/AN621CAK9-B11@9=person_whole_front.jpg ADDED
sample_images/AN621D0CY-G11@14.1=person_whole_side.jpg ADDED
sample_images/AN621D0CY-Q11@9=person_whole_front.jpg ADDED
sample_images/AN621DAA0-Q11@10=cloth_front.jpg ADDED
sample_images/AN621EA00-K13@11=person_whole_front.jpg ADDED
sample_images/AN621EA2Y-G11@15=person_whole_front.jpg ADDED
sample_images/ARC21D007-A12@7=person_half_back.jpg ADDED
sample_images/ARC21D007-A12@8=person_half_front.jpg ADDED
sample_images/AX821D000-G11@9=person_whole_front.jpg ADDED
sample_images/AX821D00G-Q11@9.1=person_whole_front.jpg ADDED
sample_images/CE321D0LN-I11@13=person_whole_front.jpg ADDED
sample_images/CE321D0LR-L11@14=person_whole_front.jpg ADDED
sample_images/CE321D0M2-A11@14=person_whole_front.jpg ADDED
sample_images/CU721E01G-C11@11.1=cloth_front.jpg ADDED
sample_images/DE121D0F5-K11@12=cloth_front.jpg ADDED
sample_images/DOA21E00H-Q11@9=person_whole_front.jpg ADDED
sample_images/DP521D0EL-J11@9=person_whole_side.jpg ADDED
sample_images/DP521E0PH-T11@13=person_whole_front.jpg ADDED
sample_images/DP521E0SD-G11@9=person_whole_front.jpg ADDED
sample_images/DP521E0SY-G11@33.1=person_whole_front.jpg ADDED
sample_images/DP621E05D-K11@9=person_whole_front.jpg ADDED
sample_images/DP621E05U-Q11@13=person_whole_front.jpg ADDED
sample_images/DP721J005-C11@10=cloth_front.jpg ADDED
sample_images/DP721J005-C11@6=person_whole_back.jpg ADDED