utnah commited on
Commit
58ada39
1 Parent(s): 0a62728

Upload folder using huggingface_hub

Browse files
Files changed (44) hide show
  1. 2_1/ViT-L-14_stats.th +0 -0
  2. 2_1/ViT-L-14_stats.th.lock +0 -0
  3. 2_1/decoder_fp16.ckpt +3 -0
  4. 2_1/decoder_fp16.ckpt.lock +0 -0
  5. 2_1/movq_final.ckpt +3 -0
  6. 2_1/movq_final.ckpt.lock +0 -0
  7. 2_1/prior_fp16.ckpt +3 -0
  8. 2_1/prior_fp16.ckpt.lock +0 -0
  9. 2_1/text_encoder/config.json +25 -0
  10. 2_1/text_encoder/config.json.lock +0 -0
  11. 2_1/text_encoder/pytorch_model.bin +3 -0
  12. 2_1/text_encoder/pytorch_model.bin.lock +0 -0
  13. 2_1/text_encoder/sentencepiece.bpe.model +3 -0
  14. 2_1/text_encoder/sentencepiece.bpe.model.lock +0 -0
  15. 2_1/text_encoder/special_tokens_map.json +1 -0
  16. 2_1/text_encoder/special_tokens_map.json.lock +0 -0
  17. 2_1/text_encoder/tokenizer.json +0 -0
  18. 2_1/text_encoder/tokenizer.json.lock +0 -0
  19. 2_1/text_encoder/tokenizer_config.json +1 -0
  20. 2_1/text_encoder/tokenizer_config.json.lock +0 -0
  21. SAM/sam_vit_b_01ec64.pth +3 -0
  22. SAM/sam_vit_h_4b8939.pth +3 -0
  23. SAM/sam_vit_l_0b3195.pth +3 -0
  24. checkpoints/model_large_caption.pth +3 -0
  25. clip_cache/ViT-L-14.pt +3 -0
  26. clip_cache/ViT-L-14_openai_artists.safetensors +3 -0
  27. clip_cache/ViT-L-14_openai_flavors.safetensors +3 -0
  28. clip_cache/ViT-L-14_openai_mediums.safetensors +3 -0
  29. clip_cache/ViT-L-14_openai_movements.safetensors +3 -0
  30. clip_cache/ViT-L-14_openai_negative.safetensors +3 -0
  31. clip_cache/ViT-L-14_openai_trendings.safetensors +3 -0
  32. esrgan/RealESRGAN_x2.pth +3 -0
  33. esrgan/RealESRGAN_x2.pth.lock +0 -0
  34. shap-e/ViT-L-14.pt +3 -0
  35. shap-e/diffusion_config.yaml +3 -0
  36. shap-e/diffusion_config.yaml.lock +0 -0
  37. shap-e/image_cond.pt +3 -0
  38. shap-e/image_cond.pt.lock +0 -0
  39. shap-e/image_cond_config.yaml +13 -0
  40. shap-e/image_cond_config.yaml.lock +0 -0
  41. shap-e/transmitter.pt +3 -0
  42. shap-e/transmitter.pt.lock +0 -0
  43. shap-e/transmitter_config.yaml +80 -0
  44. shap-e/transmitter_config.yaml.lock +0 -0
2_1/ViT-L-14_stats.th ADDED
Binary file (7.08 kB). View file
 
2_1/ViT-L-14_stats.th.lock ADDED
File without changes
2_1/decoder_fp16.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f6eaaee7d104073cfca18fd0318cba0c23e96634d83e15b91669e04d3854c6
3
+ size 2683523457
2_1/decoder_fp16.ckpt.lock ADDED
File without changes
2_1/movq_final.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b2ded4afb21c1a90d106a6aa4d37be7040fe6bce87f08e12bad2ea83d892b9a
3
+ size 271481827
2_1/movq_final.ckpt.lock ADDED
File without changes
2_1/prior_fp16.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:028e6618d320f7f4471daeeb54d47784a053a3d6b7dad377bc4f3c817c8564e3
3
+ size 2044156871
2_1/prior_fp16.ckpt.lock ADDED
File without changes
2_1/text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "XLMRobertaForMaskedLM"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 1024,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 4096,
13
+ "layer_norm_eps": 1e-05,
14
+ "max_position_embeddings": 514,
15
+ "model_type": "xlm-roberta",
16
+ "num_attention_heads": 16,
17
+ "num_hidden_layers": 24,
18
+ "output_past": true,
19
+ "pad_token_id": 1,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.17.0.dev0",
22
+ "type_vocab_size": 1,
23
+ "use_cache": true,
24
+ "vocab_size": 250002
25
+ }
2_1/text_encoder/config.json.lock ADDED
File without changes
2_1/text_encoder/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ad09a3a1ccf35d93c95c08686506fc59c966d77ac1f5833413366c4d327cb2d
3
+ size 2242872365
2_1/text_encoder/pytorch_model.bin.lock ADDED
File without changes
2_1/text_encoder/sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
2_1/text_encoder/sentencepiece.bpe.model.lock ADDED
File without changes
2_1/text_encoder/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
2_1/text_encoder/special_tokens_map.json.lock ADDED
File without changes
2_1/text_encoder/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
2_1/text_encoder/tokenizer.json.lock ADDED
File without changes
2_1/text_encoder/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "xlm-roberta-large", "tokenizer_class": "XLMRobertaTokenizer"}
2_1/text_encoder/tokenizer_config.json.lock ADDED
File without changes
SAM/sam_vit_b_01ec64.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec2df62732614e57411cdcf32a23ffdf28910380d03139ee0f4fcbe91eb8c912
3
+ size 375042383
SAM/sam_vit_h_4b8939.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7bf3b02f3ebf1267aba913ff637d9a2d5c33d3173bb679e46d9f338c26f262e
3
+ size 2564550879
SAM/sam_vit_l_0b3195.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3adcc4315b642a4d2101128f611684e8734c41232a17c648ed1693702a49a622
3
+ size 1249524607
checkpoints/model_large_caption.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d79b3b7c41478b5fe55c35b73ca6f3525a09708289371c6c0fac641e588287e
3
+ size 1785411505
clip_cache/ViT-L-14.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8cca3fd41ae0c99ba7e8951adf17d267cdb84cd88be6f7c2e0eca1737a03836
3
+ size 932768134
clip_cache/ViT-L-14_openai_artists.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bf73197c78825d443a3415f0df3e46d2068a77a874ca62eb575d39b02de76ad
3
+ size 16174296
clip_cache/ViT-L-14_openai_flavors.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:794576a1605913445f44c7d45d891ce171ef9b8c3e91beaf95a25a1afccb1580
3
+ size 155090140
clip_cache/ViT-L-14_openai_mediums.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8a18152cd3e3d72a4b30f2f6a99a9a0c65a2134bb11ddbd1b793038c067cf49
3
+ size 146127
clip_cache/ViT-L-14_openai_movements.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4944e4aa89138c85fcaceecefd17be68239e2fd243c3336d502bc9d9dd1ead56
3
+ size 307408
clip_cache/ViT-L-14_openai_negative.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19c3820ad959eab90dcc98778c38f6e54c0fba98433ed2eafea0c6d1850b4d8c
3
+ size 63184
clip_cache/ViT-L-14_openai_trendings.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a9e6bbc39318dd4bb40a5790d68faca785629068aaccd02bf2854c50bc47532
3
+ size 110799
esrgan/RealESRGAN_x2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c830d067d54fc767b9543a8432f36d91bc2de313584e8bbfe4ac26a47339e899
3
+ size 67061725
esrgan/RealESRGAN_x2.pth.lock ADDED
File without changes
shap-e/ViT-L-14.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8cca3fd41ae0c99ba7e8951adf17d267cdb84cd88be6f7c2e0eca1737a03836
3
+ size 932768134
shap-e/diffusion_config.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ mean_type: x_start
2
+ schedule: exp
3
+ timesteps: 1024
shap-e/diffusion_config.yaml.lock ADDED
File without changes
shap-e/image_cond.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb8072c64bbbcf6910488814d212227de5db291780d4ea99c6152f9346cf12aa
3
+ size 1263925407
shap-e/image_cond.pt.lock ADDED
File without changes
shap-e/image_cond_config.yaml ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ d_latent: 1048576
2
+ inner:
3
+ cond_drop_prob: 0.1
4
+ heads: 8
5
+ init_scale: 0.25
6
+ layers: 24
7
+ name: CLIPImageGridPointDiffusionTransformer
8
+ pos_emb_init_scale: 0.05
9
+ time_token_cond: true
10
+ use_pos_emb: true
11
+ width: 1024
12
+ latent_ctx: 1024
13
+ name: SplitVectorDiffusion
shap-e/image_cond_config.yaml.lock ADDED
File without changes
shap-e/transmitter.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af02a0b85a8abdfb3919584b63c540ba175f6ad4790f574a7fef4617e5acdc3b
3
+ size 1776048343
shap-e/transmitter.pt.lock ADDED
File without changes
shap-e/transmitter_config.yaml ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ encoder:
2
+ cross_attention_dataset: pcl_and_multiview_pcl
3
+ d_latent: 1024
4
+ data_ctx: 1024
5
+ fps_method: first
6
+ heads: 8
7
+ init_scale: 0.25
8
+ inner_batch_size:
9
+ - 16384
10
+ - 20
11
+ input_channels: 6
12
+ latent_bottleneck:
13
+ diffusion:
14
+ schedule: inv_parabola
15
+ schedule_args:
16
+ power: 5.0
17
+ timesteps: 1024
18
+ diffusion_prob: 0.1
19
+ name: clamp_diffusion_noise
20
+ layers: 12
21
+ max_depth: 9.0
22
+ max_unrolls: 1
23
+ min_unrolls: 1
24
+ name: PointCloudPerceiverChannelsEncoder
25
+ params_proj:
26
+ init_scale: 1.0
27
+ learned_scale: 0.0625
28
+ name: channels
29
+ use_ln: true
30
+ patch_size: 8
31
+ pointconv_hidden:
32
+ - 1024
33
+ - 1024
34
+ pointconv_padding_mode: circular
35
+ pointconv_patch_size: 8
36
+ pointconv_samples: 64
37
+ pointconv_stride: 4
38
+ pos_emb: nerf
39
+ use_depth: true
40
+ use_pointconv: true
41
+ width: 1024
42
+ name: Transmitter
43
+ renderer:
44
+ grid_size: 128
45
+ n_coarse_samples: 64
46
+ n_fine_samples: 128
47
+ name: NeRSTFRenderer
48
+ nerstf:
49
+ activation: swish
50
+ d_hidden: 256
51
+ density_activation: relu
52
+ init_scale: 0.25
53
+ initial_density_bias: 0.1
54
+ insert_direction_at: 4
55
+ meta_bias: false
56
+ meta_parameters: true
57
+ n_hidden_layers: 6
58
+ n_meta_layers: 4
59
+ name: MLPNeRSTFModel
60
+ posenc_version: nerf
61
+ separate_coarse_channels: true
62
+ separate_nerf_channels: true
63
+ trainable_meta: false
64
+ separate_shared_samples: true
65
+ void:
66
+ background:
67
+ - 0
68
+ - 0
69
+ - 0
70
+ name: VoidNeRFModel
71
+ volume:
72
+ bbox_max:
73
+ - 1.0
74
+ - 1.0
75
+ - 1.0
76
+ bbox_min:
77
+ - -1.0
78
+ - -1.0
79
+ - -1.0
80
+ name: BoundingBoxVolume
shap-e/transmitter_config.yaml.lock ADDED
File without changes