snake7gun commited on
Commit
a47128a
·
verified ·
1 Parent(s): c66042f

Upload folder using huggingface_hub

Browse files
model_index.json CHANGED
@@ -23,7 +23,7 @@
23
  ],
24
  "tokenizer": [
25
  "transformers",
26
- "CLIPTokenizerFast"
27
  ],
28
  "tokenizer_2": [
29
  "transformers",
 
23
  ],
24
  "tokenizer": [
25
  "transformers",
26
+ "CLIPTokenizer"
27
  ],
28
  "tokenizer_2": [
29
  "transformers",
text_encoder/config.json CHANGED
@@ -3,49 +3,21 @@
3
  "CLIPTextModel"
4
  ],
5
  "attention_dropout": 0.0,
 
 
6
  "hidden_act": "gelu",
7
  "hidden_size": 32,
8
  "initializer_factor": 1.0,
 
9
  "intermediate_size": 37,
10
  "layer_norm_eps": 1e-05,
11
- "logit_scale_init_value": 2.6592,
12
  "max_position_embeddings": 77,
13
- "model_type": "clip",
14
- "num_attention_heads": 8,
15
- "num_hidden_layers": 1,
16
- "projection_dim": 512,
17
- "text_config": {
18
- "attention_dropout": 0.0,
19
- "hidden_act": "quick_gelu",
20
- "hidden_size": 512,
21
- "initializer_factor": 1.0,
22
- "initializer_range": 0.02,
23
- "intermediate_size": 2048,
24
- "layer_norm_eps": 1e-05,
25
- "max_position_embeddings": 77,
26
- "model_type": "clip_text_model",
27
- "num_attention_heads": 8,
28
- "num_hidden_layers": 12,
29
- "projection_dim": 512,
30
- "vocab_size": 49408
31
- },
32
  "torch_dtype": "float32",
33
- "transformers_version": "4.54.0.dev0",
34
- "vision_config": {
35
- "attention_dropout": 0.0,
36
- "hidden_act": "quick_gelu",
37
- "hidden_size": 768,
38
- "image_size": 224,
39
- "initializer_factor": 1.0,
40
- "initializer_range": 0.02,
41
- "intermediate_size": 3072,
42
- "layer_norm_eps": 1e-05,
43
- "model_type": "clip_vision_model",
44
- "num_attention_heads": 12,
45
- "num_channels": 3,
46
- "num_hidden_layers": 12,
47
- "patch_size": 32,
48
- "projection_dim": 512
49
- },
50
- "vocab_size": 49408
51
  }
 
3
  "CLIPTextModel"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "bos_token_id": 49406,
7
+ "eos_token_id": 49407,
8
  "hidden_act": "gelu",
9
  "hidden_size": 32,
10
  "initializer_factor": 1.0,
11
+ "initializer_range": 0.02,
12
  "intermediate_size": 37,
13
  "layer_norm_eps": 1e-05,
 
14
  "max_position_embeddings": 77,
15
+ "model_type": "clip_text_model",
16
+ "num_attention_heads": 4,
17
+ "num_hidden_layers": 5,
18
+ "pad_token_id": 1,
19
+ "projection_dim": 32,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  "torch_dtype": "float32",
21
+ "transformers_version": "4.51.3",
22
+ "vocab_size": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  }
text_encoder/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17f380bc42c36b6074372fed8b6584b104757a744095b6f759d48e8fcd05d235
3
- size 6363756
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ea36437f6030f4b55001bec74b64b37d793a27cac46cb6cc6c265547dd8a28d
3
+ size 283212
text_encoder_2/config.json CHANGED
@@ -22,7 +22,7 @@
22
  "relative_attention_max_distance": 128,
23
  "relative_attention_num_buckets": 32,
24
  "torch_dtype": "float32",
25
- "transformers_version": "4.54.0.dev0",
26
  "use_cache": true,
27
  "vocab_size": 32128
28
  }
 
22
  "relative_attention_max_distance": 128,
23
  "relative_attention_num_buckets": 32,
24
  "torch_dtype": "float32",
25
+ "transformers_version": "4.51.3",
26
  "use_cache": true,
27
  "vocab_size": 32128
28
  }
text_encoder_2/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6cb317f3d2b86e1dcbda3df0af4371153a24f076a4925722e54790ec7532e9ba
3
  size 4275288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd89137a066a2235d6d33c445166d1309701822a5c34ef3ef1cb5780fde800fd
3
  size 4275288
tokenizer/vocab.json CHANGED
The diff for this file is too large to render. See raw diff
 
transformer/config.json CHANGED
@@ -1,20 +1,19 @@
1
  {
2
  "_class_name": "FluxTransformer2DModel",
3
  "_diffusers_version": "0.35.0.dev0",
4
- "_name_or_path": "black-forest-labs/FLUX.1-Kontext-dev",
5
- "attention_head_dim": 4,
6
  "axes_dims_rope": [
7
  16,
8
  56,
9
  56
10
  ],
11
- "guidance_embeds": true,
12
- "in_channels": 4,
13
- "joint_attention_dim": 8,
14
  "num_attention_heads": 2,
15
  "num_layers": 1,
16
  "num_single_layers": 1,
17
  "out_channels": null,
18
  "patch_size": 1,
19
- "pooled_projection_dim": 4
20
  }
 
1
  {
2
  "_class_name": "FluxTransformer2DModel",
3
  "_diffusers_version": "0.35.0.dev0",
4
+ "attention_head_dim": 16,
 
5
  "axes_dims_rope": [
6
  16,
7
  56,
8
  56
9
  ],
10
+ "guidance_embeds": false,
11
+ "in_channels": 20,
12
+ "joint_attention_dim": 32,
13
  "num_attention_heads": 2,
14
  "num_layers": 1,
15
  "num_single_layers": 1,
16
  "out_channels": null,
17
  "patch_size": 1,
18
+ "pooled_projection_dim": 32
19
  }
transformer/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4edf5900823e06e6450458a03206a387d0066c3a7ea4ebe764d675e45f1552f9
3
- size 39752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c33571d7e575700c18e9fbe7963fa2fe50a11303498933969e1265d2a684da77
3
+ size 284504
vae/config.json CHANGED
@@ -16,7 +16,7 @@
16
  ],
17
  "force_upcast": true,
18
  "in_channels": 3,
19
- "latent_channels": 1,
20
  "latents_mean": null,
21
  "latents_std": null,
22
  "layers_per_block": 1,
 
16
  ],
17
  "force_upcast": true,
18
  "in_channels": 3,
19
+ "latent_channels": 2,
20
  "latents_mean": null,
21
  "latents_std": null,
22
  "layers_per_block": 1,
vae/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d36a13e223ad92a2491beb9f93232a27ac6289212e3d10cb55261e847f9ee804
3
- size 247108196
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb1f6fd1391550257eb361ec4b01913b7b5c82986d8575fe22ee52b51f18eeb9
3
+ size 247163572