saddam213 commited on
Commit
dbebe8d
·
verified ·
1 Parent(s): 0f570ee

Initial Upload

Browse files
.gitattributes CHANGED
@@ -33,3 +33,10 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ controlnet/model.onnx.data filter=lfs diff=lfs merge=lfs -text
37
+ Icon.png filter=lfs diff=lfs merge=lfs -text
38
+ Sample.png filter=lfs diff=lfs merge=lfs -text
39
+ Sample2.png filter=lfs diff=lfs merge=lfs -text
40
+ Sample3.png filter=lfs diff=lfs merge=lfs -text
41
+ Sample4.png filter=lfs diff=lfs merge=lfs -text
42
+ unet/model.onnx.data filter=lfs diff=lfs merge=lfs -text
Icon.png ADDED

Git LFS Details

  • SHA256: b6004f711a353cadd0ff4b45523aa73999ed586e0e5974c052c8266f20c16045
  • Pointer size: 132 Bytes
  • Size of remote file: 1.63 MB
README.md ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Zavy Chroma XL v8 - Onnx Olive DirectML Optimized
2
+
3
+ ## Original Model
4
+ https://civitai.com/models/119229?modelVersionId=563988
5
+
6
+ ## C# Inference Demo
7
+ https://github.com/saddam213/OnnxStack
8
+
9
+ ```csharp
10
+ // Create Pipeline
11
+ var pipeline = StableDiffusionXLPipeline.CreatePipeline("D:\\Models\\ZavyChromaXL-onnx");
12
+
13
+ // Prompt
14
+ var promptOptions = new PromptOptions
15
+ {
16
+ Prompt = "Visualize a traditional wooden fishing boat anchored in a quiet harbor, with nets and fishing gear on board."
17
+ };
18
+
19
+ // Run pipeline
20
+ var result = await pipeline.GenerateImageAsync(promptOptions);
21
+
22
+ // Save Image Result
23
+ await result.SaveAsync("Result.png");
24
+ ```
25
+ ## Inference Result
26
+ ![Intro Image](Sample.png)
Sample.png ADDED

Git LFS Details

  • SHA256: daa1a0faaa5dd3763269a7e1d15baf6d9e3b463cd37de2dca59c6a1b7ec3a96c
  • Pointer size: 132 Bytes
  • Size of remote file: 1.62 MB
Sample2.png ADDED

Git LFS Details

  • SHA256: 184ff9857e2fda7bc9486f72a613ae361595b9a035cb0b983cd349c74fc033bd
  • Pointer size: 132 Bytes
  • Size of remote file: 1.47 MB
Sample3.png ADDED

Git LFS Details

  • SHA256: f2296ff7df2f1ee463b94330650d126c6b70faa7646ba68c7f57fa6884e999fd
  • Pointer size: 132 Bytes
  • Size of remote file: 1.45 MB
Sample4.png ADDED

Git LFS Details

  • SHA256: cf2ff14b3841ae1af59041e3666b71ffdd0f10c1896655d11018a1dbc3dc8484
  • Pointer size: 132 Bytes
  • Size of remote file: 1.79 MB
controlnet/config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.19.0.dev0",
4
+ "act_fn": "silu",
5
+ "addition_embed_type": "text_time",
6
+ "addition_embed_type_num_heads": 64,
7
+ "addition_time_embed_dim": 256,
8
+ "attention_head_dim": [
9
+ 5,
10
+ 10,
11
+ 20
12
+ ],
13
+ "block_out_channels": [
14
+ 320,
15
+ 640,
16
+ 1280
17
+ ],
18
+ "center_input_sample": false,
19
+ "class_embed_type": null,
20
+ "class_embeddings_concat": false,
21
+ "conv_in_kernel": 3,
22
+ "conv_out_kernel": 3,
23
+ "cross_attention_dim": 2048,
24
+ "cross_attention_norm": null,
25
+ "down_block_types": [
26
+ "DownBlock2D",
27
+ "CrossAttnDownBlock2D",
28
+ "CrossAttnDownBlock2D"
29
+ ],
30
+ "downsample_padding": 1,
31
+ "dual_cross_attention": false,
32
+ "encoder_hid_dim": null,
33
+ "encoder_hid_dim_type": null,
34
+ "flip_sin_to_cos": true,
35
+ "freq_shift": 0,
36
+ "in_channels": 4,
37
+ "layers_per_block": 2,
38
+ "mid_block_only_cross_attention": null,
39
+ "mid_block_scale_factor": 1,
40
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
41
+ "norm_eps": 1e-05,
42
+ "norm_num_groups": 32,
43
+ "num_attention_heads": null,
44
+ "num_class_embeds": null,
45
+ "only_cross_attention": false,
46
+ "out_channels": 4,
47
+ "projection_class_embeddings_input_dim": 2816,
48
+ "resnet_out_scale_factor": 1.0,
49
+ "resnet_skip_time_act": false,
50
+ "resnet_time_scale_shift": "default",
51
+ "sample_size": 128,
52
+ "time_cond_proj_dim": null,
53
+ "time_embedding_act_fn": null,
54
+ "time_embedding_dim": null,
55
+ "time_embedding_type": "positional",
56
+ "timestep_post_act": null,
57
+ "transformer_layers_per_block": [
58
+ 1,
59
+ 2,
60
+ 10
61
+ ],
62
+ "up_block_types": [
63
+ "CrossAttnUpBlock2D",
64
+ "CrossAttnUpBlock2D",
65
+ "UpBlock2D"
66
+ ],
67
+ "upcast_attention": null,
68
+ "use_linear_projection": true
69
+ }
controlnet/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:921a14a1a653309db5c3a745b36c4fe5ab2c48042261937cf07b6e96b5d71709
3
+ size 1244445
controlnet/model.onnx.data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c79714a0888f675cba0ff2c09038988d6b217a0ece8b8ea9e6a3dfe90dc0fbaf
3
+ size 5134903040
model_index.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionXLPipeline",
3
+ "_diffusers_version": "0.19.0.dev0",
4
+ "force_zeros_for_empty_prompt": true,
5
+ "add_watermarker": null,
6
+ "scheduler": [
7
+ "diffusers",
8
+ "EulerDiscreteScheduler"
9
+ ],
10
+ "text_encoder": [
11
+ "transformers",
12
+ "CLIPTextModel"
13
+ ],
14
+ "text_encoder_2": [
15
+ "transformers",
16
+ "CLIPTextModelWithProjection"
17
+ ],
18
+ "tokenizer": [
19
+ "transformers",
20
+ "CLIPTokenizer"
21
+ ],
22
+ "tokenizer_2": [
23
+ "transformers",
24
+ "CLIPTokenizer"
25
+ ],
26
+ "unet": [
27
+ "diffusers",
28
+ "UNet2DConditionModel"
29
+ ],
30
+ "vae": [
31
+ "diffusers",
32
+ "AutoencoderKL"
33
+ ]
34
+ }
model_template.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "Id": "E79BD653-4E57-4DC3-B5CD-CA4D822435B8",
3
+ "FileVersion": "1",
4
+ "Created": "2024-06-016T00:00:00",
5
+ "Name": "Zavy Chroma XL",
6
+ "ImageIcon": "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/Icon.png",
7
+ "Author": "Zavy",
8
+ "Description": "A model line that should be a continuance of the ZavyMix SD1.5 model for SDXL. The primary focus is to get a similar feeling in style and uniqueness that model had, where it's good at merging magic with realism, really merging them together seamlessly. Of course with the evolution to SDXL this model should have better quality and coherance for a lot of things, including the eyes and teeth than the SD1.5 models. This model has no need to use the refiner for great results, in fact it usually is preferable to not use the refiner. Recommended to use ultimate SD upscaler to get the most amazing results.",
9
+ "Rank": 310,
10
+ "Group": "Online",
11
+ "Template": "SDXL",
12
+ "Category": "StableDiffusion",
13
+ "StableDiffusionTemplate": {
14
+ "PipelineType": "StableDiffusionXL",
15
+ "ModelType": "Base",
16
+ "SampleSize": 1024,
17
+ "TokenizerLength": 768,
18
+ "DiffuserTypes": [
19
+ "TextToImage",
20
+ "ImageToImage",
21
+ "ImageInpaintLegacy"
22
+ ],
23
+ "SchedulerDefaults": {
24
+ "SchedulerType": "EulerAncestral",
25
+ "Steps": 30,
26
+ "StepsMin": 4,
27
+ "StepsMax": 100,
28
+ "Guidance": 5,
29
+ "GuidanceMin": 0,
30
+ "GuidanceMax": 30,
31
+ "TimestepSpacing": "Linspace",
32
+ "BetaSchedule": "ScaledLinear",
33
+ "BetaStart": 0.00085,
34
+ "BetaEnd": 0.012
35
+ }
36
+ },
37
+ "Precision": "F16",
38
+ "MemoryMin": 6,
39
+ "MemoryMax": 14,
40
+ "DownloadSize": 11.2,
41
+ "Website": "https://civitai.com/models/119229",
42
+ "Repository": "https://huggingface.co/TensorStack/ZavyChromaXL-onnx",
43
+ "RepositoryFiles": [
44
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/controlnet/model.onnx",
45
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/controlnet/model.onnx.data",
46
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/controlnet/config.json",
47
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/scheduler/scheduler_config.json",
48
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/text_encoder/model.onnx",
49
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/text_encoder/config.json",
50
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/text_encoder_2/model.onnx",
51
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/text_encoder_2/config.json",
52
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer/merges.txt",
53
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer/model.onnx",
54
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer/special_tokens_map.json",
55
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer/tokenizer_config.json",
56
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer/vocab.json",
57
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer_2/merges.txt",
58
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer_2/model.onnx",
59
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer_2/special_tokens_map.json",
60
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer_2/tokenizer_config.json",
61
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/tokenizer_2/vocab.json",
62
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/unet/model.onnx",
63
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/unet/model.onnx.data",
64
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/unet/config.json",
65
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/vae_decoder/model.onnx",
66
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/vae_decoder/config.json",
67
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/vae_encoder/model.onnx",
68
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/vae_encoder/config.json",
69
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/model_index.json",
70
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/model_template.json",
71
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/README.md"
72
+ ],
73
+ "PreviewImages": [
74
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/Sample.png",
75
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/Sample2.png",
76
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/Sample3.png",
77
+ "https://huggingface.co/TensorStack/ZavyChromaXL-onnx/resolve/main/Sample4.png"
78
+ ],
79
+ "Tags": [
80
+ "GPU",
81
+ "F16"
82
+ ]
83
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "EulerDiscreteScheduler",
3
+ "_diffusers_version": "0.19.0.dev0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "interpolation_type": "linear",
9
+ "num_train_timesteps": 1000,
10
+ "prediction_type": "epsilon",
11
+ "sample_max_value": 1.0,
12
+ "set_alpha_to_one": false,
13
+ "skip_prk_steps": true,
14
+ "steps_offset": 1,
15
+ "timestep_spacing": "leading",
16
+ "trained_betas": null,
17
+ "use_karras_sigmas": false
18
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CLIPTextModel"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 0,
7
+ "dropout": 0.0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "quick_gelu",
10
+ "hidden_size": 768,
11
+ "initializer_factor": 1.0,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 77,
16
+ "model_type": "clip_text_model",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "projection_dim": 768,
21
+ "torch_dtype": "float16",
22
+ "transformers_version": "4.32.0.dev0",
23
+ "vocab_size": 49408
24
+ }
text_encoder/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1099affd7f5ed3c2a3a6bb8497fc28b1e65b691ca08a3ab6d503f9660e367e8c
3
+ size 246481009
text_encoder_2/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CLIPTextModelWithProjection"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 0,
7
+ "dropout": 0.0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_size": 1280,
11
+ "initializer_factor": 1.0,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 5120,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 77,
16
+ "model_type": "clip_text_model",
17
+ "num_attention_heads": 20,
18
+ "num_hidden_layers": 32,
19
+ "pad_token_id": 1,
20
+ "projection_dim": 1280,
21
+ "torch_dtype": "float16",
22
+ "transformers_version": "4.32.0.dev0",
23
+ "vocab_size": 49408
24
+ }
text_encoder_2/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:400190e46fbba5abda50dc14fbcee398d5065b2b9fb810c5b8542b2d54e4c30a
3
+ size 1390289338
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63b7618c80a5bd8d1ee6fe92b28e7b72dde7aaa522963ff083f284501a9ec7df
3
+ size 1683233
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "clean_up_tokenization_spaces": true,
12
+ "do_lower_case": true,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "<|endoftext|>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "errors": "replace",
22
+ "model_max_length": 77,
23
+ "pad_token": "<|endoftext|>",
24
+ "tokenizer_class": "CLIPTokenizer",
25
+ "unk_token": {
26
+ "__type": "AddedToken",
27
+ "content": "<|endoftext|>",
28
+ "lstrip": false,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_2/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_2/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63b7618c80a5bd8d1ee6fe92b28e7b72dde7aaa522963ff083f284501a9ec7df
3
+ size 1683233
tokenizer_2/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "!",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer_2/tokenizer_config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "clean_up_tokenization_spaces": true,
12
+ "do_lower_case": true,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "<|endoftext|>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "errors": "replace",
22
+ "model_max_length": 77,
23
+ "pad_token": "!",
24
+ "tokenizer_class": "CLIPTokenizer",
25
+ "unk_token": {
26
+ "__type": "AddedToken",
27
+ "content": "<|endoftext|>",
28
+ "lstrip": false,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
+ }
tokenizer_2/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.19.0.dev0",
4
+ "act_fn": "silu",
5
+ "addition_embed_type": "text_time",
6
+ "addition_embed_type_num_heads": 64,
7
+ "addition_time_embed_dim": 256,
8
+ "attention_head_dim": [
9
+ 5,
10
+ 10,
11
+ 20
12
+ ],
13
+ "block_out_channels": [
14
+ 320,
15
+ 640,
16
+ 1280
17
+ ],
18
+ "center_input_sample": false,
19
+ "class_embed_type": null,
20
+ "class_embeddings_concat": false,
21
+ "conv_in_kernel": 3,
22
+ "conv_out_kernel": 3,
23
+ "cross_attention_dim": 2048,
24
+ "cross_attention_norm": null,
25
+ "down_block_types": [
26
+ "DownBlock2D",
27
+ "CrossAttnDownBlock2D",
28
+ "CrossAttnDownBlock2D"
29
+ ],
30
+ "downsample_padding": 1,
31
+ "dual_cross_attention": false,
32
+ "encoder_hid_dim": null,
33
+ "encoder_hid_dim_type": null,
34
+ "flip_sin_to_cos": true,
35
+ "freq_shift": 0,
36
+ "in_channels": 4,
37
+ "layers_per_block": 2,
38
+ "mid_block_only_cross_attention": null,
39
+ "mid_block_scale_factor": 1,
40
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
41
+ "norm_eps": 1e-05,
42
+ "norm_num_groups": 32,
43
+ "num_attention_heads": null,
44
+ "num_class_embeds": null,
45
+ "only_cross_attention": false,
46
+ "out_channels": 4,
47
+ "projection_class_embeddings_input_dim": 2816,
48
+ "resnet_out_scale_factor": 1.0,
49
+ "resnet_skip_time_act": false,
50
+ "resnet_time_scale_shift": "default",
51
+ "sample_size": 128,
52
+ "time_cond_proj_dim": null,
53
+ "time_embedding_act_fn": null,
54
+ "time_embedding_dim": null,
55
+ "time_embedding_type": "positional",
56
+ "timestep_post_act": null,
57
+ "transformer_layers_per_block": [
58
+ 1,
59
+ 2,
60
+ 10
61
+ ],
62
+ "up_block_types": [
63
+ "CrossAttnUpBlock2D",
64
+ "CrossAttnUpBlock2D",
65
+ "UpBlock2D"
66
+ ],
67
+ "upcast_attention": null,
68
+ "use_linear_projection": true
69
+ }
unet/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afa82fc59447e0d239343378a57af55a1dae6a04cbc6a02e36bee9637b0b90e9
3
+ size 1242150
unet/model.onnx.data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c79714a0888f675cba0ff2c09038988d6b217a0ece8b8ea9e6a3dfe90dc0fbaf
3
+ size 5134903040
vae_decoder/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.20.0.dev0",
4
+ "_name_or_path": "../sdxl-vae/",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "force_upcast": true,
19
+ "in_channels": 3,
20
+ "latent_channels": 4,
21
+ "layers_per_block": 2,
22
+ "norm_num_groups": 32,
23
+ "out_channels": 3,
24
+ "sample_size": 1024,
25
+ "scaling_factor": 0.13025,
26
+ "up_block_types": [
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D",
29
+ "UpDecoderBlock2D",
30
+ "UpDecoderBlock2D"
31
+ ]
32
+ }
vae_decoder/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20aae7848dcd708f4dfa562eda63a5df6a4434859f74b7e11dfd9553cf0fceb1
3
+ size 99039207
vae_encoder/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.20.0.dev0",
4
+ "_name_or_path": "../sdxl-vae/",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "force_upcast": true,
19
+ "in_channels": 3,
20
+ "latent_channels": 4,
21
+ "layers_per_block": 2,
22
+ "norm_num_groups": 32,
23
+ "out_channels": 3,
24
+ "sample_size": 1024,
25
+ "scaling_factor": 0.13025,
26
+ "up_block_types": [
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D",
29
+ "UpDecoderBlock2D",
30
+ "UpDecoderBlock2D"
31
+ ]
32
+ }
vae_encoder/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f3cc0532c41f52760353218f5c20c9c19b3d056c038d51c7b71c053ce57e599
3
+ size 68391734