saddam213 commited on
Commit
e30fcc2
·
verified ·
1 Parent(s): 6686444

Initial Upload

Browse files
Icon.png ADDED
README.md ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Fluently v4 LCM - Onnx Olive DirectML Optimized
2
+
3
+ ## Original Model
4
+ https://huggingface.co/fluently/Fluently-v4-LCM
5
+
6
+ ## C# Inference Demo
7
+ https://github.com/TensorStack-AI/OnnxStack
8
+
9
+ ```csharp
10
+ // Create Pipeline
11
+ var pipeline = LatentConsistencyPipeline.CreatePipeline("D:\\Models\\Fluently-v4-LCM-onnx");
12
+
13
+ // Prompt
14
+ var promptOptions = new PromptOptions
15
+ {
16
+ Prompt = "Illustrate a cheerful barista preparing a cup of coffee behind the counter of a cozy cafe."
17
+ };
18
+
19
+ // Run pipeline
20
+ var result = await pipeline.GenerateImageAsync(promptOptions);
21
+
22
+ // Save Image Result
23
+ await result.SaveAsync("Result.png");
24
+ ```
25
+ ## Inference Result
26
+ ![Intro Image](Sample.png)
Sample.png ADDED
Sample2.png ADDED
Sample3.png ADDED
Sample4.png ADDED
controlnet/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffeda6aa8ea2af08f2291585693a78e94e360e4a4fc490b6bdffe63106e8892a
3
+ size 1719495908
model_index.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.27.0.dev0",
4
+ "feature_extractor": [
5
+ null,
6
+ null
7
+ ],
8
+ "image_encoder": [
9
+ null,
10
+ null
11
+ ],
12
+ "requires_safety_checker": true,
13
+ "safety_checker": [
14
+ null,
15
+ null
16
+ ],
17
+ "scheduler": [
18
+ "diffusers",
19
+ "EulerDiscreteScheduler"
20
+ ],
21
+ "text_encoder": [
22
+ "transformers",
23
+ "CLIPTextModel"
24
+ ],
25
+ "tokenizer": [
26
+ "transformers",
27
+ "CLIPTokenizer"
28
+ ],
29
+ "unet": [
30
+ "diffusers",
31
+ "UNet2DConditionModel"
32
+ ],
33
+ "vae": [
34
+ "diffusers",
35
+ "AutoencoderKL"
36
+ ]
37
+ }
model_template.json ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "Id": "AB13DC2A-7F4E-43B1-9BA3-85039D544D18",
3
+ "FileVersion": "1",
4
+ "Created": "2024-06-17T00:00:00",
5
+ "Name": "Fluently v4 LCM",
6
+ "ImageIcon": "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/Icon.png",
7
+ "Author": "fluently",
8
+ "Description": "Latent Consistency Model - This model combines multiple checkpoints and various LoRAs to deliver exceptional results. It focuses on producing high-quality eye rendering and correct anatomical details, while simplifying prompts for better outcomes. The model generates beautifully detailed faces and eyes without additional tools, and activates stunning artistic effects when needed. It excels at inpainting and outpainting, and is particularly adept at creating impressive Anime and Comic Book art.",
9
+ "Rank": 300,
10
+ "Group": "Online",
11
+ "Template": "LCM",
12
+ "Category": "StableDiffusion",
13
+ "StableDiffusionTemplate": {
14
+ "PipelineType": "LatentConsistency",
15
+ "ModelType": "Base",
16
+ "SampleSize": 512,
17
+ "TokenizerLength": 768,
18
+ "DiffuserTypes": [
19
+ "TextToImage",
20
+ "ImageToImage",
21
+ "ImageInpaintLegacy",
22
+ "ControlNet",
23
+ "ControlNetImage"
24
+ ],
25
+ "SchedulerDefaults": {
26
+ "SchedulerType": "LCM",
27
+ "Steps": 6,
28
+ "StepsMin": 1,
29
+ "StepsMax": 100,
30
+ "Guidance": 1,
31
+ "GuidanceMin": 0,
32
+ "GuidanceMax": 2,
33
+ "TimestepSpacing": "Linspace",
34
+ "BetaSchedule": "ScaledLinear",
35
+ "BetaStart": 0.00085,
36
+ "BetaEnd": 0.01,
37
+ "DisableNegativePrompt": true
38
+ }
39
+ },
40
+ "Precision": "F16",
41
+ "MemoryMin": 2.1,
42
+ "MemoryMax": 4.6,
43
+ "DownloadSize": 4.2,
44
+ "Website": "https://huggingface.co/fluently/Fluently-v4-LCM",
45
+ "Repository": "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx",
46
+ "RepositoryFiles": [
47
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/controlnet/model.onnx",
48
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/controlnet/config.json",
49
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/scheduler/scheduler_config.json",
50
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/text_encoder/model.onnx",
51
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/text_encoder/config.json",
52
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/tokenizer/merges.txt",
53
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/tokenizer/model.onnx",
54
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/tokenizer/special_tokens_map.json",
55
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/tokenizer/tokenizer_config.json",
56
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/tokenizer/vocab.json",
57
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/unet/model.onnx",
58
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/unet/config.json",
59
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/vae_decoder/model.onnx",
60
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/vae_decoder/config.json",
61
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/vae_encoder/model.onnx",
62
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/vae_encoder/config.json",
63
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/model_index.json",
64
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/model_template.json",
65
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/README.md"
66
+ ],
67
+ "PreviewImages": [
68
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/Sample.png",
69
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/Sample2.png",
70
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/Sample3.png",
71
+ "https://huggingface.co/TensorStack/Fluently-v4-LCM-onnx/resolve/main/Sample4.png"
72
+ ],
73
+ "Tags": [
74
+ "GPU",
75
+ "F16"
76
+ ]
77
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "EulerDiscreteScheduler",
3
+ "_diffusers_version": "0.27.0.dev0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "interpolation_type": "linear",
9
+ "num_train_timesteps": 1000,
10
+ "prediction_type": "epsilon",
11
+ "rescale_betas_zero_snr": false,
12
+ "set_alpha_to_one": false,
13
+ "sigma_max": null,
14
+ "sigma_min": null,
15
+ "steps_offset": 1,
16
+ "timestep_spacing": "linspace",
17
+ "timestep_type": "discrete",
18
+ "trained_betas": null,
19
+ "use_karras_sigmas": false
20
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CLIPTextModel"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 0,
7
+ "dropout": 0.0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "quick_gelu",
10
+ "hidden_size": 768,
11
+ "initializer_factor": 1.0,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 77,
16
+ "model_type": "clip_text_model",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "projection_dim": 768,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.37.2",
23
+ "vocab_size": 49408
24
+ }
text_encoder/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc364c607b9efe000bde1d9395bee54c3c647ef2bdd32e0eae01ddc1f87bd0db
3
+ size 246481009
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63b7618c80a5bd8d1ee6fe92b28e7b72dde7aaa522963ff083f284501a9ec7df
3
+ size 1683233
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "49406": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "49407": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ }
20
+ },
21
+ "bos_token": "<|startoftext|>",
22
+ "clean_up_tokenization_spaces": true,
23
+ "do_lower_case": true,
24
+ "eos_token": "<|endoftext|>",
25
+ "errors": "replace",
26
+ "model_max_length": 77,
27
+ "pad_token": "<|endoftext|>",
28
+ "tokenizer_class": "CLIPTokenizer",
29
+ "unk_token": "<|endoftext|>"
30
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.27.0.dev0",
4
+ "act_fn": "silu",
5
+ "addition_embed_type": null,
6
+ "addition_embed_type_num_heads": 64,
7
+ "addition_time_embed_dim": null,
8
+ "attention_head_dim": 8,
9
+ "attention_type": "default",
10
+ "block_out_channels": [
11
+ 320,
12
+ 640,
13
+ 1280,
14
+ 1280
15
+ ],
16
+ "center_input_sample": false,
17
+ "class_embed_type": null,
18
+ "class_embeddings_concat": false,
19
+ "conv_in_kernel": 3,
20
+ "conv_out_kernel": 3,
21
+ "cross_attention_dim": 768,
22
+ "cross_attention_norm": null,
23
+ "down_block_types": [
24
+ "CrossAttnDownBlock2D",
25
+ "CrossAttnDownBlock2D",
26
+ "CrossAttnDownBlock2D",
27
+ "DownBlock2D"
28
+ ],
29
+ "downsample_padding": 1,
30
+ "dropout": 0.0,
31
+ "dual_cross_attention": false,
32
+ "encoder_hid_dim": null,
33
+ "encoder_hid_dim_type": null,
34
+ "flip_sin_to_cos": true,
35
+ "freq_shift": 0,
36
+ "in_channels": 4,
37
+ "layers_per_block": 2,
38
+ "mid_block_only_cross_attention": null,
39
+ "mid_block_scale_factor": 1,
40
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
41
+ "norm_eps": 1e-05,
42
+ "norm_num_groups": 32,
43
+ "num_attention_heads": null,
44
+ "num_class_embeds": null,
45
+ "only_cross_attention": false,
46
+ "out_channels": 4,
47
+ "projection_class_embeddings_input_dim": null,
48
+ "resnet_out_scale_factor": 1.0,
49
+ "resnet_skip_time_act": false,
50
+ "resnet_time_scale_shift": "default",
51
+ "reverse_transformer_layers_per_block": null,
52
+ "sample_size": 64,
53
+ "time_cond_proj_dim": null,
54
+ "time_embedding_act_fn": null,
55
+ "time_embedding_dim": null,
56
+ "time_embedding_type": "positional",
57
+ "timestep_post_act": null,
58
+ "transformer_layers_per_block": 1,
59
+ "up_block_types": [
60
+ "UpBlock2D",
61
+ "CrossAttnUpBlock2D",
62
+ "CrossAttnUpBlock2D",
63
+ "CrossAttnUpBlock2D"
64
+ ],
65
+ "upcast_attention": null,
66
+ "use_linear_projection": false
67
+ }
unet/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcf5a54b4481bac0083d6162a7535942b4c0811fea29c608b1ef4a53aed9e20d
3
+ size 1719492909
vae_decoder/config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.27.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 128,
7
+ 256,
8
+ 512,
9
+ 512
10
+ ],
11
+ "down_block_types": [
12
+ "DownEncoderBlock2D",
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D"
16
+ ],
17
+ "force_upcast": true,
18
+ "in_channels": 3,
19
+ "latent_channels": 4,
20
+ "layers_per_block": 2,
21
+ "norm_num_groups": 32,
22
+ "out_channels": 3,
23
+ "sample_size": 512,
24
+ "scaling_factor": 0.18215,
25
+ "up_block_types": [
26
+ "UpDecoderBlock2D",
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D",
29
+ "UpDecoderBlock2D"
30
+ ]
31
+ }
vae_decoder/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ce3158522f3939ab69a62be36cc524a82676285361ab0bebf09b6f540a12483
3
+ size 99039207
vae_encoder/config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.27.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 128,
7
+ 256,
8
+ 512,
9
+ 512
10
+ ],
11
+ "down_block_types": [
12
+ "DownEncoderBlock2D",
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D"
16
+ ],
17
+ "force_upcast": true,
18
+ "in_channels": 3,
19
+ "latent_channels": 4,
20
+ "layers_per_block": 2,
21
+ "norm_num_groups": 32,
22
+ "out_channels": 3,
23
+ "sample_size": 512,
24
+ "scaling_factor": 0.18215,
25
+ "up_block_types": [
26
+ "UpDecoderBlock2D",
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D",
29
+ "UpDecoderBlock2D"
30
+ ]
31
+ }
vae_encoder/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76d04393869dc0f12e75c022b55d0e30c9c9ace2ff9274dda513719d6cdb060d
3
+ size 68391734