You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 8 Next »

Test



STEPS:48162032
CFG:1

2

3

4

6

8

Test



48162032
1

2

3

4

6

8


Test



STEP:48162032
CFG:1

2

3

4

6

8



System

app: sdnext.git
updated: 2025-07-18
hash: 4ef0a4d5
url: https://github.com/vladmandic/sdnext.git/tree/dev
arch: x86_64
cpu: x86_64
system: Linux
release: 6.11.0-29-generic
python: 3.12.3
2.7.1+xpu
device: Intel(R) Arc(TM) Graphics (1)
ipex: 
ram: free:123.65 used:1.68 total:125.33
xformers: 
diffusers: 0.35.0.dev0
transformers: 4.53.2
active: xpu
dtype: torch.bfloat16
vae: torch.bfloat16
unet: torch.bfloat16

Model

Model: Diffusers/THUDM/CogView4-6B
Type: cogview4
Class: CogView4Pipeline
Size: 0 bytes
Modified: 2025-07-17 21:11:21


Module

Class

Device

DType

Params

Modules

Config

tokenizer

PreTrainedTokenizerFast

None

None

0

0

None

text_encoder

GlmForCausalLM

xpu:0

torch.bfloat16

9399951360

487

GlmConfig { "architectures": [ "GlmModel" ], "attention_bias": true, "attention_dropout": 0.0, "eos_token_id": [ 151329, 151336, 151338 ], "head_dim": 128, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 13696, "max_position_embeddings": 8192, "model_type": "glm", "num_attention_heads": 32, "num_hidden_layers": 40, "num_key_value_heads": 2, "pad_token_id": 151329, "partial_rotary_factor": 0.5, "rms_norm_eps": 1.5625e-07, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "transformers_version": "4.53.2", "use_cache": true, "vocab_size": 151552 }

vae

AutoencoderKL

xpu:0

torch.bfloat16

406100643

281

FrozenDict({'in_channels': 3, 'out_channels': 3, 'down_block_types': ['DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D'], 'up_block_types': ['UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D'], 'block_out_channels': [128, 512, 1024, 1024], 'layers_per_block': 3, 'act_fn': 'silu', 'latent_channels': 16, 'norm_num_groups': 32, 'sample_size': 1024, 'scaling_factor': 1.0, 'shift_factor': 0.0, 'latents_mean': None, 'latents_std': None, 'force_upcast': True, 'use_quant_conv': False, 'use_post_quant_conv': False, 'mid_block_add_attention': False, '_class_name': 'AutoencoderKL', '_diffusers_version': '0.33.0.dev0', '_name_or_path': '/mnt/models/Diffusers/models--THUDM--CogView4-6B/snapshots/63a52b7f6dace7033380cd6da14d0915eab3e6b5/vae'})

transformer

CogView4Transformer2DModel

xpu:0

torch.bfloat16

6369118272

638

FrozenDict({'patch_size': 2, 'in_channels': 16, 'out_channels': 16, 'num_layers': 28, 'attention_head_dim': 128, 'num_attention_heads': 32, 'text_embed_dim': 4096, 'time_embed_dim': 512, 'condition_dim': 256, 'pos_embed_max_size': 128, 'sample_size': 128, 'rope_axes_dim': [256, 256], '_class_name': 'CogView4Transformer2DModel', '_diffusers_version': '0.33.0.dev0', '_name_or_path': 'THUDM/CogView4-6B'})

scheduler

FlowMatchEulerDiscreteScheduler

None

None

0

0

FrozenDict({'num_train_timesteps': 1000, 'shift': 1.0, 'use_dynamic_shifting': True, 'base_shift': 0.25, 'max_shift': 0.75, 'base_image_seq_len': 256, 'max_image_seq_len': 4096, 'invert_sigmas': False, 'shift_terminal': None, 'use_karras_sigmas': False, 'use_exponential_sigmas': False, 'use_beta_sigmas': False, 'time_shift_type': 'linear', 'stochastic_sampling': False, '_use_default_values': ['stochastic_sampling'], '_class_name': 'FlowMatchEulerDiscreteScheduler', '_diffusers_version': '0.33.0.dev0'})

_name_or_path

str

None

None

0

0

None


  • No labels