Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

CFG5, STEP28Seed: 1620085323Seed:1931701040Seed:4075624134Seed:2736029172
bookshop girl

Image Added

Image Added

Image Added

Image Added

hand and face

Image Added

Image Added

Image Added

Image Added

legs and shoes

Image Added

Image Added

Image Added

Image Added

Test 1 - Bookshop

Prompt: masterpiece, best quality, photorealistic girl in bookshop choosing the book in romantic stories shelf. smiling

...

Test 4 - Other model Covers

Image AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage Added

Test 5 - Art collection

Image Added

Test 6 - Model own example prompts

Image AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage AddedImage Added

System info


Code Block
Sat Nov 15 10:36:54 2025
app: sdnext.git updated: 2025-11-14 hash: c02192870 url: https://github.com/liutyi/sdnext/tree/pytorch
arch: x86_64 cpu: x86_64 system: Linux release: 6.14.0-35-generic 
python: 3.12.3 torch 2.9.1+xpu
device: Intel(R) Arc(TM) Graphics (1) ipex: 
ram: free:121.86 used:3.47 total:125.33
xformers: diffusers: 0.36.0.dev0 transformers: 4.57.1
active: xpu dtype: torch.bfloat16 vae: torch.bfloat16 unet: torch.bfloat16
base: Diffusers/Photoroom/prx-1024-t2i-beta [318a05beb7] refiner: none vae: none te: none unet: none


Config

Code Block
{
 
}


Model info

...

Diffusers/Photoroom/prx-1024-t2i-beta [318a05beb7]
ModuleClassDeviceDtypeQuantParamsModulesConfig
transformerPRXTransformer2DModelxpu:0torch.bfloat16None1170691648303

FrozenDict({'in_channels': 16, 'patch_size': 2, 'context_in_dim': 2304, 'hidden_size': 1792, 'mlp_ratio': 3.5, 'num_heads': 28, 'depth': 16, 'axes_dim': [32, 32], 'theta': 10000, 'time_factor': 1000.0, 'time_max_period': 10000, '_name_or_path': 'Photoroom/prx-1024-t2i-beta'})

schedulerFlowMatchEulerDiscreteSchedulerNoneNoneNone00

FrozenDict({'num_train_timesteps': 1000, 'shift': 3.0, 'use_dynamic_shifting': False, 'base_shift': 0.5, 'max_shift': 1.15, 'base_image_seq_len': 256, 'max_image_seq_len': 4096, 'invert_sigmas': False, 'shift_terminal': None, 'use_karras_sigmas': False, 'use_exponential_sigmas': False, 'use_beta_sigmas': False, 'time_shift_type': 'exponential', 'stochastic_sampling': False, '_use_default_values': ['use_karras_sigmas', 'use_exponential_sigmas', 'max_image_seq_len', 'use_beta_sigmas', 'time_shift_type', 'shift_terminal', 'use_dynamic_shifting',

text_encoderT5GemmaEncoderxpu:0torch.bfloat16None2614341888448

T5GemmaConfig { "architectures": [ "T5GemmaEncoder" ], "attention_bias": false, "attention_dropout": 0.0, "attn_logit_softcapping": 50.0, "bos_token_id": 2, "classifier_dropout_rate": 0.0, "decoder": { "attention_bias": false, "attention_dropout": 0.0, "attn_logit_softcapping": 50.0, "cross_attention_hidden_size": 2304, "dropout_rate": 0.0, "final_logit_softcapping": 30.0, "head_dim": 256, "hidden_activation": "gelu_pytorch_tanh", "hidden_size": 2304, "initializer_range": 0.02, "intermediate_size": 9216, "is_decoder": true, "layer_types": [ "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention" ], "max_position_embeddings": 8192, "model_type": "t5_gemma_module", "num_attention_heads": 8, "num_hidden_layers": 26, "num_key_value_heads": 4, "query_pre_attn_scalar": 256, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "use_cache": true, "vocab_size": 256000 }, "dropout_rate": 0.0, "dtype": "bfloat16", "encoder": { "attention_bias": false, "attention_dropout": 0.0, "attn_logit_softcapping": 50.0, "dropout_rate": 0.0, "final_logit_softcapping": 30.0, "head_dim": 256, "hidden_activation": "gelu_pytorch_tanh", "hidden_size": 2304, "initializer_range": 0.02, "intermediate_size": 9216, "layer_types": [ "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention" ], "max_position_embeddings": 8192, "model_type": "t5_gemma_module", "num_attention_heads": 8, "num_hidden_layers": 26, "num_key_value_heads": 4, "query_pre_attn_scalar": 256, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "use_cache": true, "vocab_size": 256000 }, "eos_token_id": 1, "final_logit_softcapping": 30.0, "head_dim": 256, "hidden_activation": "gelu_pytorch_tanh", "hidden_size": 2304, "initializer_range": 0.02, "intermediate_size": 9216, "is_encoder_decoder": true, "layer_types": [ "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention", "sliding_attention", "full_attention" ], "max_position_embeddings": 8192, "model_type": "t5gemma", "num_attention_heads": 8, "num_hidden_layers": 26, "num_key_value_heads": 4, "pad_token_id": 0, "query_pre_attn_scalar": 256, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "transformers_version": "4.57.1", "use_cache": true, "vocab_size": 256000 }

tokenizerGemmaTokenizerFastNoneNoneNone00

None

vaeAutoencoderKLxpu:0torch.bfloat16None83819683241

FrozenDict({'in_channels': 3, 'out_channels': 3, 'down_block_types': ['DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D'], 'up_block_types': ['UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D'], 'block_out_channels': [128, 256, 512, 512], 'layers_per_block': 2, 'act_fn': 'silu', 'latent_channels': 16, 'norm_num_groups': 32, 'sample_size': 1024, 'scaling_factor': 0.3611, 'shift_factor': 0.1159, 'latents_mean': None, 'latents_std': None, 'force_upcast': True, 'use_quant_conv': False, 'use_post_quant_conv': False, 'mid_block_add_attention': True, '_class_name': 'AutoencoderKL', '_diffusers_version': '0.36.0.dev0', '_name_or_path': '/mnt/models/Diffusers/models--Photoroom--prx-1024-t2i-beta/snapshots/318a05beb7d65d55616e8fc17b325055be0e4756/vae'})

default_sample_sizeintNoneNoneNone00

None