bridge.diffusion.models.wan.wan_provider#
Module Contents#
Classes#
WAN 1.3B model configuration. |
|
WAN 14B model configuration. |
Data#
API#
- bridge.diffusion.models.wan.wan_provider.logger#
‘getLogger(…)’
- class bridge.diffusion.models.wan.wan_provider.WanModelProvider#
Bases:
megatron.bridge.models.transformer_config.TransformerConfig,megatron.bridge.models.model_provider.ModelProviderMixin[megatron.core.models.common.vision_module.vision_module.VisionModule]- crossattn_emb_size: int#
1536
- add_bias_linear: bool#
True
- gated_linear_unit: bool#
False
- num_layers: int#
30
1536
8960
- num_attention_heads: int#
12
- layernorm_epsilon: float#
1e-06
- normalization: str#
‘RMSNorm’
- layernorm_zero_centered_gamma: bool#
False
- layernorm_across_heads: bool#
True
- add_qkv_bias: bool#
True
- rotary_interleaved: bool#
True
- activation_func: Callable#
None
0
- attention_dropout: float#
0
- fp16_lm_cross_entropy: bool#
False
- parallel_output: bool#
True
- bf16: bool#
False
- params_dtype: torch.dtype#
None
- qkv_format: str#
‘thd’
- apply_rope_fusion: bool#
True
- bias_activation_fusion: bool#
True
- seq_length: int#
1024
False
- vocab_size: int#
None
- make_vocab_size_divisible_by: int#
128
- in_channels: int#
16
- out_channels: int#
16
- patch_spatial: int#
2
- patch_temporal: int#
1
- freq_dim: int#
256
- text_len: int#
512
- text_dim: int#
4096
- provide(
- pre_process=None,
- post_process=None,
- vp_stage=None,
- class bridge.diffusion.models.wan.wan_provider.WanModelProvider1_3B#
Bases:
bridge.diffusion.models.wan.wan_provider.WanModelProviderWAN 1.3B model configuration.
Architecture: 30 layers, hidden_size=1536, 12 attention heads, ffn_hidden_size=8960. Default seq_length=1024.
- num_layers: int#
30
1536
8960
- num_attention_heads: int#
12
- crossattn_emb_size: int#
1536
- seq_length: int#
1024
- class bridge.diffusion.models.wan.wan_provider.WanModelProvider14B#
Bases:
bridge.diffusion.models.wan.wan_provider.WanModelProviderWAN 14B model configuration.
Architecture: 40 layers, hidden_size=5120, 40 attention heads, ffn_hidden_size=13824. Default seq_length=1024.
- num_layers: int#
40
5120
13824
- num_attention_heads: int#
40
- crossattn_emb_size: int#
5120
- seq_length: int#
1024