bridge.models.glm_moe_dsa.glm5_bridge#
Module Contents#
Classes#
Megatron Bridge for GLM-5 (MoE + MLA + DSA). |
Data#
API#
- bridge.models.glm_moe_dsa.glm5_bridge.logger#
‘getLogger(…)’
- class bridge.models.glm_moe_dsa.glm5_bridge.GLM5Bridge#
Bases:
megatron.bridge.models.conversion.model_bridge.MegatronModelBridgeMegatron Bridge for GLM-5 (MoE + MLA + DSA).
This bridge handles conversion between HuggingFace GlmMoeDsaForCausalLM and Megatron-Core GPTModel formats.
GLM-5 uses Multi-Latent Attention (MLA), Dynamic Sparse Attention (DSA) indexer layers, and Mixture-of-Experts (MoE). Requires transformers>=5.2.0.
.. rubric:: Example
from megatron.bridge import AutoBridge bridge = AutoBridge.from_hf_pretrained(“zai-org/GLM-5”) provider = bridge.to_megatron_provider()
- provider_bridge(
- hf_pretrained: megatron.bridge.models.hf_pretrained.causal_lm.PreTrainedCausalLM,
- mapping_registry() megatron.bridge.models.conversion.mapping_registry.MegatronMappingRegistry#