bridge.models.glm.glm_moe_mappings#
GLM MoE mapping helpers for fused expert weights.
These are thin aliases around the shared FusedExpertMapping / FusedGatedExpertMapping classes in param_mapping.py. Kept for backwards compatibility with existing imports.
Module Contents#
Classes#
FusedExpertMapping for GLM down-projection expert weights. |
API#
- class bridge.models.glm.glm_moe_mappings.GLMExpertDownProjMapping(
- megatron_param: str,
- hf_param: str,
- permute_dims: tuple[int, ...] | None = None,
- transpose_on_export: bool = False,
Bases:
megatron.bridge.models.conversion.param_mapping.FusedExpertMappingFusedExpertMapping for GLM down-projection expert weights.
GLM down-projection weights are stored transposed relative to Megatron’s layout, so
transpose_on_exportis always enabled.Initialization