bridge.models.olmoe.olmoe_bridge#

Module Contents#

Classes#

OlMoEBridge

Megatron Bridge for OlMoE Models.

Data#

API#

bridge.models.olmoe.olmoe_bridge.logger#

‘getLogger(…)’

class bridge.models.olmoe.olmoe_bridge.OlMoEBridge#

Bases: megatron.bridge.models.conversion.model_bridge.MegatronModelBridge

Megatron Bridge for OlMoE Models.

This bridge handles the conversion between HuggingFace OlMoEForCausalLM and Megatron-Core GPTModel formats. OlMoE models use mixture of experts architecture with QK layernorm.

.. rubric:: Example

from megatron.bridge import AutoBridge bridge = AutoBridge.from_hf_pretrained(“allenai/OLMoE-1B-7B-0125”) provider = bridge.to_megatron_provider()

provider_bridge(
hf_pretrained: megatron.bridge.models.hf_pretrained.causal_lm.PreTrainedCausalLM,
) megatron.bridge.models.olmoe.olmoe_provider.OlMoEModelProvider#
mapping_registry() megatron.bridge.models.conversion.mapping_registry.MegatronMappingRegistry#