bridge.recipes.gemma3_vl.gemma3_vl#
Gemma3-VL finetuning recipes with parameterless API.
This module provides SFT and PEFT configurations for Gemma3-VL models (4B, 12B, 27B).
Module Contents#
Functions#
Return a full SFT config for Gemma3-VL 4B Instruct. |
|
Return a full SFT config for Gemma3-VL 12B Instruct. |
|
Return a full SFT config for Gemma3-VL 27B Instruct. |
|
Return a PEFT config for Gemma3-VL 4B Instruct. |
|
Return a PEFT config for Gemma3-VL 12B Instruct. |
|
Return a PEFT config for Gemma3-VL 27B Instruct. |
API#
- bridge.recipes.gemma3_vl.gemma3_vl.gemma3_vl_4b_sft_config() megatron.bridge.training.config.ConfigContainer#
Return a full SFT config for Gemma3-VL 4B Instruct.
Default configuration: 1 node, 8 GPUs
TP=1, PP=1
LR=5e-6 (full SFT)
Sequence length: 4096
- bridge.recipes.gemma3_vl.gemma3_vl.gemma3_vl_12b_sft_config() megatron.bridge.training.config.ConfigContainer#
Return a full SFT config for Gemma3-VL 12B Instruct.
Default configuration: 1 node, 8 GPUs
TP=4, PP=1
LR=5e-6 (full SFT)
Sequence length: 4096
- bridge.recipes.gemma3_vl.gemma3_vl.gemma3_vl_27b_sft_config() megatron.bridge.training.config.ConfigContainer#
Return a full SFT config for Gemma3-VL 27B Instruct.
Default configuration: 2 nodes, 16 GPUs total
TP=8, PP=2
LR=5e-6 (full SFT)
Sequence length: 4096
- bridge.recipes.gemma3_vl.gemma3_vl.gemma3_vl_4b_peft_config(
- peft_scheme: str | megatron.bridge.peft.base.PEFT = 'lora',
Return a PEFT config for Gemma3-VL 4B Instruct.
Default configuration: 1 node, 8 GPUs
TP=1, PP=1
LR=1e-4 (PEFT)
Sequence length: 4096
- Parameters:
peft_scheme – PEFT scheme - “lora”, “dora”, or a custom PEFT instance.
- bridge.recipes.gemma3_vl.gemma3_vl.gemma3_vl_12b_peft_config(
- peft_scheme: str | megatron.bridge.peft.base.PEFT = 'lora',
Return a PEFT config for Gemma3-VL 12B Instruct.
Default configuration: 1 node, 8 GPUs
TP=1, PP=1 (lower than SFT for PEFT)
LR=1e-4 (PEFT)
Sequence length: 4096
- Parameters:
peft_scheme – PEFT scheme - “lora”, “dora”, or a custom PEFT instance.
- bridge.recipes.gemma3_vl.gemma3_vl.gemma3_vl_27b_peft_config(
- peft_scheme: str | megatron.bridge.peft.base.PEFT = 'lora',
Return a PEFT config for Gemma3-VL 27B Instruct.
Default configuration: 1 node, 8 GPUs
TP=4, PP=1 (lower than SFT for PEFT)
LR=1e-4 (PEFT)
Sequence length: 4096
- Parameters:
peft_scheme – PEFT scheme - “lora”, “dora”, or a custom PEFT instance.