bridge.recipes.qwen2_audio.qwen2_audio#
Module Contents#
Functions#
Return a fine-tuning config for Qwen2-Audio 7B Instruct. |
|
Create a fine-tuning configuration for Qwen2-Audio models. |
API#
- bridge.recipes.qwen2_audio.qwen2_audio.qwen2_audio_7b_finetune_config(
- **user_kwargs,
Return a fine-tuning config for Qwen2-Audio 7B Instruct.
Default configuration: 1 node, TP=1, PP=1
LoRA/DoRA: LR=1e-4
Full SFT: LR=5e-6
See
_qwen2_audio_commonfor the full list of parameters.
- bridge.recipes.qwen2_audio.qwen2_audio._qwen2_audio_common(
- hf_path: str,
- output_dir: str | None = None,
- name: str = 'qwen2_audio_finetune',
- pretrained_checkpoint: Optional[str] = None,
- tensor_model_parallel_size: int = 1,
- pipeline_model_parallel_size: int = 1,
- pipeline_dtype: Optional[torch.dtype] = None,
- virtual_pipeline_model_parallel_size: Optional[int] = None,
- context_parallel_size: int = 1,
- sequence_parallel: bool = False,
- train_iters: int = 2000,
- global_batch_size: int = 32,
- micro_batch_size: int = 1,
- seq_length: int = 4096,
- lr: float = 0.0003,
- min_lr: float = 3e-05,
- lr_warmup_iters: int = 5,
- lr_decay_iters: Optional[int] = None,
- eval_interval: int = 500,
- save_interval: int = 200,
- precision_config: Optional[Union[megatron.bridge.training.mixed_precision.MixedPrecisionConfig, str]] = 'bf16_mixed',
- freeze_language_model: bool = False,
- freeze_audio_model: bool = False,
- freeze_audio_projection: bool = False,
- peft: Optional[Union[str, megatron.bridge.peft.base.PEFT]] = None,
- finetune_lr: Optional[float] = None,
- maker_name: str = 'make_default_audio_dataset',
- maker_kwargs: Optional[dict] = None,
- val_maker_kwargs: Optional[dict] = None,
- test_maker_kwargs: Optional[dict] = None,
- wandb_project: Optional[str] = None,
- wandb_entity: Optional[str] = None,
- wandb_exp_name: Optional[str] = None,
Create a fine-tuning configuration for Qwen2-Audio models.