core.training_args
core.training_args
extra axolotl specific training args
Classes
| Name | Description |
|---|---|
| AxolotlCPOConfig | CPO config for CPO training |
| AxolotlKTOConfig | KTO config for KTO training |
| AxolotlORPOConfig | ORPO config for ORPO training |
| AxolotlPRMConfig | PRM config for PRM training |
| AxolotlRewardConfig | Reward config for Reward training |
| AxolotlTrainingArguments | Training arguments for Causal trainer |
AxolotlCPOConfig
core.training_args.AxolotlCPOConfig(simpo_gamma=None)CPO config for CPO training
AxolotlKTOConfig
core.training_args.AxolotlKTOConfig()KTO config for KTO training
AxolotlORPOConfig
core.training_args.AxolotlORPOConfig()ORPO config for ORPO training
AxolotlPRMConfig
core.training_args.AxolotlPRMConfig()PRM config for PRM training
AxolotlRewardConfig
core.training_args.AxolotlRewardConfig()Reward config for Reward training
AxolotlTrainingArguments
core.training_args.AxolotlTrainingArguments()Training arguments for Causal trainer
This code is duplicated due to HF TrainingArguments not setting output_dir with a default value so it can’t be used as a mixin.