core.training_args
core.training_args
extra axolotl specific training args
Classes
Name | Description |
---|---|
AxolotlCPOConfig | CPO config for CPO training |
AxolotlKTOConfig | KTO config for KTO training |
AxolotlORPOConfig | ORPO config for ORPO training |
AxolotlPRMConfig | PRM config for PRM training |
AxolotlRewardConfig | Reward config for Reward training |
AxolotlTrainingArguments | Training arguments for Causal trainer |
AxolotlTrainingMixins | Mixin class for the Axolotl training args. |
AxolotlCPOConfig
core.training_args.AxolotlCPOConfig(self,
=None,
model_type=False,
lr_quadratic_warmup=False,
pretraining=False,
sample_packing=False,
multipack_real_batches=None,
eval_sample_packing=1.0,
sample_packing_efficiency=200,
sample_packing_bin_size=100000,
sample_packing_group_size=2048,
max_seq_length=None,
relora_steps=None,
relora_warmup_steps=None,
relora_anneal_steps=0.9,
relora_prune_ratio='eval',
bench_split='pharaouk/dharma-1/dharma_1_mini.json',
bench_dataset=False,
do_bench_eval=False,
do_causal_lm_eval=None,
max_bench_samples=2048,
bench_source_max_len=None,
dataloader_prefetch_factor=None,
cosine_min_lr_ratio=None,
cosine_constant_lr_ratio=None,
loraplus_lr_ratio=1e-06,
loraplus_lr_embedding=None,
embedding_lr_scale=None,
lr_groups=None,
embedding_lr=False,
qlora=None,
orpo_alpha=None,
lisa_n_layers=None,
lisa_step_interval=None,
lisa_layers_attribute=None,
curriculum_sampling=None,
alternate_optimizer=None,
alternate_lr_scheduler_type=None,
chat_template=None,
kd_ce_alpha=1.0,
kd_alpha=1.0,
kd_temperature=None,
kd_zscore_base_temp=None,
kd_top_k_before_softmax=1,
sequence_parallel_degree=None,
image_size=None,
image_resize_algorithm=None,
simpo_gamma )
CPO config for CPO training
AxolotlKTOConfig
core.training_args.AxolotlKTOConfig(self,
=None,
model_type=False,
lr_quadratic_warmup=False,
pretraining=False,
sample_packing=False,
multipack_real_batches=None,
eval_sample_packing=1.0,
sample_packing_efficiency=200,
sample_packing_bin_size=100000,
sample_packing_group_size=2048,
max_seq_length=None,
relora_steps=None,
relora_warmup_steps=None,
relora_anneal_steps=0.9,
relora_prune_ratio='eval',
bench_split='pharaouk/dharma-1/dharma_1_mini.json',
bench_dataset=False,
do_bench_eval=False,
do_causal_lm_eval=None,
max_bench_samples=2048,
bench_source_max_len=None,
dataloader_prefetch_factor=None,
cosine_min_lr_ratio=None,
cosine_constant_lr_ratio=None,
loraplus_lr_ratio=1e-06,
loraplus_lr_embedding=None,
embedding_lr_scale=None,
lr_groups=None,
embedding_lr=False,
qlora=None,
orpo_alpha=None,
lisa_n_layers=None,
lisa_step_interval=None,
lisa_layers_attribute=None,
curriculum_sampling=None,
alternate_optimizer=None,
alternate_lr_scheduler_type=None,
chat_template=None,
kd_ce_alpha=1.0,
kd_alpha=1.0,
kd_temperature=None,
kd_zscore_base_temp=None,
kd_top_k_before_softmax=1,
sequence_parallel_degree=None,
image_size=None,
image_resize_algorithm )
KTO config for KTO training
AxolotlORPOConfig
core.training_args.AxolotlORPOConfig(self,
=None,
model_type=False,
lr_quadratic_warmup=False,
pretraining=False,
sample_packing=False,
multipack_real_batches=None,
eval_sample_packing=1.0,
sample_packing_efficiency=200,
sample_packing_bin_size=100000,
sample_packing_group_size=2048,
max_seq_length=None,
relora_steps=None,
relora_warmup_steps=None,
relora_anneal_steps=0.9,
relora_prune_ratio='eval',
bench_split='pharaouk/dharma-1/dharma_1_mini.json',
bench_dataset=False,
do_bench_eval=False,
do_causal_lm_eval=None,
max_bench_samples=2048,
bench_source_max_len=None,
dataloader_prefetch_factor=None,
cosine_min_lr_ratio=None,
cosine_constant_lr_ratio=None,
loraplus_lr_ratio=1e-06,
loraplus_lr_embedding=None,
embedding_lr_scale=None,
lr_groups=None,
embedding_lr=False,
qlora=None,
orpo_alpha=None,
lisa_n_layers=None,
lisa_step_interval=None,
lisa_layers_attribute=None,
curriculum_sampling=None,
alternate_optimizer=None,
alternate_lr_scheduler_type=None,
chat_template=None,
kd_ce_alpha=1.0,
kd_alpha=1.0,
kd_temperature=None,
kd_zscore_base_temp=None,
kd_top_k_before_softmax=1,
sequence_parallel_degree=None,
image_size=None,
image_resize_algorithm )
ORPO config for ORPO training
AxolotlPRMConfig
core.training_args.AxolotlPRMConfig(self,
=None,
model_type=False,
lr_quadratic_warmup=False,
pretraining=False,
sample_packing=False,
multipack_real_batches=None,
eval_sample_packing=1.0,
sample_packing_efficiency=200,
sample_packing_bin_size=100000,
sample_packing_group_size=2048,
max_seq_length=None,
relora_steps=None,
relora_warmup_steps=None,
relora_anneal_steps=0.9,
relora_prune_ratio='eval',
bench_split='pharaouk/dharma-1/dharma_1_mini.json',
bench_dataset=False,
do_bench_eval=False,
do_causal_lm_eval=None,
max_bench_samples=2048,
bench_source_max_len=None,
dataloader_prefetch_factor=None,
cosine_min_lr_ratio=None,
cosine_constant_lr_ratio=None,
loraplus_lr_ratio=1e-06,
loraplus_lr_embedding=None,
embedding_lr_scale=None,
lr_groups=None,
embedding_lr=False,
qlora=None,
orpo_alpha=None,
lisa_n_layers=None,
lisa_step_interval=None,
lisa_layers_attribute=None,
curriculum_sampling=None,
alternate_optimizer=None,
alternate_lr_scheduler_type=None,
chat_template=None,
kd_ce_alpha=1.0,
kd_alpha=1.0,
kd_temperature=None,
kd_zscore_base_temp=None,
kd_top_k_before_softmax=1,
sequence_parallel_degree=None,
image_size=None,
image_resize_algorithm )
PRM config for PRM training
AxolotlRewardConfig
core.training_args.AxolotlRewardConfig(self,
=None,
model_type=False,
lr_quadratic_warmup=False,
pretraining=False,
sample_packing=False,
multipack_real_batches=None,
eval_sample_packing=1.0,
sample_packing_efficiency=200,
sample_packing_bin_size=100000,
sample_packing_group_size=2048,
max_seq_length=None,
relora_steps=None,
relora_warmup_steps=None,
relora_anneal_steps=0.9,
relora_prune_ratio='eval',
bench_split='pharaouk/dharma-1/dharma_1_mini.json',
bench_dataset=False,
do_bench_eval=False,
do_causal_lm_eval=None,
max_bench_samples=2048,
bench_source_max_len=None,
dataloader_prefetch_factor=None,
cosine_min_lr_ratio=None,
cosine_constant_lr_ratio=None,
loraplus_lr_ratio=1e-06,
loraplus_lr_embedding=None,
embedding_lr_scale=None,
lr_groups=None,
embedding_lr=False,
qlora=None,
orpo_alpha=None,
lisa_n_layers=None,
lisa_step_interval=None,
lisa_layers_attribute=None,
curriculum_sampling=None,
alternate_optimizer=None,
alternate_lr_scheduler_type=None,
chat_template=None,
kd_ce_alpha=1.0,
kd_alpha=1.0,
kd_temperature=None,
kd_zscore_base_temp=None,
kd_top_k_before_softmax=1,
sequence_parallel_degree=None,
image_size=None,
image_resize_algorithm )
Reward config for Reward training
AxolotlTrainingArguments
core.training_args.AxolotlTrainingArguments(self,
=None,
model_type=False,
lr_quadratic_warmup=False,
pretraining=False,
sample_packing=False,
multipack_real_batches=None,
eval_sample_packing=1.0,
sample_packing_efficiency=200,
sample_packing_bin_size=100000,
sample_packing_group_size=2048,
max_seq_length=None,
relora_steps=None,
relora_warmup_steps=None,
relora_anneal_steps=0.9,
relora_prune_ratio='eval',
bench_split='pharaouk/dharma-1/dharma_1_mini.json',
bench_dataset=False,
do_bench_eval=False,
do_causal_lm_eval=None,
max_bench_samples=2048,
bench_source_max_len=None,
dataloader_prefetch_factor=None,
cosine_min_lr_ratio=None,
cosine_constant_lr_ratio=None,
loraplus_lr_ratio=1e-06,
loraplus_lr_embedding=None,
embedding_lr_scale=None,
lr_groups=None,
embedding_lr=False,
qlora=None,
orpo_alpha=None,
lisa_n_layers=None,
lisa_step_interval=None,
lisa_layers_attribute=None,
curriculum_sampling=None,
alternate_optimizer=None,
alternate_lr_scheduler_type=None,
chat_template=None,
kd_ce_alpha=1.0,
kd_alpha=1.0,
kd_temperature=None,
kd_zscore_base_temp=None,
kd_top_k_before_softmax=1,
sequence_parallel_degree=None,
image_size=None,
image_resize_algorithm )
Training arguments for Causal trainer
This code is duplicated due to HF TrainingArguments not setting output_dir with a default value so it can’t be used as a mixin.
AxolotlTrainingMixins
core.training_args.AxolotlTrainingMixins(self,
=None,
model_type=False,
lr_quadratic_warmup=False,
pretraining=False,
sample_packing=False,
multipack_real_batches=None,
eval_sample_packing=1.0,
sample_packing_efficiency=200,
sample_packing_bin_size=100000,
sample_packing_group_size=2048,
max_seq_length=None,
relora_steps=None,
relora_warmup_steps=None,
relora_anneal_steps=0.9,
relora_prune_ratio='eval',
bench_split='pharaouk/dharma-1/dharma_1_mini.json',
bench_dataset=False,
do_bench_eval=False,
do_causal_lm_eval=None,
max_bench_samples=2048,
bench_source_max_len=None,
dataloader_prefetch_factor=None,
cosine_min_lr_ratio=None,
cosine_constant_lr_ratio=None,
loraplus_lr_ratio=1e-06,
loraplus_lr_embedding=None,
embedding_lr_scale=None,
lr_groups=None,
embedding_lr=False,
qlora=None,
orpo_alpha=None,
lisa_n_layers=None,
lisa_step_interval=None,
lisa_layers_attribute=None,
curriculum_sampling=None,
alternate_optimizer=None,
alternate_lr_scheduler_type=None,
chat_template=None,
kd_ce_alpha=1.0,
kd_alpha=1.0,
kd_temperature=None,
kd_zscore_base_temp=None,
kd_top_k_before_softmax=1,
sequence_parallel_degree=None,
image_size=None,
image_resize_algorithm )
Mixin class for the Axolotl training args.