We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
want to train dpo
#!/bin/bash
CUDA_VISIBLE_DEVICES=0 python ../../src/train_bash.py --stage dpo --do_train --model_name_or_path BASE_MODEL --adapter_name_or_path FINETUNED_MODEL --create_new_adapter --dataset orca_rlhf --dataset_dir ../../data --template default --finetuning_type lora --lora_target q_proj,v_proj --output_dir ../../saves/LLaMA2-7B/lora/dpo --overwrite_cache --overwrite_output_dir --cutoff_len 1024 --preprocessing_num_workers 16 --per_device_train_batch_size 1 --per_device_eval_batch_size 1 --gradient_accumulation_steps 8 --lr_scheduler_type cosine --logging_steps 10 --warmup_steps 20 --save_steps 100 --eval_steps 100 --evaluation_strategy steps --load_best_model_at_end --learning_rate 1e-5 --num_train_epochs 1.0 --max_samples 1000 --val_size 0.1 --dpo_ftx 1.0 --plot_loss --fp16
error. cannot run dpo. Should I just put FINETUNED_MODEL at "--model_name_or_path" and delete "--adapter_name_or_path"?
ValueError: Can't find 'adapter_config.json' at '/path/to/FINETUNED_MODEL'
No response
The text was updated successfully, but these errors were encountered:
if you have merged lora adapters to the base model, just use --model_name_or_path only, otherwise, put your sft lora to --adapter_name_or_path
Sorry, something went wrong.
No branches or pull requests
Reminder
Reproduction
want to train dpo
#!/bin/bash
CUDA_VISIBLE_DEVICES=0 python ../../src/train_bash.py
--stage dpo
--do_train
--model_name_or_path BASE_MODEL
--adapter_name_or_path FINETUNED_MODEL
--create_new_adapter
--dataset orca_rlhf
--dataset_dir ../../data
--template default
--finetuning_type lora
--lora_target q_proj,v_proj
--output_dir ../../saves/LLaMA2-7B/lora/dpo
--overwrite_cache
--overwrite_output_dir
--cutoff_len 1024
--preprocessing_num_workers 16
--per_device_train_batch_size 1
--per_device_eval_batch_size 1
--gradient_accumulation_steps 8
--lr_scheduler_type cosine
--logging_steps 10
--warmup_steps 20
--save_steps 100
--eval_steps 100
--evaluation_strategy steps
--load_best_model_at_end
--learning_rate 1e-5
--num_train_epochs 1.0
--max_samples 1000
--val_size 0.1
--dpo_ftx 1.0
--plot_loss
--fp16
Expected behavior
error. cannot run dpo.
Should I just put FINETUNED_MODEL at "--model_name_or_path" and delete "--adapter_name_or_path"?
System Info
ValueError: Can't find 'adapter_config.json' at '/path/to/FINETUNED_MODEL'
Others
No response
The text was updated successfully, but these errors were encountered: