Fine-tune LLAMA2 with DPO (Direct Preference Optimization) in AWS

0

I'm exploring fine-tuning with DPO and successfully trained facebook/opt-model (HF model) with DPO (Ref: https://huggingface.co/blog/dpo-trl). As part of DPO training, I first performed SFT training, and using the final checkpoint I performed DPO training.

Now, I'm working on fine-tuning Llama2 with DPO in AWS. I have successfully fine-tuned Llama2 in AWS SageMaker Jumpstart, but stuck there figuring out how to perform DPO using the fine-tuned model artifact which is stored in S3 bucket.

It would be helpful if anyone could share some resources or insights on how to proceed DPO training in AWS. Thanks in advance!

Jyothi
已提问 5 个月前1959 查看次数
没有答案

您未登录。 登录 发布回答。

一个好的回答可以清楚地解答问题和提供建设性反馈,并能促进提问者的职业发展。

回答问题的准则