Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unable to produce the result between LLaMA-Adapter V1 and Alpaca #133

Open
vicissitude1999 opened this issue Nov 23, 2023 · 1 comment
Open

Comments

@vicissitude1999
Copy link

vicissitude1999 commented Nov 23, 2023

I used the provided trained weights of LLaMA-Adapter V1 and compared its performance with Alpaca. I wasn't able to get the same result as in figure 6 of the LLaMA-Adapter V1 paper. As shown in the image below, there are a lot of ties.

For Alpaca weights, I followed the official guide from https://huggingface.co/tatsu-lab/alpaca-7b-wdiff. Could you please detail the exact steps to reproduce figure 6?

微信图片_20231123113224
@csuhan
Copy link
Collaborator

csuhan commented Nov 30, 2023

The generation params have a great impact on the results. In our setting, LLaMA-Adapter uses top-p=0.1 and temprature=0.75 for generation.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants