-
Notifications
You must be signed in to change notification settings - Fork 40
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Loading checkpoint of saved intervened model takes long time #45
Comments
@jeffreyzhanghc Hi, I worked on updating this codebase and can help you. May I ask what GPU setup you are using to intervene on llama3.1 8b? I used 1 H100 and loaded the model in 1 minute, so perhaps it's a GPU size issue. |
Hi, thanks for helping, yes i m using 2 A100 80GB, and it can takes up to 30mins or even more, does the case of long prompt also affect the intervention time |
Hmm, that's odd. Are you using using both devices, e.g. when running |
yes i omit the cuda=0 |
Hi Jeffrey, I just wanted to confirm the workflow you are doing:
To clarify: the warning you see of Please let me know if my understanding of your issue is correct, and I'll see how else I can help. |
Hi, thanks for the great work, I am trying to replicate the outcome on different dataset, but when I intervene llama3.1 8B

it has following warning, and the loading time is about 20min, I want to ask is that long checkpoint loading time normal?
The text was updated successfully, but these errors were encountered: