You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This generated 3 pt files. I converted the last pt file to safetensor file and put it back to the original model and then ran things with vLLM. There was no crash. But the model outputspart of the inputs and then randomly listed S10, S111, etc. at the end. Apparently, the full fine-tuned model is not working.
In comparison, if I set use_peft=True, the LoRA fine-tuned model works fine in vLLM. Wonder if I miss anything in my setup. Appreciate your help!
Error logs
No error. As mentioned, just incorrect output.
Expected behavior
The model should output safe or unsafe with error code.
The text was updated successfully, but these errors were encountered:
System Info
Versions of relevant libraries:
[pip3] mypy-extensions==1.0.0
[pip3] numpy==1.26.4
[pip3] onnx==1.17.0
[pip3] onnxruntime==1.20.1
[pip3] optree==0.13.1
[pip3] pytorch-triton==3.0.0+72734f086
[pip3] torch==2.6.0a0+df5bbc09d1.nv24.11
[pip3] torch_tensorrt==2.6.0a0
[pip3] torchao==0.7.0
[pip3] torchaudio==2.5.0a0+265bc5c
[pip3] torchaudio==2.5.0a0+265bc5c
[pip3] torchprofile==0.0.4
[pip3] torchtune==0.0.0
[pip3] torchvision==0.20.0a0
[pip3] triton==3.1.0
Information
🐛 Describe the bug
I have the same set-up as in the previous issue #831. I have
This generated 3
pt
files. I converted the lastpt
file to safetensor file and put it back to the original model and then ran things with vLLM. There was no crash. But the model outputspart of the inputs and then randomly listed S10, S111, etc. at the end. Apparently, the full fine-tuned model is not working.In comparison, if I set
use_peft=True
, the LoRA fine-tuned model works fine in vLLM. Wonder if I miss anything in my setup. Appreciate your help!Error logs
No error. As mentioned, just incorrect output.
Expected behavior
The model should output
safe
orunsafe
with error code.The text was updated successfully, but these errors were encountered: