Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

有没有详细一点的web_demo推理部署流程? #92

Open
LayKwokMing opened this issue Jan 15, 2025 · 1 comment
Open

有没有详细一点的web_demo推理部署流程? #92

LayKwokMing opened this issue Jan 15, 2025 · 1 comment

Comments

@LayKwokMing
Copy link

有没有详细一点的web_demo推理部署流程?按readme去做,接连报错。
1、先是报 Repo is must be in the form ‘repo_name’ or namespace/repo_name......,排查是readme中“mv demo_VITA_ckpt/config.json demo_VITA_ckpt/origin_config.json”这句话把config.json去掉了,把config.json补上,问题解决。
2、然后遇到 limit_mm_per_prompt is only supported for multimodal models. 把limit_mm_per_prompt={‘image’:256, 'audio':50}注释掉,问题解决。
3、又遇到vllm中['VITAQwen2ForCausalLM'] are not supported for now的问题,至此怀疑自己的操作步骤有问题。
有没有大佬可以解答一下

@lxysl
Copy link
Contributor

lxysl commented Jan 16, 2025

All the issues you encountered are due to not executing the following operations:

# Backup a new weight file
cp -rL  VITA_ckpt/ demo_VITA_ckpt/

mv demo_VITA_ckpt/config.json demo_VITA_ckpt/origin_config.json

cd ./web_demo/vllm_tools
cp -rf qwen2p5_model_weight_file/*  ../../demo_VITA_ckpt/
cp -rf vllm_file/*  your_anaconda/envs/vita_demo/lib/python3.10/site-packages/vllm/model_executor/models/

Our readme about Demo have the following structure:


Demo

readme0

📍 Basic Demo

readme1

📍 Real-Time Interactive Demo

readme2


Readme0 must be executed first before readme1 or readme2 is to be executed afterward.

The instructions here may not be clear, and we will revise them later.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants