Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llava_hf inference is extremely slow #387

Open
luomancs opened this issue Nov 1, 2024 · 1 comment
Open

llava_hf inference is extremely slow #387

luomancs opened this issue Nov 1, 2024 · 1 comment

Comments

@luomancs
Copy link

luomancs commented Nov 1, 2024

Hi there,

Thank you for the benchmark.
I have tried to this repo and get the inference of llava_hf model type and the checkpoint of llava-hf/llava-v1.6-mistral-7b-hf on infovqa, however, the inference speed is extremely slow and also it get out of memory issue even when I use A100 gpu with 80G memory. The following is my command line:

python -m accelerate.commands.launch
--num_processes=8
-m lmms_eval
--model llava_hf
--model_args pretrained="llava-hf/llava-v1.6-mistral-7b-hf"
--tasks infovqa
--batch_size 1
--log_samples
--log_samples_suffix llava_v1.6_mistral_infovqa_scienceqa_docvqa
--output_path ./logs/

On the other hand, running the inference with llava model type with checkpoint of liuhaotian/llava-v1.6-mistral-7b on the same dataset only requires 45 mins to do the inference without any OOM issue.

Thanks for your guidance.

@kcz358
Copy link
Collaborator

kcz358 commented Nov 1, 2024

There is a fix recently in #386 to fix some issues for llava_hf that possibily relate to this issue. Can you try that branch?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants