Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] <title>Failed to load clip model #313

Open
Desperat1on opened this issue Dec 11, 2024 · 2 comments
Open

[BUG] <title>Failed to load clip model #313

Desperat1on opened this issue Dec 11, 2024 · 2 comments
Labels
🐞 bug Something isn't working

Comments

@Desperat1on
Copy link

Desperat1on commented Dec 11, 2024

Issue Description

On Windows, I cannot use multimodal models like llava-v1.6-vicuna-7b:q4_0 or llava-llama-3-8b-v1.1, as it throws an error: "Error running ggml inference: Failed to load clip model: C:\Users\xxx.cache\nexa\hub\official\llava-v1.6-vicuna-7b\projector-q4_0.gguf. Please refer to our docs to install the nexaai package: https://docs.nexaai.com/getting-started/installation." However, I am able to use the omniVLM:fp16 model, and I am not sure where the issue is occurring.

Steps to Reproduce

1.$env:CMAKE_ARGS="-DGGML_CUDA=ON -DSD_CUBLAS=ON"; pip install nexaai --prefer-binary --index-url https://github.nexa.ai/whl/cu124 --extra-index-url https://pypi.org/simple --no-cache-dir
2.nexa run llava-v1.6-vicuna-7b:q4_0

OS

Windows 11

Python Version

3.12.7

Nexa SDK Version

0.0.9.6

GPU (if using one)

NVIDIA RTX 4090 Laptop,CUDA 12.3

@Desperat1on Desperat1on added the 🐞 bug Something isn't working label Dec 11, 2024
@Davidqian123
Copy link
Collaborator

What about trying to run another llava model like nexa run nanollava and check if it has the same error.

@Desperat1on
Copy link
Author

What about trying to run another llava model like nexa run nanollava and check if it has the same error.

Sorry, the situation is the same. Only the OmniVLM multimodal model is available.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
🐞 bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants