-
Notifications
You must be signed in to change notification settings - Fork 739
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Segmentation Fault during NexaVLMInference
Initialization on macOS Ventura
#208
Comments
it seems the issue related to memory handling, it happens when i give it a bit of large prompt |
Update i switched to using CPU and now models are initializing successfully but when i give a bit of a complex prompt i get the error:
|
Hi @Ysf101 in llava architecture, one image take 729+ tokens, and with a long prompt, it is easy to have OOM error on mac laptop. We have addressed this issue and plan to propose a new multiodal model architecture to support it. Seems that similar issue is also reported in llama.cpp |
Hi @zhiyuan8, Thank you for the information, Just to add to my initial bug report I've tried to run NexaAI with 5GB of RAM, then 12 GB of RAM and also tried to run it on my friend's laptop(we both run macOS in VMware in which we use NexaAI) and it produced the segmentation error on my friend laptop (same error i was having). The input does not need to be long like see in the below example:
Meanwhile a 4GB RAM Windows running NexaAI in VMware runs all models perfectly, so it's the same machine and everything except OS. |
@zhiyuan8 an update, I upgraded to sequoia macOS and the issue was solved completely for text models but still in vision models.
|
Issue Description
I'm encountering a segmentation fault when initializing
NexaVLMInference
using thenexa.gguf
package on macOS Ventura.Steps to Reproduce
nexaai
viapip install nexaai
.Environment Details:
macOS Ventura version: [e.g., 13.0]
Python version: 3.12.0
nexaaiversion: [0.0.9.0]
Architecture: x86_64
OS
macOS ventura
Python Version
3.12.0
Nexa SDK Version
0.0.9.0
GPU (if using one)
No response
The text was updated successfully, but these errors were encountered: