8000 Fix cpu inference with gpu build by zpitroda · Pull Request #323 · mindverse/Second-Me · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

Fix cpu inference with gpu build #323

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

8000 Already on GitHub? Sign in to your account

Open
wants to merge 3 commits into
base: develop
Choose a base branch
from

Conversation

zpitroda
Copy link
Contributor

When building and training with gpu and then trying to start service I get error:

local_llm_service.py:222 - Failed to start llama-server: free(): double free detected in tcache 2

I believe this is due to the env["CUDA_VISIBLE_DEVICES"] = "" line of local_llm_service causing memory conflicts since inference is currently only on cpu

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants
0