Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Integrating local LLMs #51

Open
1 of 2 tasks
ogzhanolguncu opened this issue Aug 13, 2024 · 0 comments
Open
1 of 2 tasks

Integrating local LLMs #51

ogzhanolguncu opened this issue Aug 13, 2024 · 0 comments

Comments

@ogzhanolguncu
Copy link
Contributor

ogzhanolguncu commented Aug 13, 2024

Currently, the RAG SDK only supports hosted models. If we could enable the use of local models, similar to web-llm, that would be great. The only issue is that while they are OpenAI-compatible, they don’t provide an endpoint to query from. I believe we could write a simple HTTP server using bun.js and trigger it only if the user decides to use one of the local LLMs.

For example:

  1. Start the model.
  2. Hook it up to a simple server.
  3. Generate a predefined URL and provide it to our base LLM client. And, add it like any other model.

The rest should be straightforward.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant