8000 Integrating local LLMs · Issue #51 · upstash/rag-chat · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content
Integrating local LLMs #51
Open
Open
@ogzhanolguncu

Description

@ogzhanolguncu

Currently, the RAG SDK only supports hosted models. If we could enable the use of local models, similar to web-llm, that would be great. The only issue is that while they are OpenAI-compatible, they don’t provide an endpoint to query from. I believe we could write a simple HTTP server using bun.js and trigger it only if the user decides to use one of the local LLMs.

For example:

  1. Start the model.
  2. Hook it up to a simple server.
  3. Generate a predefined URL and provide it to our base LLM client. And, add it like any other model.

The rest should be straightforward.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions

      0