Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for vLLM embedding models #29

Merged
merged 1 commit into from
Jun 14, 2024
Merged

Conversation

petermuller
Copy link
Contributor

This change adds explicit support for vLLM embedding models (all one of them so far) for use within LISA. Either LiteLLM or LangChain was defaulting the embeddings API to using base64 as the encoding format, which the intfloat model threw errors on. By changing the encoding format to float, we preserve the functionality of the existing RAG implementation, and prevent the intfloat model from failing when it's called with default parameters.

Tested by deploying to my account and using the intfloat model for document upload and search within the chat UI.

By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.

@petermuller petermuller requested a review from estohlmann June 13, 2024 23:23
@petermuller petermuller self-assigned this Jun 13, 2024
@petermuller petermuller merged commit 52fb8ae into main Jun 14, 2024
2 checks passed
@petermuller petermuller deleted the feature/vllm-embeddings branch June 14, 2024 18:36
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants