-
Notifications
You must be signed in to change notification settings - Fork 3
Semantic Chunking of Content Files #2005
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
…gle semantic chunker
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM but the task was extremely slow until i set QDRANT_ENCODER=vector_search.encoders.litellm.LiteLLMEncoder . I'm not sure how vectorizing all the content files is going to work
what does this do ? |
This is the setting that allows us to toggle between fastembed for local environments and litellm when deployed. By setting this to "vector_search.encoders.litellm.LiteLLMEncoder" it routes all requests through litelllm |
litellm is a proxy. what encoder does it use ? |
The embedding model is a separate setting configured via |
What are the relevant tickets?
Closes https://github.com/mitodl/hq/issues/6602
Description (What does it do?)
This PR adds the ability to toggle the use of the Semantic chunker when embedding contentfiles.
How can this be tested?
Testing interactively
To quickly play with the different params and see how the chunker works interactively:
Testing via embedding command
docker compose down celery redis
docker compose up -d
python manage.y generate_embeddings --resource-ids 1,3,2