-
Notifications
You must be signed in to change notification settings - Fork 103
Description
Hey folks,
this package is absolutely awesome! I'm always watching out for performant small models, so this is a goldmine for me.
I have some questions/possible feature ideas for getting static models to support real life use cases.
-
CLI for embeddings: I'd love a simple CLI for embeddings, similar to what llama.cpp offers. The background is that small models can be quickly loaded and used to generate a query vector for an existing set of embeddings. My personal use case would be a minimal note taking app with advances search but low memory footprint. The major advantage here would be that one does not need to keep the model loaded (using VRAM) all the time.
-
An integration in transformers.js would be amazing! This way, downstream projects using embeddings like the one I'm working on SemanticFinder could be accelerated so much!
Maybe you could ping @xenova for this if interested. Alternatively, is there already a way to use a distllied static model in JS somehow? If so, could you document it somewhere? -
Could you also open Discussions in this repo?
Really excited to give these models a try, thanks for building this!