-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Description
Feature request
Hey, it'd be super nice to have the same endpoint when working with variety of libraries just like with Amazon S3.
Motivation
Many existing libraries support OpenAI as one of the endpoints. It's kinda of a must. However the open source models are becoming just as good to be a drop in replacement whether for custom fine tune, data privacy or other motivation.
I recently discovered that many fine tuned models include tokenizer with chat template. I even saw that on my fine tunes when inspecting tokenizer config after training with axolotl
.
https://huggingface.co/docs/transformers/main/chat_templating
These are two offerings from HuggingFace and if they were to have a baby that would mean an OpenAI compatible endpoint like vLLM has, it'd be lit. 🔥
Example from vLLM:
https://vllm.readthedocs.io/en/latest/getting_started/quickstart.html#openai-compatible-server
This would allow HuggingFace models to be a complete drop-in replacement for customer without needing to ever open a terminal, code editor or server and spin up an endpoint using existing GUI for Spaces.
Your contribution
Let's RFC first. I'm afraid I don't have the expertise to contribute to Rust at this point. I'm sorry.