Skip to content

OpenAI endpoint using Templates for Chat Models #1160

@flexchar

Description

@flexchar

Feature request

Hey, it'd be super nice to have the same endpoint when working with variety of libraries just like with Amazon S3.

Motivation

Many existing libraries support OpenAI as one of the endpoints. It's kinda of a must. However the open source models are becoming just as good to be a drop in replacement whether for custom fine tune, data privacy or other motivation.

I recently discovered that many fine tuned models include tokenizer with chat template. I even saw that on my fine tunes when inspecting tokenizer config after training with axolotl.
https://huggingface.co/docs/transformers/main/chat_templating

These are two offerings from HuggingFace and if they were to have a baby that would mean an OpenAI compatible endpoint like vLLM has, it'd be lit. 🔥

Example from vLLM:
https://vllm.readthedocs.io/en/latest/getting_started/quickstart.html#openai-compatible-server

This would allow HuggingFace models to be a complete drop-in replacement for customer without needing to ever open a terminal, code editor or server and spin up an endpoint using existing GUI for Spaces.

Your contribution

Let's RFC first. I'm afraid I don't have the expertise to contribute to Rust at this point. I'm sorry.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions