Skip to content

Conversation

@yossiovadia
Copy link
Collaborator

Summary

Fixes llm-katan OpenAI API compatibility issues causing OpenWebUI to hang when connecting to llm-katan endpoints.

Changes Made

  • Add missing OpenAI API response fields: system_fingerprint, logprobs, detailed usage object
  • Fix streaming response Content-Type: Changed from text/plain to text/event-stream
  • Add token_usage alias: For better SDK compatibility
  • Apply fixes to both backends: TransformersBackend and VLLMBackend
  • Bump version to 0.1.9: Published to PyPI

Problem Solved

  • Resolves OpenWebUI hanging issue when using llm-katan as backend
  • Improves OpenAI SDK compatibility with proper response format
  • Fixes streaming responses with correct SSE format and final usage chunk

Testing

  • Tested with curl (non-streaming and streaming)
  • Verified all OpenAI compatibility fields present
  • Confirmed OpenWebUI now works correctly
  • Published v0.1.9 to PyPI

Fixes #241

…t#241

- Add missing OpenAI API response fields (system_fingerprint, logprobs, detailed usage)
- Fix streaming response Content-Type from text/plain to text/event-stream
- Ensure both static and streaming responses include all compatibility fields
- Add token_usage alias for better SDK compatibility
- Apply fixes to both TransformersBackend and VLLMBackend

Resolves OpenWebUI hanging issue when connecting to llm-katan endpoints.

Signed-off-by: Yossi Ovadia <[email protected]>
Published llm-katan v0.1.9 to PyPI with OpenAI API compatibility fixes.

Signed-off-by: Yossi Ovadia <[email protected]>
@yossiovadia yossiovadia requested a review from rootfs as a code owner October 6, 2025 20:34
@netlify
Copy link

netlify bot commented Oct 6, 2025

Deploy Preview for vllm-semantic-router ready!

Name Link
🔨 Latest commit 528bc6d
🔍 Latest deploy log https://app.netlify.com/projects/vllm-semantic-router/deploys/68e4954519cb16000809f343
😎 Deploy Preview https://deploy-preview-354--vllm-semantic-router.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

rootfs
rootfs previously approved these changes Oct 6, 2025
@rootfs
Copy link
Collaborator

rootfs commented Oct 6, 2025

@yossiovadia can you run pre-commit?

Trigger CI re-run to verify if Black formatting issues are resolved.

Signed-off-by: Yossi Ovadia <[email protected]>
@github-actions
Copy link

github-actions bot commented Oct 7, 2025

👥 vLLM Semantic Team Notification

The following members have been identified for the changed files in this PR and have been automatically assigned:

📁 e2e-tests

Owners: @yossiovadia
Files changed:

  • e2e-tests/llm-katan/llm_katan/model.py
  • e2e-tests/llm-katan/llm_katan/server.py
  • e2e-tests/llm-katan/pyproject.toml

vLLM

🎉 Thanks for your contributions!

This comment was automatically generated based on the OWNER files in the repository.

@Xunzhuo Xunzhuo merged commit 88eec24 into vllm-project:main Oct 7, 2025
9 checks passed
joyful-ii-V-I pushed a commit to joyful-ii-V-I/semantic-router that referenced this pull request Oct 13, 2025
…t#241 (vllm-project#354)

* fix: enhance llm-katan OpenAI API compatibility for issue vllm-project#241

- Add missing OpenAI API response fields (system_fingerprint, logprobs, detailed usage)
- Fix streaming response Content-Type from text/plain to text/event-stream
- Ensure both static and streaming responses include all compatibility fields
- Add token_usage alias for better SDK compatibility
- Apply fixes to both TransformersBackend and VLLMBackend

Resolves OpenWebUI hanging issue when connecting to llm-katan endpoints.

Signed-off-by: Yossi Ovadia <[email protected]>

* bump llm-katan version to 0.1.9 for PyPI release

Published llm-katan v0.1.9 to PyPI with OpenAI API compatibility fixes.

Signed-off-by: Yossi Ovadia <[email protected]>

* chore: trigger CI re-run to check pre-commit status

Trigger CI re-run to verify if Black formatting issues are resolved.

Signed-off-by: Yossi Ovadia <[email protected]>

* trigger pre-commit formatting fix

Signed-off-by: Yossi Ovadia <[email protected]>

* fix: apply black formatting to llm-katan Python files

Signed-off-by: Yossi Ovadia <[email protected]>

* fix: apply black formatting to llm-katan Python files for CI compliance

Signed-off-by: Yossi Ovadia <[email protected]>

---------

Signed-off-by: Yossi Ovadia <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Enhance llm-katan OpenAI API compatibility with missing response fields

3 participants