Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

epic: llama.cpp params are settable via API call or model.yaml #1151

@dan-menlo

Description

@dan-menlo

Goal

  • Cortex can handle all llama.cpp params correctly
  • Model running params (i.e. POST /v1/models/<model_id>/start)
  • Inference params (i.e. POST /chat/completions)
  • Function Calling, eg for llama.cpp

Tasklist

I am using this epic to aggregate all llama.cpp params issues, including llama3.1 function calling + tool use

model.yaml

Out-of-scope:

Related

Metadata

Metadata

Labels

P0: criticalMission criticalcategory: model runningInference ux, handling context/parameters, runtimetype: epicA major feature or initiative

Type

No type

Projects

Status

Completed

Milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions