@@ -1020,6 +1020,8 @@ async def oai_chat_completions(user_prompt,
1020
1020
event_data = line .split (': ' , 1 )
1021
1021
assert event_data [0 ] == 'data' , f'Bad event code received: ```{ event_data } ```'
1022
1022
chunk_raw = event_data [1 ]
1023
+ if chunk_raw == '[DONE]' :
1024
+ continue
1023
1025
1024
1026
chunk = json .loads (chunk_raw )
1025
1027
assert len (chunk ['choices' ]) == 1 , f"no choices provided, line ```{ line } ```"
@@ -1322,14 +1324,16 @@ def start_server_background(context):
1322
1324
if 'LLAMA_SERVER_BIN_PATH' in os .environ :
1323
1325
context .server_path = os .environ ['LLAMA_SERVER_BIN_PATH' ]
1324
1326
server_listen_addr = context .server_fqdn
1327
+ # server_listen_addr = "127.0.0.1"
1325
1328
server_args = [
1326
1329
'--host' , server_listen_addr ,
1327
1330
'--port' , context .server_port ,
1328
1331
]
1332
+
1329
1333
if context .model_file :
1330
1334
server_args .extend (['--model' , context .model_file ])
1331
- if context .model_url :
1332
- server_args .extend (['--model-url' , context .model_url ])
1335
+ # if context.model_url:
1336
+ # server_args.extend(['--model-url', context.model_url])
1333
1337
if context .model_hf_repo :
1334
1338
server_args .extend (['--hf-repo' , context .model_hf_repo ])
1335
1339
if context .model_hf_file :
0 commit comments