File tree Expand file tree Collapse file tree 1 file changed +1
-6
lines changed
src/embeddedllm/entrypoints Expand file tree Collapse file tree 1 file changed +1
-6
lines changed Original file line number Diff line number Diff line change @@ -81,16 +81,11 @@ async def create_chat_completion(request: ChatCompletionRequest,
8181
8282 generator = await openai_chat_server .create_chat_completion (
8383 request , raw_request )
84- # print(generator)
8584 if isinstance (generator , ErrorResponse ):
8685 return JSONResponse (content = generator .model_dump (),
8786 status_code = generator .code )
8887 if request .stream :
89- async def event_generator ():
90- async for resp in generator :
91- yield resp
92- print ("Return StreamingResponse" )
93- return StreamingResponse (content = event_generator (),
88+ return StreamingResponse (content = generator ,
9489 media_type = "text/event-stream" )
9590 else :
9691 # return JSONResponse(content="Non-streaming Chat Generation Yet to be Implemented.",
You can’t perform that action at this time.
0 commit comments