@@ -382,7 +382,7 @@ def test_format_chunk_unknown_type(model):
382382
383383
384384@pytest .mark .asyncio
385- async def test_stream (openai_client , model , agenerator , alist ):
385+ async def test_stream (openai_client , model_id , model , agenerator , alist ):
386386 mock_tool_call_1_part_1 = unittest .mock .Mock (index = 0 )
387387 mock_tool_call_2_part_1 = unittest .mock .Mock (index = 1 )
388388 mock_delta_1 = unittest .mock .Mock (
@@ -465,7 +465,7 @@ async def test_stream(openai_client, model, agenerator, alist):
465465 # Verify that format_request was called with the correct arguments
466466 expected_request = {
467467 "max_tokens" : 1 ,
468- "model" : "m1" ,
468+ "model" : model_id ,
469469 "messages" : [{"role" : "user" , "content" : [{"text" : "calculate 2+2" , "type" : "text" }]}],
470470 "stream" : True ,
471471 "stream_options" : {"include_usage" : True },
@@ -475,14 +475,13 @@ async def test_stream(openai_client, model, agenerator, alist):
475475
476476
477477@pytest .mark .asyncio
478- async def test_stream_empty (openai_client , model , agenerator , alist ):
478+ async def test_stream_empty (openai_client , model_id , model , agenerator , alist ):
479479 mock_delta = unittest .mock .Mock (content = None , tool_calls = None , reasoning_content = None )
480- mock_usage = unittest .mock .Mock (prompt_tokens = 0 , completion_tokens = 0 , total_tokens = 0 )
481480
482481 mock_event_1 = unittest .mock .Mock (choices = [unittest .mock .Mock (finish_reason = None , delta = mock_delta )])
483482 mock_event_2 = unittest .mock .Mock (choices = [unittest .mock .Mock (finish_reason = "stop" , delta = mock_delta )])
484483 mock_event_3 = unittest .mock .Mock ()
485- mock_event_4 = unittest .mock .Mock (usage = mock_usage )
484+ mock_event_4 = unittest .mock .Mock (usage = None )
486485
487486 openai_client .chat .completions .create = unittest .mock .AsyncMock (
488487 return_value = agenerator ([mock_event_1 , mock_event_2 , mock_event_3 , mock_event_4 ]),
@@ -497,13 +496,12 @@ async def test_stream_empty(openai_client, model, agenerator, alist):
497496 {"contentBlockStart" : {"start" : {}}},
498497 {"contentBlockStop" : {}},
499498 {"messageStop" : {"stopReason" : "end_turn" }},
500- {"metadata" : {"usage" : {"inputTokens" : 0 , "outputTokens" : 0 , "totalTokens" : 0 }, "metrics" : {"latencyMs" : 0 }}},
501499 ]
502500
503501 assert len (tru_events ) == len (exp_events )
504502 expected_request = {
505503 "max_tokens" : 1 ,
506- "model" : "m1" ,
504+ "model" : model_id ,
507505 "messages" : [],
508506 "stream" : True ,
509507 "stream_options" : {"include_usage" : True },
0 commit comments