Skip to content

Commit 8689714

Browse files
committed
remove time.sleepp; add logging
1 parent adcbef5 commit 8689714

File tree

2 files changed

+8
-9
lines changed

2 files changed

+8
-9
lines changed

src/embeddedllm/engine.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ def __init__(self, model_path: str, vision: bool):
4848
sliding_window_len=self.get_hf_config_sliding_window(),
4949
)
5050

51-
logger.info(self.max_model_len)
51+
logger.info("Model Context Lenght: " + str(self.max_model_len))
5252

5353
try:
5454
logger.info("Attempt to load fast tokenizer")
@@ -122,7 +122,7 @@ async def generate_vision(
122122
tmpfile.flush()
123123

124124
# logger.trace("Loading from temporary file: {name}", name=image_path.as_posix())
125-
logger.debug("Loading from temporary file: {name}", name=image_path.as_posix())
125+
# logger.debug("Loading from temporary file: {name}", name=image_path.as_posix())
126126

127127
# if not os.path.exists(image_path.as_posix()):
128128
# raise FileNotFoundError(f"Image file not found: {image_path.as_posix()}")
@@ -312,7 +312,7 @@ async def generate(
312312
prompt_text = inputs["prompt"]
313313
input_token_length = None
314314
input_tokens = None # for text only use case
315-
logger.debug("inputs: " + prompt_text)
315+
# logger.debug("inputs: " + prompt_text)
316316

317317
input_tokens = self.onnx_tokenizer.encode(prompt_text)
318318
input_token_length = len(input_tokens)

src/embeddedllm/entrypoints/chat_server.py

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ def _load_chat_template(self, chat_template: Optional[str]):
109109
elif tokenizer.chat_template is not None:
110110
logger.info("Using default chat template:\n%s", tokenizer.chat_template)
111111
else:
112-
logger.warning("No chat template provided. Chat API will not work.")
112+
logger.warning("No chat template provided. Chat API will not work as intended.")
113113

114114
def create_error_response(
115115
self,
@@ -130,7 +130,7 @@ def _parse_chat_message_content_parts(
130130
for part in parts:
131131
# logger.debug(f"part: {str(part)}")
132132
part_type = part["type"]
133-
logger.debug(f"part_type: {part_type}")
133+
# logger.debug(f"part_type: {part_type}")
134134
if part_type == "text":
135135
text = cast(ChatCompletionContentPartTextParam, part)["text"]
136136

@@ -153,8 +153,8 @@ def _parse_chat_message_content_parts(
153153

154154
file_data, mime_type = decode_base64(image_url["url"])
155155

156-
logger.debug(f"file_data: {type(file_data)}")
157-
logger.debug(f"mime_type: {str(mime_type)}")
156+
# logger.debug(f"file_data: {type(file_data)}")
157+
# logger.debug(f"mime_type: {str(mime_type)}")
158158

159159
image_future: ImagePixelData = {
160160
"image_pixel_data": file_data,
@@ -174,7 +174,7 @@ def _parse_chat_message_content_parts(
174174

175175
messages = [ConversationMessage(role=role, content=text_prompt)]
176176

177-
logger.debug(f"messages: {str(messages)}")
177+
# logger.debug(f"messages: {str(messages)}")
178178
return ChatMessageParseResult(messages=messages, image_futures=image_futures)
179179

180180
def _parse_chat_message_content(
@@ -343,7 +343,6 @@ async def chat_completion_stream_generator(
343343
first_iteration = False
344344

345345
for output in res.outputs:
346-
time.sleep(0.5)
347346
i = output.index
348347

349348
if finish_reason_sent[i]:

0 commit comments

Comments
 (0)