File tree Expand file tree Collapse file tree 2 files changed +10
-6
lines changed Expand file tree Collapse file tree 2 files changed +10
-6
lines changed Original file line number Diff line number Diff line change @@ -157,15 +157,15 @@ def llama_context_default_params() -> llama_context_params:
157157_lib .llama_context_default_params .restype = llama_context_params
158158
159159
160- def llama_mmap_supported () -> c_bool :
160+ def llama_mmap_supported () -> bool :
161161 return _lib .llama_mmap_supported ()
162162
163163
164164_lib .llama_mmap_supported .argtypes = []
165165_lib .llama_mmap_supported .restype = c_bool
166166
167167
168- def llama_mlock_supported () -> c_bool :
168+ def llama_mlock_supported () -> bool :
169169 return _lib .llama_mlock_supported ()
170170
171171
@@ -387,7 +387,9 @@ def llama_n_embd(ctx: llama_context_p) -> c_int:
387387# Can be mutated in order to change the probabilities of the next token
388388# Rows: n_tokens
389389# Cols: n_vocab
390- def llama_get_logits (ctx : llama_context_p ): # type: (...) -> Array[float] # type: ignore
390+ def llama_get_logits (
391+ ctx : llama_context_p ,
392+ ): # type: (...) -> Array[float] # type: ignore
391393 return _lib .llama_get_logits (ctx )
392394
393395
@@ -397,7 +399,9 @@ def llama_get_logits(ctx: llama_context_p): # type: (...) -> Array[float] # typ
397399
398400# Get the embeddings for the input
399401# shape: [n_embd] (1-dimensional)
400- def llama_get_embeddings (ctx : llama_context_p ): # type: (...) -> Array[float] # type: ignore
402+ def llama_get_embeddings (
403+ ctx : llama_context_p ,
404+ ): # type: (...) -> Array[float] # type: ignore
401405 return _lib .llama_get_embeddings (ctx )
402406
403407
Original file line number Diff line number Diff line change @@ -27,11 +27,11 @@ class Settings(BaseSettings):
2727 )
2828 f16_kv : bool = Field (default = True , description = "Whether to use f16 key/value." )
2929 use_mlock : bool = Field (
30- default = bool ( llama_cpp .llama_mlock_supported (). value ),
30+ default = llama_cpp .llama_mlock_supported (),
3131 description = "Use mlock." ,
3232 )
3333 use_mmap : bool = Field (
34- default = bool ( llama_cpp .llama_mmap_supported (). value ),
34+ default = llama_cpp .llama_mmap_supported (),
3535 description = "Use mmap." ,
3636 )
3737 embedding : bool = Field (default = True , description = "Whether to use embeddings." )
You can’t perform that action at this time.
0 commit comments