We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b426e7e commit 4e10efbCopy full SHA for 4e10efb
vec_inf/cli/_cli.py
@@ -63,6 +63,11 @@ def cli():
63
type=str,
64
help='Time limit for job, this should comply with QoS, default to max walltime of the chosen QoS'
65
)
66
+@click.option(
67
+ "--vocab-size",
68
+ type=int,
69
+ help='Vocabulary size, this option is intended for custom models'
70
+)
71
@click.option(
72
"--data-type",
73
@@ -93,6 +98,7 @@ def launch(
93
98
num_gpus: int=None,
94
99
qos: str=None,
95
100
time: str=None,
101
+ vocab_size: int=None,
96
102
data_type: str=None,
97
103
venv: str=None,
104
log_dir: str=None,
@@ -109,16 +115,20 @@ def launch(
109
115
110
116
models_df = load_models_df()
111
117
112
- if model_name not in models_df['model_name'].values:
113
- raise ValueError(f"Model name {model_name} not found in available models")
114
-
- default_args = load_default_args(models_df, model_name)
- for arg in default_args:
118
- if arg in locals() and locals()[arg] is not None:
119
- default_args[arg] = locals()[arg]
120
- renamed_arg = arg.replace("_", "-")
121
- launch_cmd += f" --{renamed_arg} {default_args[arg]}"
+ if model_name in models_df['model_name'].values:
+ default_args = load_default_args(models_df, model_name)
+ for arg in default_args:
+ if arg in locals() and locals()[arg] is not None:
122
+ default_args[arg] = locals()[arg]
123
+ renamed_arg = arg.replace("_", "-")
124
+ launch_cmd += f" --{renamed_arg} {default_args[arg]}"
125
+ else:
126
+ model_args = models_df.columns.tolist()
127
+ excluded_keys = ['model_name', 'pipeline_parallelism']
128
+ for arg in model_args:
129
+ if arg not in excluded_keys and locals()[arg] is not None:
130
131
+ launch_cmd += f" --{renamed_arg} {locals()[arg]}"
132
133
output = run_bash_command(launch_cmd)
134
vec_inf/models/models.csv
@@ -42,4 +42,5 @@ Mixtral-8x7B-Instruct-v0.1,Mixtral,8x7B-Instruct-v0.1,a40,m2,08:00:00,4,1,32000,
42
Mixtral-8x22B-v0.1,Mixtral,8x22B-v0.1,a40,m2,08:00:00,4,2,32768,65536,auto,singularity,default,false
43
Mixtral-8x22B-Instruct-v0.1,Mixtral,8x22B-Instruct-v0.1,a40,m2,08:00:00,4,2,32768,65536,auto,singularity,default,false
44
Phi-3-medium-128k-instruct,Phi-3,medium-128k-instruct,a40,m2,08:00:00,2,1,32064,131072,auto,singularity,default,false
45
-Phi-3-vision-128k-instruct,Phi-3,vision-128k-instruct,a40,m2,08:00:00,2,1,32064,65536,auto,singularity,default,false
+Phi-3-vision-128k-instruct,Phi-3,vision-128k-instruct,a40,m2,08:00:00,2,1,32064,65536,auto,singularity,default,false
46
+Llama3-OpenBioLLM-70B,Llama3-OpenBioLLM,70B,a40,m2,08:00:00,4,1,128256,8192,auto,singularity,default,false
vec_inf/multinode_vllm.slurm
@@ -93,7 +93,6 @@ if [ "$VENV_BASE" = "singularity" ]; then
--pipeline-parallel-size ${PIPELINE_PARALLEL_SIZE} \
--tensor-parallel-size ${TENSOR_PARALLEL_SIZE} \
--dtype ${VLLM_DATA_TYPE} \
- --load-format safetensors \
--trust-remote-code \
--max-logprobs ${VLLM_MAX_LOGPROBS} \
--max-model-len ${VLLM_MAX_MODEL_LEN}
@@ -107,7 +106,6 @@ else
107
106
108
0 commit comments