Skip to content
This repository was archived by the owner on May 1, 2025. It is now read-only.

Commit 22ae1da

Browse files
author
Nghi Bui
committed
update configs
1 parent f7f0017 commit 22ae1da

File tree

4 files changed

+7
-11
lines changed

4 files changed

+7
-11
lines changed

codetf/configs/inference/causal_lm.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,8 @@ causallm-codegen-350M-multi-pretrained:
77
tokenizer_url: "Salesforce/codegen-350M-multi"
88
max_prediction_length: 512
99
causallm-codegen-350M-nl-pretrained:
10-
huggingface_url: "Salesforce/codegen-350-nl"
11-
tokenizer_url: "Salesforce/codegen-350-nl"
10+
huggingface_url: "Salesforce/codegen-350M-nl"
11+
tokenizer_url: "Salesforce/codegen-350M-nl"
1212
max_prediction_length: 512
1313
causallm-codegen-2B-mono-pretrained:
1414
huggingface_url: "Salesforce/codegen-2B-mono"

codetf/configs/inference/codet5.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -140,20 +140,20 @@ codet5-plus-770M-python-pretrained:
140140
max_prediction_length: 512
141141
beam_size: 5
142142
trust_remote_code: False
143-
device_map: True
143+
device_map: False
144144
codet5-plus-770M-pretrained:
145145
huggingface_url: "Salesforce/codet5p-770m"
146146
tokenizer_url: "Salesforce/codet5p-770m"
147147
max_source_length: 512
148148
max_prediction_length: 512
149149
beam_size: 5
150150
trust_remote_code: False
151-
device_map: True
151+
device_map: False
152152
codet5-plus-220M-pretrained:
153153
huggingface_url: "Salesforce/codet5p-220m"
154154
tokenizer_url: "Salesforce/codet5p-220m"
155155
max_source_length: 512
156156
max_prediction_length: 512
157157
beam_size: 5
158158
trust_remote_code: False
159-
device_map: True
159+
device_map: False

codetf/configs/training/causal_lm.yaml

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,9 +6,6 @@ hyperparameters:
66
num_train_epochs: 10
77
auto_find_batch_size: True
88
batch_size: 4
9-
max_steps: 1000
10-
eval_steps: 100
11-
save_steps: 1000
129
logging_steps: 100
1310
per_device_train_batch_size: 8
1411
per_device_eval_batch_size: 8
@@ -30,3 +27,4 @@ hyperparameters:
3027
beam_size: 5
3128
max_grad_norm: 5.0
3229
adam_epsilon : 1e-06
30+
load_best_model_at_end: True

codetf/configs/training/codet5.yaml

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,9 +6,6 @@ hyperparameters:
66
num_train_epochs: 1
77
auto_find_batch_size: True
88
batch_size: 4
9-
max_steps: 1000
10-
eval_steps: 100
11-
save_steps: 1000
129
logging_steps: 100
1310
per_device_train_batch_size: 8
1411
per_device_eval_batch_size: 8
@@ -30,6 +27,7 @@ hyperparameters:
3027
beam_size: 5
3128
max_grad_norm: 5.0
3229
adam_epsilon : 1e-06
30+
load_best_model_at_end: True
3331
lora:
3432
r: 8
3533
lora_alpha: 32

0 commit comments

Comments
 (0)