@@ -42,22 +42,22 @@ def vllm_infer(
4242 raise ValueError ("Pipeline parallel size should be smaller than the number of gpus." )
4343
4444 model_args , data_args , _ , generating_args = get_infer_args (
45- dict (
46- model_name_or_path = model_name_or_path ,
47- adapter_name_or_path = adapter_name_or_path ,
48- dataset = dataset ,
49- dataset_dir = dataset_dir ,
50- template = template ,
51- cutoff_len = cutoff_len ,
52- max_samples = max_samples ,
53- preprocessing_num_workers = 16 ,
54- vllm_config = vllm_config ,
55- temperature = temperature ,
56- top_p = top_p ,
57- top_k = top_k ,
58- max_new_tokens = max_new_tokens ,
59- repetition_penalty = repetition_penalty ,
60- )
45+ {
46+ " model_name_or_path" : model_name_or_path ,
47+ " adapter_name_or_path" : adapter_name_or_path ,
48+ " dataset" : dataset ,
49+ " dataset_dir" : dataset_dir ,
50+ " template" : template ,
51+ " cutoff_len" : cutoff_len ,
52+ " max_samples" : max_samples ,
53+ " preprocessing_num_workers" : 16 ,
54+ " vllm_config" : vllm_config ,
55+ " temperature" : temperature ,
56+ " top_p" : top_p ,
57+ " top_k" : top_k ,
58+ " max_new_tokens" : max_new_tokens ,
59+ " repetition_penalty" : repetition_penalty ,
60+ }
6161 )
6262
6363 tokenizer_module = load_tokenizer (model_args )
0 commit comments