mirror of
https://github.com/hiyouga/LlamaFactory.git
synced 2026-03-23 02:33:24 +08:00
fix generating args
Former-commit-id: 52805a8441bd7b324bd89489de60f18f103c8e4c
This commit is contained in:
@@ -30,8 +30,8 @@ def main():
|
||||
# Override the decoding parameters of Seq2SeqTrainer
|
||||
training_args.generation_max_length = training_args.generation_max_length if \
|
||||
training_args.generation_max_length is not None else data_args.max_target_length
|
||||
training_args.generation_num_beams = data_args.num_beams if \
|
||||
data_args.num_beams is not None else training_args.generation_num_beams
|
||||
training_args.generation_num_beams = data_args.eval_num_beams if \
|
||||
data_args.eval_num_beams is not None else training_args.generation_num_beams
|
||||
|
||||
# Split the dataset
|
||||
if training_args.do_train:
|
||||
|
||||
Reference in New Issue
Block a user