File tree Expand file tree Collapse file tree 2 files changed +7
-0
lines changed Expand file tree Collapse file tree 2 files changed +7
-0
lines changed Original file line number Diff line number Diff line change @@ -72,6 +72,8 @@ class GRPOConfig(TrainingArguments):
72
72
Float that penalizes new tokens based on whether they appear in the prompt and the generated text so far.
73
73
Values > `1.0` encourage the model to use new tokens, while values < `1.0` encourage the model to repeat
74
74
tokens.
75
+ cache_implementation (`str` or `None`, *optional*, defaults to `None`):
76
+ Implementation of the cache method for faster generation when use_vllm is set to False.
75
77
76
78
> Parameters that control generation acceleration powered by vLLM
77
79
@@ -217,6 +219,10 @@ class GRPOConfig(TrainingArguments):
217
219
"to repeat tokens."
218
220
},
219
221
)
222
+ cache_implementation : Optional [str ] = field (
223
+ default = None ,
224
+ metadata = {"help" : "Implementation of the cache method for faster generation when use_vllm is set to False." },
225
+ )
220
226
221
227
# Parameters that control generation acceleration powered by vLLM
222
228
use_vllm : Optional [bool ] = field (
Original file line number Diff line number Diff line change @@ -548,6 +548,7 @@ def new_group_context():
548
548
top_k = args .top_k ,
549
549
min_p = args .min_p ,
550
550
repetition_penalty = args .repetition_penalty ,
551
+ cache_implementation = args .cache_implementation ,
551
552
)
552
553
553
554
# Gradient accumulation requires scaled loss. Normally, loss scaling in the parent class depends on whether the
You can’t perform that action at this time.
0 commit comments