Skip to content

Commit 6e39f8d

Browse files
committed
More tweaks
1 parent 3bdcfcb commit 6e39f8d

File tree

1 file changed

+8
-8
lines changed

1 file changed

+8
-8
lines changed

en/config.json

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -99,14 +99,14 @@
9999
"llm.prediction.tools/serverPageDescriptionAddon": "Pass this through the request body as `tools` when using the server API",
100100
"llm.prediction.promptTemplate/title": "Prompt Template",
101101
"llm.prediction.promptTemplate/subTitle": "The format in which messages in chat are sent to the model. Changing this may introduce unexpected behavior - make sure you know what you're doing!",
102-
"llm.prediction.mlx.speculativeDecoding.numDraftTokensExact/title": "Draft Tokens to Generate",
103-
"llm.prediction.mlx.speculativeDecoding.numDraftTokensExact/subTitle": "The number of tokens to generate with the draft model per main model token. Find the sweet spot of compute vs. reward",
104-
"llm.prediction.llama.speculativeDecoding.minContinueDraftingProbability/title": "Drafting Probability Cutoff",
105-
"llm.prediction.llama.speculativeDecoding.minContinueDraftingProbability/subTitle": "Continue drafting until a token's probability falls below this threshold. Higher values generally mean lower risk, lower reward",
106-
"llm.prediction.llama.speculativeDecoding.minDraftLengthToConsider/title": "Min Draft Size",
107-
"llm.prediction.llama.speculativeDecoding.minDraftLengthToConsider/subTitle": "Drafts smaller than this will be ignored by the main model. Higher values generally mean lower risk, lower reward",
108-
"llm.prediction.llama.speculativeDecoding.maxTokensToDraft/title": "Max Draft Size",
109-
"llm.prediction.llama.speculativeDecoding.maxTokensToDraft/subTitle": "Max number of tokens allowed in a draft. Ceiling if all token probs are > the cutoff. Lower values generally mean lower risk, lower reward",
102+
"llm.prediction.speculativeDecoding.numDraftTokensExact/title": "Draft Tokens to Generate",
103+
"llm.prediction.speculativeDecoding.numDraftTokensExact/subTitle": "The number of tokens to generate with the draft model per main model token. Find the sweet spot of compute vs. reward",
104+
"llm.prediction.speculativeDecoding.minContinueDraftingProbability/title": "Drafting Probability Cutoff",
105+
"llm.prediction.speculativeDecoding.minContinueDraftingProbability/subTitle": "Continue drafting until a token's probability falls below this threshold. Higher values generally mean lower risk, lower reward",
106+
"llm.prediction.speculativeDecoding.minDraftLengthToConsider/title": "Min Draft Size",
107+
"llm.prediction.speculativeDecoding.minDraftLengthToConsider/subTitle": "Drafts smaller than this will be ignored by the main model. Higher values generally mean lower risk, lower reward",
108+
"llm.prediction.speculativeDecoding.maxTokensToDraft/title": "Max Draft Size",
109+
"llm.prediction.speculativeDecoding.maxTokensToDraft/subTitle": "Max number of tokens allowed in a draft. Ceiling if all token probs are > the cutoff. Lower values generally mean lower risk, lower reward",
110110
"llm.prediction.speculativeDecoding.draftModel/title": "Draft Model",
111111
"llm.prediction.reasoning.parsing/title": "Reasoning Section Parsing",
112112
"llm.prediction.reasoning.parsing/subTitle": "How to parse reasoning sections in the model's output",

0 commit comments

Comments
 (0)