-
Notifications
You must be signed in to change notification settings - Fork 19
Config Parameters
Yiwei Mao edited this page Oct 14, 2025
·
6 revisions
| parameter | type | default | description |
|---|---|---|---|
| projects | string[] | [] | If empty, all projects will pass. Value is like "@web-bench/calculator". |
| agentMode | "local" | "http" | "local" | |
| agentEndPoint | string | "" | When agentMode is set to "http", set http API for network requests. |
| models | string[] | [] | 'models' field in apps/eval/src/model.json
|
| maxdop | number | 30 | max degree of parallelism |
| logLevel | "info" | "warn" | "debug" | "error" | "info" | |
| httpLimit | number | 10 | When agentMode is set to "http", maximum concurrent requests |
| fileDiffLog | boolean | false | Whether to log the diff of files generated by llm. Only enable in 'debug' log level. Note: This affects performance, don't enable it during all-project evaluation. |
| screenshotLog | boolean | false | Whether to log the screenshot. Only enable in 'debug' log level. Note: This affects performance, don't enable it during all-project evaluation. |
| startTask | string | the first task of tasks.jsonl
|
Task executed starts from, including startTask. |
| endTask | string | last task of tasks.jsonl
|
Task executed ends to, including endTask. |
- 'local': This mode has the basic capability to interact with LLM. It can specify the corresponding model in the
apps/eval/src/model.json. - 'http': Through this mode, it calls the configured
agentEndPointto send a request to the custom Agent. You can read Agent Server for more details.
- For models deployed on OpenRouter, use the native OpenRouter provider with the following configuration:
{
"title": "anthropic/claude-3-opus",
"provider": "openrouter",
"model": "anthropic/claude-3-opus",
"apiBase": "https://openrouter.ai/api/v1",
"apiKey": "{{OPENROUTER_API_KEY}}"
}-
If existing providers do not meet your requirements, you can evaluate specific models by creating a new
Provider. This is achieved by extending theBaseLLM:-
export abstract class BaseLLM { abstract provider: string abstract option: LLMOption info: Model abstract chat( compiledMessages: ChatMessage[], originOptions: CompletionOptions ): Promise<{ request: string error?: string response: string }> }
-
option– define parameters for LLM requests: -
export interface LLMOption { contextLength: number maxTokens: number temperature?: number apiBase: string }
-
info– model metadata inapps/eval/src/model.json. -
chat– custom request method that returns the generated text from the LLM.
-