Skip to content

Commit ed9f75a

Browse files
committed
replace hyperlink with modular path
Signed-off-by: h-guo18 <[email protected]>
1 parent 80dfcb5 commit ed9f75a

File tree

4 files changed

+17
-18
lines changed

4 files changed

+17
-18
lines changed

docs/source/torch/auto_deploy/advanced/example_run.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
1-
# Example Run Script ([`build_and_run_ad.py`](../../../../../examples/auto_deploy/build_and_run_ad.py))
1+
# Example Run Script
22

3-
To build and run AutoDeploy example, use the [`build_and_run_ad.py`](../../../../../examples/auto_deploy/build_and_run_ad.py) script:
3+
To build and run AutoDeploy example, use `examples/auto_deploy/build_and_run_ad.py` script:
44

55
```bash
66
cd examples/auto_deploy
@@ -33,7 +33,7 @@ Below is a non-exhaustive list of common config options:
3333
| `--prompt.batch-size` | Number of queries to generate |
3434
| `--benchmark.enabled` | Whether to run the built-in benchmark (true/false) |
3535

36-
For default values and additional configuration options, refer to the `ExperimentConfig` class in [build_and_run_ad.py](../../../../../examples/auto_deploy/build_and_run_ad.py) file.
36+
For default values and additional configuration options, refer to the `ExperimentConfig` class in `examples/auto_deploy/build_and_run_ad.py` file.
3737

3838
Here is a more complete example of using the script:
3939

docs/source/torch/auto_deploy/advanced/expert_configurations.md

Lines changed: 10 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,29 +1,28 @@
11
# Expert Configuration of LLM API
22

3-
For expert TensorRT-LLM users, we also expose the full set of [`LlmArgs`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py)
3+
For expert TensorRT-LLM users, we also expose the full set of `tensorrt_llm._torch.auto_deploy.llm_args.LlmArgs`
44
*at your own risk* (the argument list diverges from TRT-LLM's argument list):
55

66
- All config fields that are used by the AutoDeploy core pipeline (i.e. the `InferenceOptimizer`) are
7-
_exclusively_ exposed in the [`AutoDeployConfig` class](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py).
7+
_exclusively_ exposed in the `AutoDeployConfig` in `tensorrt_llm._torch.auto_deploy.llm_args`.
88
Please make sure to refer to those first.
9-
- For expert users we expose the full set of [`LlmArgs`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py)
10-
that can be used to configure the [AutoDeploy `LLM` API](../../../../../tensorrt_llm/_torch/auto_deploy/llm.py) including runtime options.
11-
- Note that some fields in the full [`LlmArgs`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py)
9+
- For expert users we expose the full set of `LlmArgs` in `tensorrt_llm._torch.auto_deploy.llm_args`
10+
that can be used to configure the AutoDeploy `LLM` API including runtime options.
11+
- Note that some fields in the full `LlmArgs`
1212
object are overlapping, duplicated, and/or _ignored_ in AutoDeploy, particularly arguments
1313
pertaining to configuring the model itself since AutoDeploy's model ingestion+optimize pipeline
1414
significantly differs from the default manual workflow in TensorRT-LLM.
15-
- However, with the proper care the full [`LlmArgs`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py)
15+
- However, with the proper care the full `LlmArgs`
1616
objects can be used to configure advanced runtime options in TensorRT-LLM.
17-
- Note that any valid field can be simply provided as keyword argument ("`**kwargs`") to the
18-
[AutoDeploy `LLM` API](../../../../../tensorrt_llm/_torch/auto_deploy/llm.py).
17+
- Note that any valid field can be simply provided as keyword argument ("`**kwargs`") to the AutoDeploy `LLM` API.
1918

2019
# Expert Configuration of `build_and_run_ad.py`
2120

2221
For expert users, `build_and_run_ad.py` provides advanced configuration capabilities through a flexible argument parser powered by PyDantic Settings and OmegaConf. You can use dot notation for CLI arguments, provide multiple YAML configuration files, and leverage sophisticated configuration precedence rules to create complex deployment configurations.
2322

2423
## CLI Arguments with Dot Notation
2524

26-
The script supports flexible CLI argument parsing using dot notation to modify nested configurations dynamically. You can target any field in both the [`ExperimentConfig`](../../../../../examples/auto_deploy/build_and_run_ad.py) and nested [`AutoDeployConfig`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py)/[`LlmArgs`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.) objects:
25+
The script supports flexible CLI argument parsing using dot notation to modify nested configurations dynamically. You can target any field in both the `ExperimentConfig` in `examples/auto_deploy/build_and_run_ad.py` and nested `AutoDeployConfig`/`LlmArgs` objects in `tensorrt_llm._torch.auto_deploy.llm_args`:
2726

2827
```bash
2928
# Configure model parameters
@@ -56,7 +55,7 @@ python build_and_run_ad.py \
5655

5756
## YAML Configuration Files
5857

59-
Both [`ExperimentConfig`](../../../../../examples/auto_deploy/build_and_run_ad.py) and [`AutoDeployConfig`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py)/[`LlmArgs`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py) inherit from [`DynamicYamlMixInForSettings`](../../../../../tensorrt_llm/_torch/auto_deploy/utils/_config.py), enabling you to provide multiple YAML configuration files that are automatically deep-merged at runtime.
58+
Both `ExperimentConfig` and `AutoDeployConfig`/`LlmArgs` inherit from `DynamicYamlMixInForSettings`, enabling you to provide multiple YAML configuration files that are automatically deep-merged at runtime.
6059

6160
Create a YAML configuration file (e.g., `my_config.yaml`):
6261

@@ -167,7 +166,7 @@ python build_and_run_ad.py \
167166

168167
## Built-in Default Configuration
169168

170-
Both [`AutoDeployConfig`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py) and [`LlmArgs`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py) classes automatically load a built-in [`default.yaml`](../../../../../tensorrt_llm/_torch/auto_deploy/config/default.yaml) configuration file that provides sensible defaults for the AutoDeploy inference optimizer pipeline. This file is specified in the [`_get_config_dict()`](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py) function and defines default transform configurations for graph optimization stages.
169+
Both `AutoDeployConfig` and `LlmArgs` classes automatically load a built-in `default.yaml` configuration file that provides sensible defaults for the AutoDeploy inference optimizer pipeline. This file is specified in the `_get_config_dict()` function in `tensorrt_llm._torch.auto_deploy.llm_args` and defines default transform configurations for graph optimization stages.
171170

172171
The built-in defaults are automatically merged with your configurations at the lowest priority level, ensuring that your custom settings always override the defaults. You can inspect the current default configuration to understand the baseline transform pipeline:
173172

docs/source/torch/auto_deploy/advanced/workflow.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,6 @@ llm = LLM(
2727
2828
```
2929

30-
Please consult the [AutoDeploy `LLM` API](../../../../../tensorrt_llm/_torch/auto_deploy/llm.py) and the
31-
[`AutoDeployConfig` class](../../../../../tensorrt_llm/_torch/auto_deploy/llm_args.py)
30+
Please consult the AutoDeploy `LLM` API in `tensorrt_llm._torch.auto_deploy.llm` and the
31+
`AutoDeployConfig` class in `tensorrt_llm._torch.auto_deploy.llm_args`
3232
for more detail on how AutoDeploy is configured via the `**kwargs` of the `LLM` API.

docs/source/torch/auto_deploy/auto-deploy.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -77,5 +77,5 @@ We are actively expanding AutoDeploy to support a broader range of model archite
7777

7878
- Speculative Decoding for accelerated generation
7979

80-
To track development progress and contribute, visit our [Github Project Board](https://github.com/orgs/NVIDIA/projects/83).
81-
We welcome community contributions, see our [`CONTRIBUTING.md`](../../../../../examples/auto_deploy/CONTRIBUTING.md) for guidelines.
80+
To track development progress and contribute, visit our [Github Project Board](https://github.com/orgs/NVIDIA/projects/83/views/13).
81+
We welcome community contributions, see `examples/auto_deploy/CONTRIBUTING.md` for guidelines.

0 commit comments

Comments
 (0)