Skip to content
This repository was archived by the owner on Nov 1, 2024. It is now read-only.

Commit 48b9b6c

Browse files
committed
Docs
1 parent 0118e3c commit 48b9b6c

File tree

1 file changed

+10
-1
lines changed

1 file changed

+10
-1
lines changed

projects/OPT/download_opt175b.md

Lines changed: 10 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
After receiving an email with a presigned URL to access the model weights, follow the below set of instructions to get started with hosting the model.
44

55
## Download all shards
6-
Since we trained OPT-175B on 124 hosts, we have 124*8 = 992 files corresponding to the model parameters (8x tensor parallelism). The pre-signed URL that you receive in your email will look something like the following:
6+
Since we trained OPT-175B on 124 hosts, we have 124\*8 = 992 files corresponding to the model parameters (8x tensor parallelism). The pre-signed URL that you receive in your email will look something like the following:
77

88
```
99
https://<cloudfront_url>/175b/checkpoint_last_20220503/stubbed.pt?&<super_long_query_string>
@@ -20,6 +20,15 @@ To consolidate the 992 shards into 8 files model-parallel evaluation, run (assum
2020
bash metaseq/scripts/reshard_sbatch.sh <directory_where_all_the_shards_are>/checkpoint_last <output_dir>/ 8 1
2121
```
2222

23+
Note that most of our models expect to run with Model (Tensor) Parallelism. For smaller models, some
24+
users may find it easier to eliminate model parallelism. The checkpoints can be converted
25+
to eliminate use of MP with the `consolidate_fsdp_shards.py` script:
26+
27+
```bash
28+
python metaseq.scripts.consolidate_fsdp_shards ${FOLDER_PATH}/checkpoint_last --new-arch-name transformer_lm_gpt --save-prefix ${FOLDER_PATH}/consolidated
29+
```
30+
31+
2332
## Run the API
2433
Follow the instructions in the [API docs](../../docs/api.md) to spin up the API. You will need to update the constants in `metaseq/service/constants.py` to point to right directories.
2534

0 commit comments

Comments
 (0)