Skip to content

Commit a27195b

Browse files
authored
Update longt5.mdx (#18634)
1 parent fd9aa82 commit a27195b

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/source/en/model_doc/longt5.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ Tips:
3737
- [`LongT5ForConditionalGeneration`] is an extension of [`T5ForConditionalGeneration`] exchanging the traditional
3838
encoder *self-attention* layer with efficient either *local* attention or *transient-global* (*tglobal*) attention.
3939
- Unlike the T5 model, LongT5 does not use a task prefix. Furthermore, it uses a different pre-training objective
40-
inspired by the pre-training of `[PegasusForConditionalGeneration]`.
40+
inspired by the pre-training of [`PegasusForConditionalGeneration`].
4141
- LongT5 model is designed to work efficiently and very well on long-range *sequence-to-sequence* tasks where the
4242
input sequence exceeds commonly used 512 tokens. It is capable of handling input sequences of a length up to 16,384 tokens.
4343
- For *Local Attention*, the sparse sliding-window local attention operation allows a given token to attend only `r`

0 commit comments

Comments
 (0)