Skip to content
This repository was archived by the owner on Sep 10, 2025. It is now read-only.

Commit 776a15d

Browse files
authored
fix max sequence length for xlmr transform (#1495)
1 parent 6d9e6df commit 776a15d

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

torchtext/models/roberta/bundler.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@ def encoderConf(self) -> RobertaEncoderConf:
158158
transform=lambda: T.Sequential(
159159
T.SentencePieceTokenizer(urljoin(_TEXT_BUCKET, "xlmr.sentencepiece.bpe.model")),
160160
T.VocabTransform(load_state_dict_from_url(urljoin(_TEXT_BUCKET, "xlmr.vocab.pt"))),
161-
T.Truncate(510),
161+
T.Truncate(254),
162162
T.AddToken(token=0, begin=True),
163163
T.AddToken(token=2, begin=False),
164164
)

0 commit comments

Comments
 (0)