We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hello,
Pretrained-Language-Model/TinyBERT/data_augmentation.py
Lines 147 to 154 in 54ca698
In line 154, the tokenized text is sliced to len <= 512, if it exceeds 512 tokens but the corresponding tokenized_len in line 149 is not updated.
The segment_ids in the subsequent lines seem to be using the un-updated tokenized_len and causing errors in the forward pass.
File "/lusnlsas/paramsiddhi/iitm/vinodg/glue_data_generation/plm/TinyBERT/transformer/modeling.py", line 361, in forward embeddings = words_embeddings + position_embeddings + token_type_embeddings RuntimeError: The size of tensor a (512) must match the size of tensor b (763) at non-singleton dimension 1
^ this bug occurs when I try to generate data augmentations using the bert-base-cased model.
bert-base-cased
The text was updated successfully, but these errors were encountered:
Thanks! We agree with your comment and will fix this bug. Your pull request is also welcome.
Sorry, something went wrong.
@zwjyyc I've submitted a pull request for the same. Can you please help review this?
Successfully merging a pull request may close this issue.
Hello,
Pretrained-Language-Model/TinyBERT/data_augmentation.py
Lines 147 to 154 in 54ca698
In line 154, the tokenized text is sliced to len <= 512, if it exceeds 512 tokens but the corresponding tokenized_len in line 149 is not updated.
The segment_ids in the subsequent lines seem to be using the un-updated tokenized_len and causing errors in the forward pass.
^ this bug occurs when I try to generate data augmentations using the
bert-base-cased
model.The text was updated successfully, but these errors were encountered: