Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

train_byte_level_bpe.py 'pad_to_max_length' error #11

Open
busraoguzoglu opened this issue Jul 21, 2022 · 1 comment
Open

train_byte_level_bpe.py 'pad_to_max_length' error #11

busraoguzoglu opened this issue Jul 21, 2022 · 1 comment

Comments

@busraoguzoglu
Copy link

Hello,

I am trying to use pretrained tokenizers to test out encoding using the train_byte_level_bpe.py file. (From Restoring model from learned vocab/merges)
Even though I have created an environment using conda.yml file, I am having an error with .encode() function:
TypeError: encode() got an unexpected keyword argument 'pad_to_max_length'
When I searched, I have found that it can be related to the version of tokenizers or transformers library, do you know how can I solve it?

Thanks a lot for your help

@drugilsberg
Copy link
Contributor

Hi @busraoguzoglu, this is exactly the issue you are observing, please try using the version of the packages suggested here.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants