Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

method related files #1

Open
tshu-w opened this issue Nov 5, 2021 · 5 comments
Open

method related files #1

tshu-w opened this issue Nov 5, 2021 · 5 comments

Comments

@tshu-w
Copy link

tshu-w commented Nov 5, 2021

Hi, I'm interested your implementations, can you point me to method related files, I found it's hard to find it since you change trasformers and fairseq directly. BTW, I think inheritance or override class and function without change them directly is a better way.

@tshu-w tshu-w closed this as completed Nov 6, 2021
@ivanmontero
Copy link
Owner

ivanmontero commented Nov 10, 2021

Hey! Thanks for your interest, I'll update the repo soon to include the configuration files we used in the paper, and update this issue. Regarding using both fairseq and transformers: we use transformers for their pretrained model checkpoints (RoBERTa-{base, large}) and classification finetuning (GLUE), and use fairseq for any generation-related tasks (autoencoding, sentiment transfer)

Regarding inheritance and overriding: I completely agree. The main reason I did what I did was to follow closely to the way the transformers (and fairseq) do things, since they, even though two models may have very similar implementations, they copy and paste a lot of the code (i.e., the RoBERTa and BERT implementations)

@ivanmontero ivanmontero reopened this Nov 10, 2021
@tshu-w
Copy link
Author

tshu-w commented Nov 11, 2021

Hi, I read you paper carefully and understand how you do it. So code is not necessary for me to understand your work now. 😄

@NakagawaAkira
Copy link

Hi,
I'm very interested in your paper.
Fixed length embedding is very easy to treat!
I would like to run your code.
I'm looking forward to updating your document including example usage and configurations!

@ivanmontero
Copy link
Owner

Thanks for the reminder on this! I just pushed a commit that starts some documentation on the data processing, pretraining, and finetuning stages, as well as some configurations to perform those. I'll work on fleshing them out more in the near future!

@NakagawaAkira
Copy link

Thank you very much! Now it is clear for me to pretrain / fine-tune the model.
I'm also looking forward to your further update ;-)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants