Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Minimal loading pretrain BEiT model #373

Closed
thunanguyen opened this issue Aug 1, 2021 · 3 comments
Closed

Minimal loading pretrain BEiT model #373

thunanguyen opened this issue Aug 1, 2021 · 3 comments

Comments

@thunanguyen
Copy link

I want to ask if there is a minimal loading pre-trained model of BEiT because I want to use BEiT as a backbone in my research about multi-label classification. I don't need all the code in this repo but only the loading code for BEiT, but currently the loading code is so confusing for me, I think you should provide a torch hub model so that everyone can easily access to your model more.

@donglixp
Copy link
Contributor

donglixp commented Aug 4, 2021

Hi @thunanguyen , thanks for the good suggestion! Huggingface is working on merging BEiT models in their model hub (huggingface/transformers#12994 (comment) ). It would be much easier to load the checkpoints with user-friendly APIs.

@donglixp donglixp closed this as completed Aug 4, 2021
@NielsRogge
Copy link

BEiT has been merged: https://huggingface.co/transformers/master/model_doc/beit.html :)

@thunanguyen
Copy link
Author

thunanguyen commented Aug 4, 2021

Great! Will try it immediately! One more question though. Do you think the use of a Squeeze layer like in SqueezeBERT can make the model smaller and faster, @donglixp? Since your large model is too large.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants