Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add new pretrained weights #1043

Open
9 of 21 tasks
calebrob6 opened this issue Jan 25, 2023 · 7 comments
Open
9 of 21 tasks

Add new pretrained weights #1043

calebrob6 opened this issue Jan 25, 2023 · 7 comments
Labels
good first issue A good issue for a new contributor to work on models Models and pretrained weights

Comments

@calebrob6
Copy link
Member

calebrob6 commented Jan 25, 2023

Summary

This issue is to track progress on implementing new pretrained weights from related literature into torchgeo:

and many, many more:

Rationale

Foundation Models are one of the most substantial developments in recent ML research. FMs trained on ImageNet are one of the core components of torchvision and transformers that make them so popular. TorchGeo serves as a collection of EO FMs, allowing researchers to quickly and easily experiment with and design new FMs. This is critical for researchers to apply FMs to transfer learning on downstream tasks with small labeled datasets.

Implementation

See #2057, #1903, #1884, etc. for recent PRs adding new FMs.

If you would like to volunteer to add a particular FM, please comment on this issue to say that you're working on this.

Not sure where to get started? FMs that can be considered "multi-modal" (the same set of pre-trained weights can dynamically handle imagery from many different sensors) are the highest priority!

@adamjstewart adamjstewart added the models Models and pretrained weights label Jan 25, 2023
@nilsleh
Copy link
Collaborator

nilsleh commented Jan 25, 2023

  • Which of the 100k or 1M models were implemented?

1 M models were implemented

  • This is partially done (@nilsleh do you mind commenting which you didn't import and why?)

From there we are missing MAE and Date2Vec. I remember when first trying to extract the state dict for those, I was running into some weird issue. And then I forgot about it, but I will look into it again!

@nilsleh
Copy link
Collaborator

nilsleh commented Jan 27, 2023

Regarding these weights, there are three things I am running into:

  1. Their checkpoint contains model, optimizer states and more so at some point you would have to do state_dict = state_dict["model"] whereas all our other pretrained weights are just the model checkpoint already
  2. Their checkpoint file was saved from GPU, and I actually don't know where I can add the map_location=torch.device("cpu") argument except in the torchvision source code. We are calling get_state_dict() but it does not accept an argument for map_location but load_state_dict_from_url() does. So I think we need to open an issue with torchvision?
  3. several of their models are ViTAE models and I cannot find those in timm

It feels like the above two issues could happen again down the line with other pretrained weights as well, if they don't have the license for us to upload them to huggingface in the format we prefer.

@adamjstewart
Copy link
Collaborator

For 1 and 2, my preference would be to modify and save only the backbone on the CPU so that users or Lightning can map it to the GPU themselves. But of course that requires a favorable license. I would start by inquiring about licenses. If there's an issue with the license we can suggest changing it and describe the use case we have in mind. If they don't respond or won't change, then and only then can we think about more complex code in TorchGeo or changes to torchvision.

@calebrob6
Copy link
Member Author

Yep I think re-save the weights is the way to go.

@adamjstewart
Copy link
Collaborator

@adamjstewart adamjstewart added the good first issue A good issue for a new contributor to work on label May 23, 2024
@adamjstewart adamjstewart pinned this issue May 29, 2024
@wangyi111
Copy link
Contributor

I will work on adding bigger ViTs from SSL4EO-S12, FG-MAE, DeCUR, and SoftCon these days.

@Navoditamathur
Copy link

Navoditamathur commented Nov 21, 2024

I am familiar and have worked with Prithvi, and have been experimenting with SatMAE & SatMAE++ for landcover classification. I can work on them.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
good first issue A good issue for a new contributor to work on models Models and pretrained weights
Projects
None yet
Development

No branches or pull requests

5 participants