Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Paper] AtmoRep: A stochastic model of atmosphere dynamics using large scale representation learning #103

Open
jacobbieker opened this issue Apr 29, 2024 · 2 comments
Labels
enhancement New feature or request

Comments

@jacobbieker
Copy link
Member

jacobbieker commented Apr 29, 2024

Arxiv/Blog/Paper Link

Paper: https://arxiv.org/abs/2308.13280
Code: https://github.com/clessig/atmorep

Detailed Description

Self-supervised representation learning that uses historical record and observations to constrain the representation of the atmosphere. The point of the model is to feed into downstream networks that take this representation as input (I.e. instead of analysis fields like most weather models, or direct observations, like in #64

Some interesting thoughts:

  1. They also struggle with getting areas from multi-dimensional NWP data efficiently, and try to mask the loading through data-parallel loading and making batches in parallel in memory.
  2. MultiFormer is their name for having one essentially vision transformer per input field (i.e. temperature, u wind, v wind, etc.) that is combined into one large model. The rationale for the multiple transformers is that each one can better learn the processes for how that field varies in time and space, and can have individually smaller transformers that can be swapped in and out for the final model predictions. So if you don't need the temperature, for example, then you don't need to load the temperature transformer, and the rest of MultiFormer would still run. They use cross-attention heads to have the different transformers communicate to each other.
  3. U-Net like connections between the encoder and decoder are used to keep multi-resolution features.
  4. Trained with a masked-reconstruction loss, has an internal ensemble prediction that gives probabilities for reconstructing the input masked tokens.
@jacobbieker jacobbieker added the enhancement New feature or request label Apr 29, 2024
@jacobbieker
Copy link
Member Author

image

@JackKelly another group of people who've had to work around getting random data in high-dimensional datasets. I think this is what we tried in nowcasting-dataset?

@JackKelly
Copy link
Member

Oooh, very nice, thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants