8000 GitHub - saiprabhakar/ChannelMAE: ChannelMAE is a combination of MAE and ChannelViT
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

saiprabhakar/ChannelMAE

 
 

Repository files navigation

ChannelMAE (Masked Autoencoders + Channel Vision Transformer): A PyTorch Implementation

This is a fun repo, it combines Masked AutoEncoders (MAE) with Channel Vision Transformers (ChannelVit) to Channel Masked AutoEncoders (ChannelMAE). It is essent 7C8F ially channels getting rolled out during MAE pretraining. This repos also has support for subsequent fine-tunning.

Masked AutoEncoders (MAE): Are a powerfull pretraining model were we mask out often 75% of the images and make the model predict the rest.

Channel Vision Transformers (ChannelVit): Are useful in non-traditional image applications like cell-painting images or satellite images, where each channels conveys very different information and it doesnt makes sense stacking them up.

Channel Masked AutoEncoders (ChannelMAE): Combining them can be useful for pretraining models for non-traditional image applications like cell-painting images or satellite images. This is based on MAE and Vit implementation https://github.com/facebookresearch/mae, modified to add ChannelVit and ChannelMAE

Contributions

  • Pre-training code for MAE and ChannelMAE
  • Fine-tuning code for Vit and ChannelVit (encoders of MAE and ChannelMAE)
  • Linprobe code for Vit and ChannelVit (encoders of MAE and ChannelMAE)

We implemented ChannelMAE in models_chamae.py We implemented ChannelVit in models_vit.py

Pre-training

Sample testing implementation use: python main_pretrain.py

For submitting jobs: The instruction is in PRETRAIN.md.

Fintunning and Linear Probing

Sample testing implementation use: python main_finetune.py python main_lineprobe.py

For submitting jobs: The instruction is in FINETUNE.md.

Masked Autoencoders

Masked Autoencoders Are Scalable Vision Learners

Channel Vit

Channel Vision Transformer: An Image Is Worth C x 16 x 16 Words

my alt text


License

This project is under the CC-BY-NC 4.0 license. See LICENSE for details.

0