Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions generate_reconstructions.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -26,9 +26,9 @@
"import numpy as np\n",
"from PIL import Image\n",
"from matplotlib import pyplot as plt\n",
"from huggingface_mae import MAEModel\n",
"from maes_microscopy.huggingface_mae import MAEModel\n",
"from skimage.transform import resize\n",
"from mae_utils import unflatten_tokens"
"from maes_microscopy.mae_utils import unflatten_tokens"
]
},
{
Expand Down
Empty file added maes_microscopy/__init__.py
Empty file.
10 changes: 5 additions & 5 deletions huggingface_mae.py → maes_microscopy/huggingface_mae.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,11 +5,11 @@

from transformers import PretrainedConfig, PreTrainedModel

from loss import FourierLoss
from normalizer import Normalizer
from mae_modules import CAMAEDecoder, MAEDecoder, MAEEncoder
from mae_utils import flatten_images
from vit import (
from maes_microscopy.loss import FourierLoss
from maes_microscopy.normalizer import Normalizer
from maes_microscopy.mae_modules import CAMAEDecoder, MAEDecoder, MAEEncoder
from maes_microscopy.mae_utils import flatten_images
from maes_microscopy.vit import (
generate_2d_sincos_pos_embeddings,
sincos_positional_encoding_vit,
vit_small_patch16_256,
Expand Down
File renamed without changes.
4 changes: 2 additions & 2 deletions mae_modules.py → maes_microscopy/mae_modules.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,8 @@
from timm.models.helpers import checkpoint_seq
from timm.models.vision_transformer import Block, Mlp, VisionTransformer

from masking import transformer_random_masking
from vit import channel_agnostic_vit
from maes_microscopy.masking import transformer_random_masking
from maes_microscopy.vit import channel_agnostic_vit

# If interested in training new MAEs, combine an encoder and decoder into a new module, and you should
# leverage the flattening and unflattening utilities as needed from mae_utils.py.
Expand Down
File renamed without changes.
File renamed without changes.
File renamed without changes.
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import pytest
import torch

from huggingface_mae import MAEModel
from maes_microscopy.huggingface_mae import MAEModel

huggingface_openphenom_model_dir = "."
# huggingface_modelpath = "recursionpharma/OpenPhenom"
Expand Down
File renamed without changes.
File renamed without changes.
2 changes: 1 addition & 1 deletion pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -30,4 +30,4 @@ dependencies = [
]

[tool.setuptools]
py-modules = []
packages=["maes_microscopy"]