minerva.models.nets.image.vit
Attributes
Classes
A sequential container. |
|
Base class for all neural network modules. |
|
Base class for all neural network modules. |
|
Base class for all neural network modules. |
|
Masked Autoencoder with VisionTransformer backbone. |
|
Simple pipeline for supervised models. |
|
A sequential container. |
|
Vision Transformer with support for patch or hybrid CNN input stage |
|
Vision Transformer with support for global average pooling |
|
Transformer Model Encoder for sequence to sequence translation. |
|
Vision Transformer as per https://arxiv.org/abs/2010.11929. |
Functions
|
|
|
|
|
|
|
Module Contents
- class minerva.models.nets.image.vit.Conv2dReLU(in_channels, out_channels, kernel_size, padding=0, stride=1, use_batchnorm=True)[source]
Bases:
torch.nn.Sequential
A sequential container.
Modules will be added to it in the order they are passed in the constructor. Alternatively, an
OrderedDict
of modules can be passed in. Theforward()
method ofSequential
accepts any input and forwards it to the first module it contains. It then “chains” outputs to inputs sequentially for each subsequent module, finally returning the output of the last module.The value a
Sequential
provides over manually calling a sequence of modules is that it allows treating the whole container as a single module, such that performing a transformation on theSequential
applies to each of the modules it stores (which are each a registered submodule of theSequential
).What’s the difference between a
Sequential
and atorch.nn.ModuleList
? AModuleList
is exactly what it sounds like–a list for storingModule
s! On the other hand, the layers in aSequential
are connected in a cascading way.Example:
# Using Sequential to create a small model. When `model` is run, # input will first be passed to `Conv2d(1,20,5)`. The output of # `Conv2d(1,20,5)` will be used as the input to the first # `ReLU`; the output of the first `ReLU` will become the input # for `Conv2d(20,64,5)`. Finally, the output of # `Conv2d(20,64,5)` will be used as input to the second `ReLU` model = nn.Sequential( nn.Conv2d(1,20,5), nn.ReLU(), nn.Conv2d(20,64,5), nn.ReLU() ) # Using Sequential with OrderedDict. This is functionally the # same as the above code model = nn.Sequential(OrderedDict([ ('conv1', nn.Conv2d(1,20,5)), ('relu1', nn.ReLU()), ('conv2', nn.Conv2d(20,64,5)), ('relu2', nn.ReLU()) ]))
Initialize internal Module state, shared by both nn.Module and ScriptModule.
- class minerva.models.nets.image.vit.DecoderBlock(in_channels, out_channels, skip_channels=0, use_batchnorm=True)[source]
Bases:
torch.nn.Module
Base class for all neural network modules.
Your models should also subclass this class.
Modules can also contain other Modules, allowing them to be nested in a tree structure. You can assign the submodules as regular attributes:
import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self) -> None: super().__init__() self.conv1 = nn.Conv2d(1, 20, 5) self.conv2 = nn.Conv2d(20, 20, 5) def forward(self, x): x = F.relu(self.conv1(x)) return F.relu(self.conv2(x))
Submodules assigned in this way will be registered, and will also have their parameters converted when you call
to()
, etc.Note
As per the example above, an
__init__()
call to the parent class must be made before assignment on the child.- Variables:
training (bool) – Boolean represents whether this module is in training or evaluation mode.
Initialize internal Module state, shared by both nn.Module and ScriptModule.
- conv1
- conv2
- up
- class minerva.models.nets.image.vit.DecoderCup[source]
Bases:
torch.nn.Module
Base class for all neural network modules.
Your models should also subclass this class.
Modules can also contain other Modules, allowing them to be nested in a tree structure. You can assign the submodules as regular attributes:
import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self) -> None: super().__init__() self.conv1 = nn.Conv2d(1, 20, 5) self.conv2 = nn.Conv2d(20, 20, 5) def forward(self, x): x = F.relu(self.conv1(x)) return F.relu(self.conv2(x))
Submodules assigned in this way will be registered, and will also have their parameters converted when you call
to()
, etc.Note
As per the example above, an
__init__()
call to the parent class must be made before assignment on the child.- Variables:
training (bool) – Boolean represents whether this module is in training or evaluation mode.
Initialize internal Module state, shared by both nn.Module and ScriptModule.
- blocks
- conv_feature1
- conv_feature2
- conv_feature3
- conv_feature4
- conv_more
- up2
- up3
- up4
- class minerva.models.nets.image.vit.MLAHead(mla_channels=256, mlahead_channels=128, norm_cfg=None)[source]
Bases:
torch.nn.Module
Base class for all neural network modules.
Your models should also subclass this class.
Modules can also contain other Modules, allowing them to be nested in a tree structure. You can assign the submodules as regular attributes:
import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self) -> None: super().__init__() self.conv1 = nn.Conv2d(1, 20, 5) self.conv2 = nn.Conv2d(20, 20, 5) def forward(self, x): x = F.relu(self.conv1(x)) return F.relu(self.conv2(x))
Submodules assigned in this way will be registered, and will also have their parameters converted when you call
to()
, etc.Note
As per the example above, an
__init__()
call to the parent class must be made before assignment on the child.- Variables:
training (bool) – Boolean represents whether this module is in training or evaluation mode.
Initialize internal Module state, shared by both nn.Module and ScriptModule.
- head2
- head3
- head4
- head5
- class minerva.models.nets.image.vit.MaskedAutoencoderViT(img_size=224, patch_size=16, in_chans=1, embed_dim=1024, depth=24, num_heads=16, decoder_embed_dim=512, decoder_depth=8, decoder_num_heads=16, mlp_ratio=4.0, norm_layer=nn.LayerNorm, norm_pix_loss=False)[source]
Bases:
lightning.LightningModule
Masked Autoencoder with VisionTransformer backbone.
- Args:
img_size (int): Size of input image. patch_size (int): Size of image patch. in_chans (int): Number of input channels. embed_dim (int): Dimension of token embeddings. depth (int): Number of transformer blocks. num_heads (int): Number of attention heads. decoder_embed_dim (int): Dimension of decoder embeddings. decoder_depth (int): Number of decoder transformer blocks. decoder_num_heads (int): Number of decoder attention heads. mlp_ratio (float): Ratio of MLP hidden layer size to embedding size. norm_layer (torch.nn.LayerNorm): Normalization layer. norm_pix_loss (bool): Whether to normalize pixel loss.
- References:
Initialize internal Module state, shared by both nn.Module and ScriptModule.
- blocks
- cls_token
- decoder_blocks
- decoder_embed
- decoder_norm
- decoder_pos_embed
- decoder_pred
- forward(imgs, mask_ratio=0.75)[source]
Forward pass.
- Args:
imgs (torch.Tensor): Input images of shape (N, C, H, W). mask_ratio (float): Ratio of values to mask.
- Returns:
Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: Loss value, predicted output, binary mask.
- forward_decoder(x, ids_restore)[source]
Forward pass through the decoder.
- Args:
x (torch.Tensor): Input tensor of shape (N, L, D). ids_restore (torch.Tensor): Indices to restore the original order of patches.
- Returns:
torch.Tensor: Decoded output tensor of shape (N, L, patch_size^2 * in_chans).
- forward_encoder(x, mask_ratio)[source]
Forward pass through the encoder.
- Args:
x (torch.Tensor): Input tensor of shape (N, C, H, W). mask_ratio (float): Ratio of values to mask.
- Returns:
Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: Encoded representation, binary mask, shuffled indices.
- forward_loss(imgs, pred, mask)[source]
Calculate the loss.
- Args:
imgs (torch.Tensor): Input images of shape (N, C, H, W). pred (torch.Tensor): Predicted output of shape (N, L, patch_size^2 * in_chans). mask (torch.Tensor): Binary mask of shape (N, L).
- Returns:
torch.Tensor: Computed loss value.
- in_chans = 1
- mask_token
- norm
- norm_pix_loss = False
- patch_embed
- patchify(imgs)[source]
Extract patches from input images.
- Args:
imgs (torch.Tensor): Input images of shape (N, C, H, W).
- Returns:
torch.Tensor: Patches of shape (N, num_patches, patch_size^2 * in_chans).
- pos_embed
- random_masking(x, mask_ratio)[source]
Perform per-sample random masking by per-sample shuffling.
- Args:
x (torch.Tensor): Input tensor of shape (N, L, D). mask_ratio (float): Ratio of values to mask.
- Returns:
Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: Masked input, binary mask, shuffled indices.
- training_step(batch, batch_idx)[source]
Training step.
- Args:
batch (Tuple[torch.Tensor]): Input batch of images and corresponding labels. batch_idx (int): Index of the current batch.
- Returns:
Dict[str, torch.Tensor]: Dictionary containing the loss value for the current step.
- class minerva.models.nets.image.vit.SFM_BasePatch16_Downstream(img_size=(512, 512), num_classes=6, in_chans=1, loss_fn=None, learning_rate=0.001, **kwargs)[source]
Bases:
minerva.models.nets.base.SimpleSupervisedModel
Simple pipeline for supervised models.
This class implements a very common deep learning pipeline, which is composed by the following steps:
Make a forward pass with the input data on the backbone model;
Make a forward pass with the input data on the fc model;
Compute the loss between the output and the label data;
Optimize the model (backbone and FC) parameters with respect to the loss.
This reduces the code duplication for autoencoder models, and makes it easier to implement new models by only changing the backbone model. More complex models, that does not follow this pipeline, should not inherit from this class. Note that, for this class the input data is a tuple of tensors, where the first tensor is the input data and the second tensor is the mask or label.
Create a SFM model with a ViT base backbone. The ViT-Base-16 backbone has the following configuration: - Patch size: 16 - Embedding dimension: 768 - Depth: 12 - Number of heads: 12
Parameters
- img_sizeUnion[int, Tuple[int, …]]
Size of the input image. Note that, to use default pre-trained SFM model, the size should be (512, 512).
- num_classesint
Number of classes for segmentation head. Default is 6.
- in_chansint
Number of input channels. Default is 1.
- loss_fnOptional[torch.nn.Module], optional
Loss function, by default None
- learning_ratefloat, optional
Learning rate value, by default 1e-3
- _single_step(batch, batch_idx, step_name)[source]
Perform a single train/validation/test step. It consists in making a forward pass with the input data on the backbone model, computing the loss between the output and the input data, and logging the loss.
Parameters
- batchtorch.Tensor
The input data. It must be a 2-element tuple of tensors, where the first tensor is the input data and the second tensor is the mask.
- batch_idxint
The index of the batch.
- step_namestr
The name of the step. It will be used to log the loss. The possible values are: “train”, “val” and “test”. The loss will be logged as “{step_name}_loss”.
Returns
- torch.Tensor
A tensor with the loss value.
- Parameters:
batch (torch.Tensor)
batch_idx (int)
step_name (str)
- Return type:
torch.Tensor
- predict_step(batch, batch_idx, dataloader_idx=0)[source]
Step function called during
predict()
. By default, it callsforward()
. Override to add any processing logic.The
predict_step()
is used to scale inference on multi-devices.To prevent an OOM error, it is possible to use
BasePredictionWriter
callback to write the predictions to disk or database after each batch or on epoch end.The
BasePredictionWriter
should be used while using a spawn based accelerator. This happens forTrainer(strategy="ddp_spawn")
or training on 8 TPU cores withTrainer(accelerator="tpu", devices=8)
as predictions won’t be returned.- Args:
batch: The output of your data iterable, normally a
DataLoader
. batch_idx: The index of this batch. dataloader_idx: The index of the dataloader that produced this batch.(only if multiple dataloaders used)
- Return:
Predicted output (optional).
Example
class MyModel(LightningModule): def predict_step(self, batch, batch_idx, dataloader_idx=0): return self(batch) dm = ... model = MyModel() trainer = Trainer(accelerator="gpu", devices=2) predictions = trainer.predict(model, dm)
- Parameters:
batch (torch.Tensor)
batch_idx (int)
dataloader_idx (int)
- Return type:
torch.Tensor
- Parameters:
img_size (Union[int, Tuple[int, Ellipsis]])
num_classes (int)
in_chans (int)
loss_fn (Optional[torch.nn.Module])
learning_rate (float)
- class minerva.models.nets.image.vit.SegmentationHead(in_channels, out_channels, kernel_size=3, upsampling=1)[source]
Bases:
torch.nn.Sequential
A sequential container.
Modules will be added to it in the order they are passed in the constructor. Alternatively, an
OrderedDict
of modules can be passed in. Theforward()
method ofSequential
accepts any input and forwards it to the first module it contains. It then “chains” outputs to inputs sequentially for each subsequent module, finally returning the output of the last module.The value a
Sequential
provides over manually calling a sequence of modules is that it allows treating the whole container as a single module, such that performing a transformation on theSequential
applies to each of the modules it stores (which are each a registered submodule of theSequential
).What’s the difference between a
Sequential
and atorch.nn.ModuleList
? AModuleList
is exactly what it sounds like–a list for storingModule
s! On the other hand, the layers in aSequential
are connected in a cascading way.Example:
# Using Sequential to create a small model. When `model` is run, # input will first be passed to `Conv2d(1,20,5)`. The output of # `Conv2d(1,20,5)` will be used as the input to the first # `ReLU`; the output of the first `ReLU` will become the input # for `Conv2d(20,64,5)`. Finally, the output of # `Conv2d(20,64,5)` will be used as input to the second `ReLU` model = nn.Sequential( nn.Conv2d(1,20,5), nn.ReLU(), nn.Conv2d(20,64,5), nn.ReLU() ) # Using Sequential with OrderedDict. This is functionally the # same as the above code model = nn.Sequential(OrderedDict([ ('conv1', nn.Conv2d(1,20,5)), ('relu1', nn.ReLU()), ('conv2', nn.Conv2d(20,64,5)), ('relu2', nn.ReLU()) ]))
Initialize internal Module state, shared by both nn.Module and ScriptModule.
- class minerva.models.nets.image.vit.VIT_MLAHead(img_size=768, mla_channels=256, mlahead_channels=128, num_classes=6, norm_layer=nn.BatchNorm2d, norm_cfg=None, **kwargs)[source]
Bases:
torch.nn.Module
Vision Transformer with support for patch or hybrid CNN input stage
Initialize internal Module state, shared by both nn.Module and ScriptModule.
- BatchNorm
- cls
- img_size = 768
- mla_channels = 256
- mlahead
- mlahead_channels = 128
- norm_cfg = None
- num_classes = 6
- class minerva.models.nets.image.vit.VisionTransformer(global_pool=False, **kwargs)[source]
Bases:
timm.models.vision_transformer.VisionTransformer
,lightning.LightningModule
Vision Transformer with support for global average pooling
- Args:
img_size: Input image size. patch_size: Patch size. in_chans: Number of image input channels. num_classes: Number of classes for classification head. global_pool: Type of global pooling for final sequence (default: ‘token’). embed_dim: Transformer embedding dimension. depth: Depth of transformer. num_heads: Number of attention heads. mlp_ratio: Ratio of mlp hidden dim to embedding dim. qkv_bias: Enable bias for qkv projections if True. init_values: Layer-scale init values (layer-scale enabled if not None). class_token: Use class token. no_embed_class: Don’t include position embeddings for class (or reg) tokens. reg_tokens: Number of register tokens. pre_norm: Enable norm after embeddings, before transformer blocks (standard in CLIP ViT). final_norm: Enable norm after transformer blocks, before head (standard in most ViT). fc_norm: Move final norm after pool (instead of before), if None, enabled when global_pool == ‘avg’. drop_rate: Head dropout rate. pos_drop_rate: Position embedding dropout rate. attn_drop_rate: Attention dropout rate. drop_path_rate: Stochastic depth rate. weight_init: Weight initialization scheme. fix_init: Apply weight initialization fix (scaling w/ layer index). embed_layer: Patch embedding layer. embed_norm_layer: Normalization layer to use / override in patch embed module. norm_layer: Normalization layer. act_layer: MLP activation layer. block_fn: Transformer block layer.
- decoder
- global_pool = False
- loss_fn
- segmentation_head
- class minerva.models.nets.image.vit._Encoder(seq_length, num_layers, num_heads, hidden_dim, mlp_dim, dropout, attention_dropout, aux_output=False, aux_output_layers=None, norm_layer=partial(nn.LayerNorm, eps=1e-06))[source]
Bases:
torch.nn.Module
Transformer Model Encoder for sequence to sequence translation.
Initialize internal Module state, shared by both nn.Module and ScriptModule.
- Parameters:
seq_length (int)
num_layers (int)
num_heads (int)
hidden_dim (int)
mlp_dim (int)
dropout (float)
attention_dropout (float)
aux_output (bool)
aux_output_layers (Optional[List[int]])
norm_layer (Callable[Ellipsis, torch.nn.Module])
- aux_output = False
- aux_output_layers = None
- dropout
- layers
- ln
- pos_embedding
- class minerva.models.nets.image.vit._VisionTransformerBackbone(image_size, patch_size, num_layers, num_heads, hidden_dim, mlp_dim, original_resolution=None, dropout=0.0, attention_dropout=0.0, num_classes=1000, aux_output=False, aux_output_layers=None, norm_layer=partial(nn.LayerNorm, eps=1e-06), conv_stem_configs=None)[source]
Bases:
torch.nn.Module
Vision Transformer as per https://arxiv.org/abs/2010.11929.
Initializes a Vision Transformer (ViT) model.
Parameters
- image_sizeint or Tuple[int, int]
The size of the input image. If an int is provided, it is assumed to be a square image. If a tuple of ints is provided, it represents the height and width of the image.
- patch_sizeint
The size of each patch in the image.
- num_layersint
The number of transformer layers in the model.
- num_headsint
The number of attention heads in the transformer layers.
- hidden_dimint
The dimensionality of the hidden layers in the transformer.
- mlp_dimint
The dimensionality of the feed-forward MLP layers in the transformer
- original_resolutionTuple[int, int], optional
The original resolution of the input image in the pre-training weights. When None, positional embeddings will not be interpolated. Defaults to None.
- dropoutfloat, optional
The dropout rate to apply. Defaults to 0.0.
- attention_dropoutfloat, optional
The dropout rate to apply to the attention weights. Defaults to 0.0
- num_classesint, optional
The number of output classes. Defaults to 1000.
- norm_layerCallable[…, torch.nn.Module], optional
The normalization layer to use. Defaults to nn.LayerNorm with epsilon=1e-6.
- conv_stem_configsList[ConvStemConfig], optional
The configuration for the convolutional stem layers. If provided, the input image will be processed by these convolutional layers before being passed to the transformer. Defaults to None.
- _process_input(x)[source]
Process the input tensor and return the reshaped tensor and dimensions.
- Args:
x (torch.Tensor): The input tensor.
- Returns:
Tuple[torch.Tensor, int, int]: The reshaped tensor, number of rows, and number of columns.
- Parameters:
x (torch.Tensor)
- Return type:
Tuple[torch.Tensor, int, int]
- attention_dropout = 0.0
- aux_output = False
- aux_output_layers = None
- class_token
- dropout = 0.0
- encoder
- forward(x)[source]
Forward pass of the Vision Transformer Backbone.
- Args:
x (torch.Tensor): The input tensor.
- Returns:
torch.Tensor: The output tensor.
- Parameters:
x (torch.Tensor)
- image_size
- interpolate_pos_embeddings(pretrained_pos_embed, new_img_size)[source]
Interpolate encoder’s positional embeddings to fit a new input size.
- Args:
pretrained_pos_embed (torch.Tensor): Pretrained positional embeddings. new_img_size (Tuple[int, int]): New height and width of the input image.
- load_backbone(path, freeze=False)[source]
Loads pretrained weights and handles positional embedding resizing if necessary.
- Parameters:
path (str)
freeze (bool)
- mlp_dim
- norm_layer
- num_classes = 1000
- original_resolution
- patch_size
- seq_length
- Parameters:
image_size (Union[int, Tuple[int, int]])
patch_size (int)
num_layers (int)
num_heads (int)
hidden_dim (int)
mlp_dim (int)
original_resolution (Optional[Tuple[int, int]])
dropout (float)
attention_dropout (float)
num_classes (int)
aux_output (bool)
aux_output_layers (Optional[List[int]])
norm_layer (Callable[Ellipsis, torch.nn.Module])
conv_stem_configs (Optional[List[torchvision.models.vision_transformer.ConvStemConfig]])
- minerva.models.nets.image.vit.interpolate_pos_embed(model, checkpoint_model, newsize1=None, newsize2=None)[source]
- minerva.models.nets.image.vit.mae_vit_base_patch16
- minerva.models.nets.image.vit.mae_vit_base_patch16D4d256
- minerva.models.nets.image.vit.mae_vit_huge_patch14
- minerva.models.nets.image.vit.mae_vit_large_patch16
- minerva.models.nets.image.vit.mae_vit_large_patch16D4d256
- minerva.models.nets.image.vit.mae_vit_small_patch16