Source code for torchio.transforms.augmentation.spatial.random_anisotropy

import warnings
from typing import Tuple
from typing import Union

import torch

from .. import RandomTransform
from ....data.subject import Subject
from ....typing import TypeRangeFloat
from ....utils import to_tuple
from ...preprocessing import Resample


[docs] class RandomAnisotropy(RandomTransform): r"""Downsample an image along an axis and upsample to initial space. This transform simulates an image that has been acquired using anisotropic spacing and resampled back to its original spacing. Similar to the work by Billot et al.: `Partial Volume Segmentation of Brain MRI Scans of any Resolution and Contrast <https://link.springer.com/chapter/10.1007/978-3-030-59728-3_18>`_. Args: axes: Axis or tuple of axes along which the image will be downsampled. downsampling: Downsampling factor :math:`m \gt 1`. If a tuple :math:`(a, b)` is provided then :math:`m \sim \mathcal{U}(a, b)`. image_interpolation: Image interpolation used to upsample the image back to its initial spacing. Downsampling is performed using nearest neighbor interpolation. See :ref:`Interpolation` for supported interpolation types. scalars_only: Apply only to instances of :class:`torchio.ScalarImage`. This is useful when the segmentation quality needs to be kept, as in `Billot et al. <https://link.springer.com/chapter/10.1007/978-3-030-59728-3_18>`_. **kwargs: See :class:`~torchio.transforms.Transform` for additional keyword arguments. Example: >>> import torchio as tio >>> transform = tio.RandomAnisotropy(axes=1, downsampling=2) >>> transform = tio.RandomAnisotropy( ... axes=(0, 1, 2), ... downsampling=(2, 5), ... ) # Multiply spacing of one of the 3 axes by a factor randomly chosen in [2, 5] >>> colin = tio.datasets.Colin27() >>> transformed = transform(colin) """ # noqa: B950 def __init__( self, axes: Union[int, Tuple[int, ...]] = (0, 1, 2), downsampling: TypeRangeFloat = (1.5, 5), image_interpolation: str = 'linear', scalars_only: bool = True, **kwargs, ): super().__init__(**kwargs) self.axes = self.parse_axes(axes) self.downsampling_range = self._parse_range( downsampling, 'downsampling', min_constraint=1, ) parsed_interpolation = self.parse_interpolation(image_interpolation) self.image_interpolation = parsed_interpolation self.scalars_only = scalars_only def get_params( self, axes: Tuple[int, ...], downsampling_range: Tuple[float, float], ) -> Tuple[int, float]: axis = axes[torch.randint(0, len(axes), (1,))] downsampling = self.sample_uniform(*downsampling_range) return axis, downsampling @staticmethod def parse_axes(axes: Union[int, Tuple[int, ...]]): axes_tuple = to_tuple(axes) for axis in axes_tuple: is_int = isinstance(axis, int) if not is_int or axis not in (0, 1, 2): raise ValueError('All axes must be 0, 1 or 2') return axes_tuple def apply_transform(self, subject: Subject) -> Subject: is_2d = subject.get_first_image().is_2d() if is_2d and 2 in self.axes: warnings.warn( f'Input image is 2D, but "2" is in axes: {self.axes}', RuntimeWarning, stacklevel=2, ) self.axes = list(self.axes) self.axes.remove(2) axis, downsampling = self.get_params( self.axes, self.downsampling_range, ) target_spacing = list(subject.spacing) target_spacing[axis] *= downsampling arguments = { 'image_interpolation': 'nearest', 'scalars_only': self.scalars_only, } sx, sy, sz = target_spacing # for mypy downsample = Resample( target=(sx, sy, sz), **self.add_include_exclude(arguments) ) downsampled = downsample(subject) image = subject.get_first_image() target = image.spatial_shape, image.affine upsample = Resample( target=target, # type: ignore[arg-type] image_interpolation=self.image_interpolation, scalars_only=self.scalars_only, ) upsampled = upsample(downsampled) assert isinstance(upsampled, Subject) return upsampled