Python torchvision.transforms.functional.vflip() Examples

The following are 26 code examples of torchvision.transforms.functional.vflip(). You can vote up the ones you like or vote down the ones you don't like, and go to the original project or source file by following the links above each example. You may also want to check out all available functions/classes of the module torchvision.transforms.functional , or try the search function .
Example #1
Source File: cvfunctional.py    From opencv_transforms_torchvision with MIT License 6 votes vote down vote up
def cv_transform(img):
    # img = resize(img, size=(100, 300))
    # img = to_tensor(img)
    # img = normalize(img, mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
    # img = pad(img, padding=(10, 10, 20, 20), fill=(255, 255, 255), padding_mode='constant')
    # img = pad(img, padding=(100, 100, 100, 100), fill=5, padding_mode='symmetric')
    # img = crop(img, -40, -20, 1000, 1000)
    # img = center_crop(img, (310, 300))
    # img = resized_crop(img, -10.3, -20, 330, 220, (500, 500))
    # img = hflip(img)
    # img = vflip(img)
    # tl, tr, bl, br, center = five_crop(img, 100)
    # img = adjust_brightness(img, 2.1)
    # img = adjust_contrast(img, 1.5)
    # img = adjust_saturation(img, 2.3)
    # img = adjust_hue(img, 0.5)
    # img = adjust_gamma(img, gamma=3, gain=0.1)
    # img = rotate(img, 10, resample='BILINEAR', expand=True, center=None)
    # img = to_grayscale(img, 3)
    # img = affine(img, 10, (0, 0), 1, 0, resample='BICUBIC', fillcolor=(255,255,0))
    # img = gaussion_noise(img)
    # img = poisson_noise(img)
    img = salt_and_pepper(img)
    return to_tensor(img) 
Example #2
Source File: cvfunctional.py    From opencv_transforms_torchvision with MIT License 6 votes vote down vote up
def pil_transform(img):
    # img = functional.resize(img, size=(100, 300))
    # img = functional.to_tensor(img)
    # img = functional.normalize(img, mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
    # img = functional.pad(img, padding=(10, 10, 20, 20), fill=(255, 255, 255), padding_mode='constant')
    # img = functional.pad(img, padding=(100, 100, 100, 100), padding_mode='symmetric')
    # img = functional.crop(img, -40, -20, 1000, 1000)
    # img = functional.center_crop(img, (310, 300))
    # img = functional.resized_crop(img, -10.3, -20, 330, 220, (500, 500))
    # img = functional.hflip(img)
    # img = functional.vflip(img)
    # tl, tr, bl, br, center = functional.five_crop(img, 100)
    # img = functional.adjust_brightness(img, 2.1)
    # img = functional.adjust_contrast(img, 1.5)
    # img = functional.adjust_saturation(img, 2.3)
    # img = functional.adjust_hue(img, 0.5)
    # img = functional.adjust_gamma(img, gamma=3, gain=0.1)
    # img = functional.rotate(img, 10, resample=PIL.Image.BILINEAR, expand=True, center=None)
    # img = functional.to_grayscale(img, 3)
    # img = functional.affine(img, 10, (0, 0), 1, 0, resample=PIL.Image.BICUBIC, fillcolor=(255,255,0))

    return functional.to_tensor(img) 
Example #3
Source File: transforms.py    From RCRNet-Pytorch with MIT License 6 votes vote down vote up
def __call__(self, sample):
        image, label = sample['image'], sample['label']
        if self.rand_flip_index is None or self.image_mode:
            self.rand_flip_index = random.randint(-1,2)
        # 0: horizontal flip, 1: vertical flip, -1: horizontal and vertical flip
        if self.rand_flip_index == 0:
            image = F.hflip(image)
            label = F.hflip(label)
        elif self.rand_flip_index == 1:
            image = F.vflip(image)
            label = F.vflip(label)
        elif self.rand_flip_index == 2:
            image = F.vflip(F.hflip(image))
            label = F.vflip(F.hflip(label))
        sample['image'], sample['label'] = image, label
        return sample 
Example #4
Source File: transforms.py    From ACDRNet with Apache License 2.0 5 votes vote down vote up
def __call__(self, img, mask):
        if random.random() < self.p:
            return F.vflip(img), F.vflip(mask)
        return img 
Example #5
Source File: imagepreprocess.py    From fast-MPN-COV with MIT License 5 votes vote down vote up
def center_crop_with_flip(img, size, vertical_flip=False):
    crop_h, crop_w = size
    first_crop = F.center_crop(img, (crop_h, crop_w))
    if vertical_flip:
        img = F.vflip(img)
    else:
         img = F.hflip(img)
    second_crop = F.center_crop(img, (crop_h, crop_w))
    return (first_crop, second_crop) 
Example #6
Source File: benchmark.py    From albumentations with MIT License 5 votes vote down vote up
def torchvision_transform(self, img):
        return torchvision.vflip(img) 
Example #7
Source File: benchmark.py    From albumentations with MIT License 5 votes vote down vote up
def albumentations(self, img):
        return albumentations.vflip(img) 
Example #8
Source File: ext_transforms.py    From DeepLabV3Plus-Pytorch with MIT License 5 votes vote down vote up
def __call__(self, img, lbl):
        """
        Args:
            img (PIL Image): Image to be flipped.
            lbl (PIL Image): Label to be flipped.
        Returns:
            PIL Image: Randomly flipped image.
            PIL Image: Randomly flipped label.
        """
        if random.random() < self.p:
            return F.vflip(img), F.vflip(lbl)
        return img, lbl 
Example #9
Source File: test_loader.py    From mobile-hair-segmentation-pytorch with MIT License 5 votes vote down vote up
def transform(image, mask, image_size=224):
    resize = transforms.Resize(size=(image_size, image_size))
    image = resize(image)
    mask = resize(mask)

    if random() > 0.5:
        image = TF.vflip(image)
        mask = TF.vflip(mask)

    if random() > 0.5:
        image = TF.hflip(image)
        mask = TF.hflip(mask)

    angle = random() * 12 - 6
    image = TF.rotate(image, angle)
    mask = TF.rotate(mask, angle)

    pad_size = random() * image_size
    image = TF.pad(image, pad_size, padding_mode='edge')
    mask = TF.pad(mask, pad_size, padding_mode='edge')

    # Transform to tensor
    image = TF.to_tensor(image)
    mask = TF.to_tensor(mask)

    # Normalize Data
    image = TF.normalize(image, (0.5, 0.5, 0.5), (0.5, 0.5, 0.5))

    return image, mask 
Example #10
Source File: video_transforms.py    From pvse with MIT License 5 votes vote down vote up
def __call__(self, img):
        """
        Args:
            img (PIL Image): Image to be flipped.

        Returns:
            PIL Image: Randomly flipped image.
        """
        if random.random() < self.p:
            return F.vflip(img)
        return img 
Example #11
Source File: augmentations.py    From pytorch-meta with MIT License 5 votes vote down vote up
def __call__(self, image):
        return F.vflip(image) 
Example #12
Source File: boundingBoxes.py    From aerial_wildlife_detection with MIT License 5 votes vote down vote up
def _verticalFlip(img, bboxes=None, labels=None):
    img = F.vflip(img)
    if bboxes is not None and len(labels):
        bboxes[:,1] = img.size[1] - (bboxes[:,1] + bboxes[:,3])
    return img, bboxes, labels 
Example #13
Source File: points.py    From aerial_wildlife_detection with MIT License 5 votes vote down vote up
def _verticalFlip(img, points=None, labels=None):
    img = F.vflip(img)
    if points is not None and len(labels):
        points[:,1] = img.size[1] - points[:,1]
    return img, points, labels 
Example #14
Source File: augmentation.py    From nni with MIT License 5 votes vote down vote up
def __call__(self, img):
        trans = {
            0: lambda x: x,
            1: lambda x: F.hflip(x),
            2: lambda x: F.vflip(x),
            3: lambda x: F.vflip(F.hflip(x)),
            4: lambda x: F.rotate(x, 90, False, False),
            5: lambda x: F.hflip(F.rotate(x, 90, False, False)),
            6: lambda x: F.vflip(F.rotate(x, 90, False, False)),
            7: lambda x: F.vflip(F.hflip(F.rotate(x, 90, False, False)))
        }
        return trans[self.index](img)

# i is tta index, 0: no change, 1: horizon flip, 2: vertical flip, 3: do both 
Example #15
Source File: augmentation.py    From nni with MIT License 5 votes vote down vote up
def __call__(self, *imgs):
        if random.random() < self.p:
            return map(F.vflip, imgs)
        else:
            return imgs 
Example #16
Source File: functional.py    From Jacinle with MIT License 5 votes vote down vote up
def vflip(img, coor):
    coor = coor.copy()
    coor[:, 1] = 1 - coor[:, 1]
    return TF.vflip(img), coor 
Example #17
Source File: functional.py    From Jacinle with MIT License 5 votes vote down vote up
def vflip(img, bbox):
    bbox = bbox.copy()
    bbox[:, 1] = 1 - bbox[:, 1]
    bbox[:, 3] = 1 - bbox[:, 3]
    return TF.vflip(img), bbox 
Example #18
Source File: transforms.py    From maskrcnn-benchmark with MIT License 5 votes vote down vote up
def __call__(self, image, target):
        if random.random() < self.prob:
            image = F.vflip(image)
            target = target.transpose(1)
        return image, target 
Example #19
Source File: transforms.py    From sampling-free with MIT License 5 votes vote down vote up
def __call__(self, image, target):
        if random.random() < self.prob:
            image = F.vflip(image)
            target = target.transpose(1)
        return image, target 
Example #20
Source File: cvfunctional.py    From opencv_transforms_torchvision with MIT License 5 votes vote down vote up
def ten_crop(img, size, vertical_flip=False):
    """Crop the given CV Image into four corners and the central crop plus the
       flipped version of these (horizontal flipping is used by default).

    .. Note::
        This transform returns a tuple of images and there may be a
        mismatch in the number of inputs and targets your ``Dataset`` returns.

       Args:
           size (sequence or int): Desired output size of the crop. If size is an
               int instead of sequence like (h, w), a square crop (size, size) is
               made.
           vertical_flip (bool): Use vertical flipping instead of horizontal

        Returns:
            tuple: tuple (tl, tr, bl, br, center, tl_flip, tr_flip, bl_flip,
                br_flip, center_flip) corresponding top left, top right,
                bottom left, bottom right and center crop and same for the
                flipped image.
    """
    if isinstance(size, numbers.Number):
        size = (int(size), int(size))
    else:
        assert len(size) == 2, "Please provide only two dimensions (h, w) for size."

    first_five = five_crop(img, size)

    if vertical_flip:
        img = vflip(img)
    else:
        img = hflip(img)

    second_five = five_crop(img, size)
    return first_five + second_five 
Example #21
Source File: cvfunctional.py    From opencv_transforms_torchvision with MIT License 5 votes vote down vote up
def vflip(img):
    """Vertically flip the given PIL Image.

    Args:
        img (CV Image): Image to be flipped.

    Returns:
        PIL Image:  Vertically flipped image.
    """
    if not _is_numpy_image(img):
        raise TypeError('img should be PIL Image. Got {}'.format(type(img)))

    return cv2.flip(img, 0) 
Example #22
Source File: __init__.py    From Pytorch_Lightweight_Network with MIT License 5 votes vote down vote up
def __call__(self, img, anns):
        if random.random() < self.p:
            img = VF.vflip(img)
            anns = HF.vflip(anns, img.size)
            return img, anns
        return img, anns 
Example #23
Source File: source_target_transforms.py    From pytorch-zssr with Apache License 2.0 5 votes vote down vote up
def __call__(self, data):
        """
        Args:
            img (PIL Image): Image to be flipped.
        Returns:
            PIL Image: Randomly flipped image.
        """
        hr, lr = data
        if random.random() < 0.5:
            return F.vflip(hr), F.vflip(lr)
        return hr, lr 
Example #24
Source File: transforms.py    From DetNAS with MIT License 5 votes vote down vote up
def __call__(self, image, target):
        if random.random() < self.prob:
            image = F.vflip(image)
            target = target.transpose(1)
        return image, target 
Example #25
Source File: transforms.py    From Clothing-Detection with GNU General Public License v3.0 5 votes vote down vote up
def __call__(self, image, target):
        if random.random() < self.prob:
            image = F.vflip(image)
            target = target.transpose(1)
        return image, target 
Example #26
Source File: dataloader.py    From mobile-hair-segmentation-pytorch with MIT License 4 votes vote down vote up
def transform(image, mask, image_size=224):
    # Resize
    resized_num = int(random.random() * image_size)
    resize = transforms.Resize(size=(image_size + resized_num, image_size + resized_num))
    image = resize(image)
    mask = resize(mask)

    # num_pad = int(random.random() * image_size)
    # image = TF.pad(image, num_pad, padding_mode='edge')
    # mask = TF.pad(mask, num_pad)

    # # Random crop
    # i, j, h, w = transforms.RandomCrop.get_params(
    #     image, output_size=(image_size, image_size))
    # image = TF.crop(image, i, j, h, w)
    # mask = TF.crop(mask, i, j, h, w)


    # # Random horizontal flipping
    # if random.random() > 0.5:
    #     image = TF.hflip(image)
    #     mask = TF.hflip(mask)
    #
    # # Random vertical flipping
    # if random.random() > 0.5:
    #     image = TF.vflip(image)
    #     mask = TF.vflip(mask)

    resize = transforms.Resize(size=(image_size, image_size))
    image = resize(image)
    mask = resize(mask)

    # Make gray scale image
    gray_image = TF.to_grayscale(image)

    # Transform to tensor
    image = TF.to_tensor(image)
    mask = TF.to_tensor(mask)
    gray_image = TF.to_tensor(gray_image)

    # Normalize Data
    image = TF.normalize(image, (0.5, 0.5, 0.5), (0.5, 0.5, 0.5))

    return image, gray_image, mask