Python torchvision.transforms.functional.vflip() Examples
The following are 26
code examples of torchvision.transforms.functional.vflip().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
torchvision.transforms.functional
, or try the search function
.
Example #1
Source File: cvfunctional.py From opencv_transforms_torchvision with MIT License | 6 votes |
def cv_transform(img): # img = resize(img, size=(100, 300)) # img = to_tensor(img) # img = normalize(img, mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) # img = pad(img, padding=(10, 10, 20, 20), fill=(255, 255, 255), padding_mode='constant') # img = pad(img, padding=(100, 100, 100, 100), fill=5, padding_mode='symmetric') # img = crop(img, -40, -20, 1000, 1000) # img = center_crop(img, (310, 300)) # img = resized_crop(img, -10.3, -20, 330, 220, (500, 500)) # img = hflip(img) # img = vflip(img) # tl, tr, bl, br, center = five_crop(img, 100) # img = adjust_brightness(img, 2.1) # img = adjust_contrast(img, 1.5) # img = adjust_saturation(img, 2.3) # img = adjust_hue(img, 0.5) # img = adjust_gamma(img, gamma=3, gain=0.1) # img = rotate(img, 10, resample='BILINEAR', expand=True, center=None) # img = to_grayscale(img, 3) # img = affine(img, 10, (0, 0), 1, 0, resample='BICUBIC', fillcolor=(255,255,0)) # img = gaussion_noise(img) # img = poisson_noise(img) img = salt_and_pepper(img) return to_tensor(img)
Example #2
Source File: cvfunctional.py From opencv_transforms_torchvision with MIT License | 6 votes |
def pil_transform(img): # img = functional.resize(img, size=(100, 300)) # img = functional.to_tensor(img) # img = functional.normalize(img, mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) # img = functional.pad(img, padding=(10, 10, 20, 20), fill=(255, 255, 255), padding_mode='constant') # img = functional.pad(img, padding=(100, 100, 100, 100), padding_mode='symmetric') # img = functional.crop(img, -40, -20, 1000, 1000) # img = functional.center_crop(img, (310, 300)) # img = functional.resized_crop(img, -10.3, -20, 330, 220, (500, 500)) # img = functional.hflip(img) # img = functional.vflip(img) # tl, tr, bl, br, center = functional.five_crop(img, 100) # img = functional.adjust_brightness(img, 2.1) # img = functional.adjust_contrast(img, 1.5) # img = functional.adjust_saturation(img, 2.3) # img = functional.adjust_hue(img, 0.5) # img = functional.adjust_gamma(img, gamma=3, gain=0.1) # img = functional.rotate(img, 10, resample=PIL.Image.BILINEAR, expand=True, center=None) # img = functional.to_grayscale(img, 3) # img = functional.affine(img, 10, (0, 0), 1, 0, resample=PIL.Image.BICUBIC, fillcolor=(255,255,0)) return functional.to_tensor(img)
Example #3
Source File: transforms.py From RCRNet-Pytorch with MIT License | 6 votes |
def __call__(self, sample): image, label = sample['image'], sample['label'] if self.rand_flip_index is None or self.image_mode: self.rand_flip_index = random.randint(-1,2) # 0: horizontal flip, 1: vertical flip, -1: horizontal and vertical flip if self.rand_flip_index == 0: image = F.hflip(image) label = F.hflip(label) elif self.rand_flip_index == 1: image = F.vflip(image) label = F.vflip(label) elif self.rand_flip_index == 2: image = F.vflip(F.hflip(image)) label = F.vflip(F.hflip(label)) sample['image'], sample['label'] = image, label return sample
Example #4
Source File: transforms.py From ACDRNet with Apache License 2.0 | 5 votes |
def __call__(self, img, mask): if random.random() < self.p: return F.vflip(img), F.vflip(mask) return img
Example #5
Source File: imagepreprocess.py From fast-MPN-COV with MIT License | 5 votes |
def center_crop_with_flip(img, size, vertical_flip=False): crop_h, crop_w = size first_crop = F.center_crop(img, (crop_h, crop_w)) if vertical_flip: img = F.vflip(img) else: img = F.hflip(img) second_crop = F.center_crop(img, (crop_h, crop_w)) return (first_crop, second_crop)
Example #6
Source File: benchmark.py From albumentations with MIT License | 5 votes |
def torchvision_transform(self, img): return torchvision.vflip(img)
Example #7
Source File: benchmark.py From albumentations with MIT License | 5 votes |
def albumentations(self, img): return albumentations.vflip(img)
Example #8
Source File: ext_transforms.py From DeepLabV3Plus-Pytorch with MIT License | 5 votes |
def __call__(self, img, lbl): """ Args: img (PIL Image): Image to be flipped. lbl (PIL Image): Label to be flipped. Returns: PIL Image: Randomly flipped image. PIL Image: Randomly flipped label. """ if random.random() < self.p: return F.vflip(img), F.vflip(lbl) return img, lbl
Example #9
Source File: test_loader.py From mobile-hair-segmentation-pytorch with MIT License | 5 votes |
def transform(image, mask, image_size=224): resize = transforms.Resize(size=(image_size, image_size)) image = resize(image) mask = resize(mask) if random() > 0.5: image = TF.vflip(image) mask = TF.vflip(mask) if random() > 0.5: image = TF.hflip(image) mask = TF.hflip(mask) angle = random() * 12 - 6 image = TF.rotate(image, angle) mask = TF.rotate(mask, angle) pad_size = random() * image_size image = TF.pad(image, pad_size, padding_mode='edge') mask = TF.pad(mask, pad_size, padding_mode='edge') # Transform to tensor image = TF.to_tensor(image) mask = TF.to_tensor(mask) # Normalize Data image = TF.normalize(image, (0.5, 0.5, 0.5), (0.5, 0.5, 0.5)) return image, mask
Example #10
Source File: video_transforms.py From pvse with MIT License | 5 votes |
def __call__(self, img): """ Args: img (PIL Image): Image to be flipped. Returns: PIL Image: Randomly flipped image. """ if random.random() < self.p: return F.vflip(img) return img
Example #11
Source File: augmentations.py From pytorch-meta with MIT License | 5 votes |
def __call__(self, image): return F.vflip(image)
Example #12
Source File: boundingBoxes.py From aerial_wildlife_detection with MIT License | 5 votes |
def _verticalFlip(img, bboxes=None, labels=None): img = F.vflip(img) if bboxes is not None and len(labels): bboxes[:,1] = img.size[1] - (bboxes[:,1] + bboxes[:,3]) return img, bboxes, labels
Example #13
Source File: points.py From aerial_wildlife_detection with MIT License | 5 votes |
def _verticalFlip(img, points=None, labels=None): img = F.vflip(img) if points is not None and len(labels): points[:,1] = img.size[1] - points[:,1] return img, points, labels
Example #14
Source File: augmentation.py From nni with MIT License | 5 votes |
def __call__(self, img): trans = { 0: lambda x: x, 1: lambda x: F.hflip(x), 2: lambda x: F.vflip(x), 3: lambda x: F.vflip(F.hflip(x)), 4: lambda x: F.rotate(x, 90, False, False), 5: lambda x: F.hflip(F.rotate(x, 90, False, False)), 6: lambda x: F.vflip(F.rotate(x, 90, False, False)), 7: lambda x: F.vflip(F.hflip(F.rotate(x, 90, False, False))) } return trans[self.index](img) # i is tta index, 0: no change, 1: horizon flip, 2: vertical flip, 3: do both
Example #15
Source File: augmentation.py From nni with MIT License | 5 votes |
def __call__(self, *imgs): if random.random() < self.p: return map(F.vflip, imgs) else: return imgs
Example #16
Source File: functional.py From Jacinle with MIT License | 5 votes |
def vflip(img, coor): coor = coor.copy() coor[:, 1] = 1 - coor[:, 1] return TF.vflip(img), coor
Example #17
Source File: functional.py From Jacinle with MIT License | 5 votes |
def vflip(img, bbox): bbox = bbox.copy() bbox[:, 1] = 1 - bbox[:, 1] bbox[:, 3] = 1 - bbox[:, 3] return TF.vflip(img), bbox
Example #18
Source File: transforms.py From maskrcnn-benchmark with MIT License | 5 votes |
def __call__(self, image, target): if random.random() < self.prob: image = F.vflip(image) target = target.transpose(1) return image, target
Example #19
Source File: transforms.py From sampling-free with MIT License | 5 votes |
def __call__(self, image, target): if random.random() < self.prob: image = F.vflip(image) target = target.transpose(1) return image, target
Example #20
Source File: cvfunctional.py From opencv_transforms_torchvision with MIT License | 5 votes |
def ten_crop(img, size, vertical_flip=False): """Crop the given CV Image into four corners and the central crop plus the flipped version of these (horizontal flipping is used by default). .. Note:: This transform returns a tuple of images and there may be a mismatch in the number of inputs and targets your ``Dataset`` returns. Args: size (sequence or int): Desired output size of the crop. If size is an int instead of sequence like (h, w), a square crop (size, size) is made. vertical_flip (bool): Use vertical flipping instead of horizontal Returns: tuple: tuple (tl, tr, bl, br, center, tl_flip, tr_flip, bl_flip, br_flip, center_flip) corresponding top left, top right, bottom left, bottom right and center crop and same for the flipped image. """ if isinstance(size, numbers.Number): size = (int(size), int(size)) else: assert len(size) == 2, "Please provide only two dimensions (h, w) for size." first_five = five_crop(img, size) if vertical_flip: img = vflip(img) else: img = hflip(img) second_five = five_crop(img, size) return first_five + second_five
Example #21
Source File: cvfunctional.py From opencv_transforms_torchvision with MIT License | 5 votes |
def vflip(img): """Vertically flip the given PIL Image. Args: img (CV Image): Image to be flipped. Returns: PIL Image: Vertically flipped image. """ if not _is_numpy_image(img): raise TypeError('img should be PIL Image. Got {}'.format(type(img))) return cv2.flip(img, 0)
Example #22
Source File: __init__.py From Pytorch_Lightweight_Network with MIT License | 5 votes |
def __call__(self, img, anns): if random.random() < self.p: img = VF.vflip(img) anns = HF.vflip(anns, img.size) return img, anns return img, anns
Example #23
Source File: source_target_transforms.py From pytorch-zssr with Apache License 2.0 | 5 votes |
def __call__(self, data): """ Args: img (PIL Image): Image to be flipped. Returns: PIL Image: Randomly flipped image. """ hr, lr = data if random.random() < 0.5: return F.vflip(hr), F.vflip(lr) return hr, lr
Example #24
Source File: transforms.py From DetNAS with MIT License | 5 votes |
def __call__(self, image, target): if random.random() < self.prob: image = F.vflip(image) target = target.transpose(1) return image, target
Example #25
Source File: transforms.py From Clothing-Detection with GNU General Public License v3.0 | 5 votes |
def __call__(self, image, target): if random.random() < self.prob: image = F.vflip(image) target = target.transpose(1) return image, target
Example #26
Source File: dataloader.py From mobile-hair-segmentation-pytorch with MIT License | 4 votes |
def transform(image, mask, image_size=224): # Resize resized_num = int(random.random() * image_size) resize = transforms.Resize(size=(image_size + resized_num, image_size + resized_num)) image = resize(image) mask = resize(mask) # num_pad = int(random.random() * image_size) # image = TF.pad(image, num_pad, padding_mode='edge') # mask = TF.pad(mask, num_pad) # # Random crop # i, j, h, w = transforms.RandomCrop.get_params( # image, output_size=(image_size, image_size)) # image = TF.crop(image, i, j, h, w) # mask = TF.crop(mask, i, j, h, w) # # Random horizontal flipping # if random.random() > 0.5: # image = TF.hflip(image) # mask = TF.hflip(mask) # # # Random vertical flipping # if random.random() > 0.5: # image = TF.vflip(image) # mask = TF.vflip(mask) resize = transforms.Resize(size=(image_size, image_size)) image = resize(image) mask = resize(mask) # Make gray scale image gray_image = TF.to_grayscale(image) # Transform to tensor image = TF.to_tensor(image) mask = TF.to_tensor(mask) gray_image = TF.to_tensor(gray_image) # Normalize Data image = TF.normalize(image, (0.5, 0.5, 0.5), (0.5, 0.5, 0.5)) return image, gray_image, mask