Python cv2.CV_32F Examples
The following are 30
code examples of cv2.CV_32F().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
cv2
, or try the search function
.
Example #1
Source File: predict.py From License-Plate-Recognition with MIT License | 6 votes |
def preprocess_hog(digits): samples = [] for img in digits: gx = cv2.Sobel(img, cv2.CV_32F, 1, 0) gy = cv2.Sobel(img, cv2.CV_32F, 0, 1) mag, ang = cv2.cartToPolar(gx, gy) bin_n = 16 bin = np.int32(bin_n*ang/(2*np.pi)) bin_cells = bin[:10,:10], bin[10:,:10], bin[:10,10:], bin[10:,10:] mag_cells = mag[:10,:10], mag[10:,:10], mag[:10,10:], mag[10:,10:] hists = [np.bincount(b.ravel(), m.ravel(), bin_n) for b, m in zip(bin_cells, mag_cells)] hist = np.hstack(hists) # transform to Hellinger kernel eps = 1e-7 hist /= hist.sum() + eps hist = np.sqrt(hist) hist /= norm(hist) + eps samples.append(hist) return np.float32(samples) #不能保证包括所有省份
Example #2
Source File: spfunctions.py From spfeas with MIT License | 6 votes |
def get_mag_avg(img): img = np.sqrt(img) kernels = get_kernels() mag = np.zeros(img.shape, dtype='float32') for kernel_filter in kernels: gx = cv2.filter2D(np.float32(img), cv2.CV_32F, kernel_filter[1], borderType=cv2.BORDER_REFLECT) gy = cv2.filter2D(np.float32(img), cv2.CV_32F, kernel_filter[0], borderType=cv2.BORDER_REFLECT) mag += cv2.magnitude(gx, gy) mag /= len(kernels) return np.uint8(mag)
Example #3
Source File: spfunctions.py From spfeas with MIT License | 6 votes |
def get_mag_ang(img): """ Gets image gradient (magnitude) and orientation (angle) Args: img Returns: Gradient, orientation """ img = np.sqrt(img) gx = cv2.Sobel(np.float32(img), cv2.CV_32F, 1, 0) gy = cv2.Sobel(np.float32(img), cv2.CV_32F, 0, 1) mag, ang = cv2.cartToPolar(gx, gy) return mag, ang, gx, gy
Example #4
Source File: vector_field.py From Pointillism with MIT License | 6 votes |
def from_gradient(gray): fieldx = cv2.Scharr(gray, cv2.CV_32F, 1, 0) / 15.36 fieldy = cv2.Scharr(gray, cv2.CV_32F, 0, 1) / 15.36 return VectorField(fieldx, fieldy)
Example #5
Source File: omnirobot_simulator_server.py From robotics-rl-srl with MIT License | 6 votes |
def renderEnvLuminosityNoise(self, origin_image, noise_var=0.1, in_RGB=False, out_RGB=False): """ render the different environment luminosity """ # variate luminosity and color origin_image_LAB = cv2.cvtColor( origin_image, cv2.COLOR_RGB2LAB if in_RGB else cv2.COLOR_BGR2LAB, cv2.CV_32F) origin_image_LAB[:, :, 0] = origin_image_LAB[:, :, 0] * (np.random.randn() * noise_var + 1.0) origin_image_LAB[:, :, 1] = origin_image_LAB[:, :, 1] * (np.random.randn() * noise_var + 1.0) origin_image_LAB[:, :, 2] = origin_image_LAB[:, :, 2] * (np.random.randn() * noise_var + 1.0) out_image = cv2.cvtColor( origin_image_LAB, cv2.COLOR_LAB2RGB if out_RGB else cv2.COLOR_LAB2BGR, cv2.CV_8UC3) return out_image
Example #6
Source File: svm_train.py From vehicle-license-plate-recognition with MIT License | 6 votes |
def preprocess_hog(digits): samples = [] for img in digits: gx = cv2.Sobel(img, cv2.CV_32F, 1, 0) gy = cv2.Sobel(img, cv2.CV_32F, 0, 1) mag, ang = cv2.cartToPolar(gx, gy) bin_n = 16 bin = np.int32(bin_n*ang/(2*np.pi)) bin_cells = bin[:10,:10], bin[10:,:10], bin[:10,10:], bin[10:,10:] mag_cells = mag[:10,:10], mag[10:,:10], mag[:10,10:], mag[10:,10:] hists = [np.bincount(b.ravel(), m.ravel(), bin_n) for b, m in zip(bin_cells, mag_cells)] hist = np.hstack(hists) # transform to Hellinger kernel eps = 1e-7 hist /= hist.sum() + eps hist = np.sqrt(hist) hist /= norm(hist) + eps samples.append(hist) return np.float32(samples)
Example #7
Source File: agent.py From DRLwithTL with MIT License | 6 votes |
def get_state(self): responses1 = self.client.simGetImages([ # depth visualization image airsim.ImageRequest("1", airsim.ImageType.Scene, False, False)]) # scene vision image in uncompressed RGBA array response = responses1[0] img1d = np.fromstring(response.image_data_uint8, dtype=np.uint8) # get numpy array img_rgba = img1d.reshape(response.height, response.width, 3) img = Image.fromarray(img_rgba) img_rgb = img.convert('RGB') self.iter = self.iter+1 state = np.asarray(img_rgb) state = cv2.resize(state, (self.input_size, self.input_size), cv2.INTER_LINEAR) state = cv2.normalize(state, state, 0, 1, cv2.NORM_MINMAX, cv2.CV_32F) state_rgb = [] state_rgb.append(state[:, :, 0:3]) state_rgb = np.array(state_rgb) state_rgb = state_rgb.astype('float32') return state_rgb
Example #8
Source File: utils.py From answer-sheet-scan with MIT License | 6 votes |
def get_init_process_img(roi_img): """ 对图片进行初始化处理,包括,梯度化,高斯模糊,二值化,腐蚀,膨胀和边缘检测 :param roi_img: ndarray :return: ndarray """ h = cv2.Sobel(roi_img, cv2.CV_32F, 0, 1, -1) v = cv2.Sobel(roi_img, cv2.CV_32F, 1, 0, -1) img = cv2.add(h, v) img = cv2.convertScaleAbs(img) img = cv2.GaussianBlur(img, (3, 3), 0) ret, img = cv2.threshold(img, 120, 255, cv2.THRESH_BINARY) kernel = np.ones((1, 1), np.uint8) img = cv2.erode(img, kernel, iterations=1) img = cv2.dilate(img, kernel, iterations=2) img = cv2.erode(img, kernel, iterations=1) img = cv2.dilate(img, kernel, iterations=2) img = auto_canny(img) return img
Example #9
Source File: file_function.py From gradcam.pytorch with MIT License | 6 votes |
def resize_and_contrast(in_dir, out_dir, target_size): check_and_mkdir(out_dir) clahe = cv2.createCLAHE(clipLimit=2.0, tileGridSize=(8,8)) for subdir, dirs, files in os.walk(in_dir): for f in files: file_path = subdir + os.sep + f if (is_image(f)): img = cv2.imread(file_path, 0) resized_img = cv2.resize(img, (target_size, target_size), interpolation = cv2.INTER_CUBIC) class_dir = out_dir + os.sep + file_path.split("/")[-2] check_and_mkdir(class_dir) file_name = class_dir + os.sep + file_path.split("/")[-1] print(file_name) norm_image = cv2.normalize(resized_img, alpha=0, beta=1, norm_type=cv2.NORM_MINMAX, dtype=cv2.CV_32F) * 256 # norm_image = clahe.apply(resized_img) cv2.imwrite(file_name, norm_image) # count the direct one-step sub directories (which will represent the class name)
Example #10
Source File: barcodeD&D_zbar.py From Barcode-Detection-and-Decoding with Apache License 2.0 | 6 votes |
def preprocess(image): # load the image image = cv2.imread(args["image"]) #resize image image = cv2.resize(image,None,fx=0.7, fy=0.7, interpolation = cv2.INTER_CUBIC) #convert to grayscale gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) #calculate x & y gradient gradX = cv2.Sobel(gray, ddepth = cv2.CV_32F, dx = 1, dy = 0, ksize = -1) gradY = cv2.Sobel(gray, ddepth = cv2.CV_32F, dx = 0, dy = 1, ksize = -1) # subtract the y-gradient from the x-gradient gradient = cv2.subtract(gradX, gradY) gradient = cv2.convertScaleAbs(gradient) # blur the image blurred = cv2.blur(gradient, (3, 3)) # threshold the image (_, thresh) = cv2.threshold(blurred, 225, 255, cv2.THRESH_BINARY) thresh = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) return thresh
Example #11
Source File: digits.py From PyCV-time with MIT License | 6 votes |
def preprocess_hog(digits): samples = [] for img in digits: gx = cv2.Sobel(img, cv2.CV_32F, 1, 0) gy = cv2.Sobel(img, cv2.CV_32F, 0, 1) mag, ang = cv2.cartToPolar(gx, gy) bin_n = 16 bin = np.int32(bin_n*ang/(2*np.pi)) bin_cells = bin[:10,:10], bin[10:,:10], bin[:10,10:], bin[10:,10:] mag_cells = mag[:10,:10], mag[10:,:10], mag[:10,10:], mag[10:,10:] hists = [np.bincount(b.ravel(), m.ravel(), bin_n) for b, m in zip(bin_cells, mag_cells)] hist = np.hstack(hists) # transform to Hellinger kernel eps = 1e-7 hist /= hist.sum() + eps hist = np.sqrt(hist) hist /= norm(hist) + eps samples.append(hist) return np.float32(samples)
Example #12
Source File: coherence.py From PyCV-time with MIT License | 6 votes |
def coherence_filter(img, sigma = 11, str_sigma = 11, blend = 0.5, iter_n = 4): h, w = img.shape[:2] for i in xrange(iter_n): print i, gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) eigen = cv2.cornerEigenValsAndVecs(gray, str_sigma, 3) eigen = eigen.reshape(h, w, 3, 2) # [[e1, e2], v1, v2] x, y = eigen[:,:,1,0], eigen[:,:,1,1] gxx = cv2.Sobel(gray, cv2.CV_32F, 2, 0, ksize=sigma) gxy = cv2.Sobel(gray, cv2.CV_32F, 1, 1, ksize=sigma) gyy = cv2.Sobel(gray, cv2.CV_32F, 0, 2, ksize=sigma) gvv = x*x*gxx + 2*x*y*gxy + y*y*gyy m = gvv < 0 ero = cv2.erode(img, None) dil = cv2.dilate(img, None) img1 = ero img1[m] = dil[m] img = np.uint8(img*(1.0 - blend) + img1*blend) print 'done' return img
Example #13
Source File: mosse.py From PyCV-time with MIT License | 6 votes |
def __init__(self, frame, rect): x1, y1, x2, y2 = rect w, h = map(cv2.getOptimalDFTSize, [x2-x1, y2-y1]) x1, y1 = (x1+x2-w)//2, (y1+y2-h)//2 self.pos = x, y = x1+0.5*(w-1), y1+0.5*(h-1) self.size = w, h img = cv2.getRectSubPix(frame, (w, h), (x, y)) self.win = cv2.createHanningWindow((w, h), cv2.CV_32F) g = np.zeros((h, w), np.float32) g[h//2, w//2] = 1 g = cv2.GaussianBlur(g, (-1, -1), 2.0) g /= g.max() self.G = cv2.dft(g, flags=cv2.DFT_COMPLEX_OUTPUT) self.H1 = np.zeros_like(self.G) self.H2 = np.zeros_like(self.G) for i in xrange(128): a = self.preprocess(rnd_warp(img)) A = cv2.dft(a, flags=cv2.DFT_COMPLEX_OUTPUT) self.H1 += cv2.mulSpectrums(self.G, A, 0, conjB=True) self.H2 += cv2.mulSpectrums( A, A, 0, conjB=True) self.update_kernel() self.update(frame)
Example #14
Source File: mosse.py From OpenCV-Python-Tutorial with MIT License | 6 votes |
def __init__(self, frame, rect): x1, y1, x2, y2 = rect w, h = map(cv2.getOptimalDFTSize, [x2-x1, y2-y1]) x1, y1 = (x1+x2-w)//2, (y1+y2-h)//2 self.pos = x, y = x1+0.5*(w-1), y1+0.5*(h-1) self.size = w, h img = cv2.getRectSubPix(frame, (w, h), (x, y)) self.win = cv2.createHanningWindow((w, h), cv2.CV_32F) g = np.zeros((h, w), np.float32) g[h//2, w//2] = 1 g = cv2.GaussianBlur(g, (-1, -1), 2.0) g /= g.max() self.G = cv2.dft(g, flags=cv2.DFT_COMPLEX_OUTPUT) self.H1 = np.zeros_like(self.G) self.H2 = np.zeros_like(self.G) for i in xrange(128): a = self.preprocess(rnd_warp(img)) A = cv2.dft(a, flags=cv2.DFT_COMPLEX_OUTPUT) self.H1 += cv2.mulSpectrums(self.G, A, 0, conjB=True) self.H2 += cv2.mulSpectrums( A, A, 0, conjB=True) self.update_kernel() self.update(frame)
Example #15
Source File: coherence.py From OpenCV-Python-Tutorial with MIT License | 6 votes |
def coherence_filter(img, sigma = 11, str_sigma = 11, blend = 0.5, iter_n = 4): h, w = img.shape[:2] for i in xrange(iter_n): print(i) gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) eigen = cv2.cornerEigenValsAndVecs(gray, str_sigma, 3) eigen = eigen.reshape(h, w, 3, 2) # [[e1, e2], v1, v2] x, y = eigen[:,:,1,0], eigen[:,:,1,1] gxx = cv2.Sobel(gray, cv2.CV_32F, 2, 0, ksize=sigma) gxy = cv2.Sobel(gray, cv2.CV_32F, 1, 1, ksize=sigma) gyy = cv2.Sobel(gray, cv2.CV_32F, 0, 2, ksize=sigma) gvv = x*x*gxx + 2*x*y*gxy + y*y*gyy m = gvv < 0 ero = cv2.erode(img, None) dil = cv2.dilate(img, None) img1 = ero img1[m] = dil[m] img = np.uint8(img*(1.0 - blend) + img1*blend) print('done') return img
Example #16
Source File: digits.py From OpenCV-Python-Tutorial with MIT License | 6 votes |
def preprocess_hog(digits): samples = [] for img in digits: gx = cv2.Sobel(img, cv2.CV_32F, 1, 0) gy = cv2.Sobel(img, cv2.CV_32F, 0, 1) mag, ang = cv2.cartToPolar(gx, gy) bin_n = 16 bin = np.int32(bin_n*ang/(2*np.pi)) bin_cells = bin[:10,:10], bin[10:,:10], bin[:10,10:], bin[10:,10:] mag_cells = mag[:10,:10], mag[10:,:10], mag[:10,10:], mag[10:,10:] hists = [np.bincount(b.ravel(), m.ravel(), bin_n) for b, m in zip(bin_cells, mag_cells)] hist = np.hstack(hists) # transform to Hellinger kernel eps = 1e-7 hist /= hist.sum() + eps hist = np.sqrt(hist) hist /= norm(hist) + eps samples.append(hist) return np.float32(samples)
Example #17
Source File: coherence.py From PyCV-time with MIT License | 6 votes |
def coherence_filter(img, sigma = 11, str_sigma = 11, blend = 0.5, iter_n = 4): h, w = img.shape[:2] for i in xrange(iter_n): print i, gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) eigen = cv2.cornerEigenValsAndVecs(gray, str_sigma, 3) eigen = eigen.reshape(h, w, 3, 2) # [[e1, e2], v1, v2] x, y = eigen[:,:,1,0], eigen[:,:,1,1] gxx = cv2.Sobel(gray, cv2.CV_32F, 2, 0, ksize=sigma) gxy = cv2.Sobel(gray, cv2.CV_32F, 1, 1, ksize=sigma) gyy = cv2.Sobel(gray, cv2.CV_32F, 0, 2, ksize=sigma) gvv = x*x*gxx + 2*x*y*gxy + y*y*gyy m = gvv < 0 ero = cv2.erode(img, None) dil = cv2.dilate(img, None) img1 = ero img1[m] = dil[m] img = np.uint8(img*(1.0 - blend) + img1*blend) print 'done' return img
Example #18
Source File: mosse.py From PyCV-time with MIT License | 6 votes |
def __init__(self, frame, rect): x1, y1, x2, y2 = rect w, h = map(cv2.getOptimalDFTSize, [x2-x1, y2-y1]) x1, y1 = (x1+x2-w)//2, (y1+y2-h)//2 self.pos = x, y = x1+0.5*(w-1), y1+0.5*(h-1) self.size = w, h img = cv2.getRectSubPix(frame, (w, h), (x, y)) self.win = cv2.createHanningWindow((w, h), cv2.CV_32F) g = np.zeros((h, w), np.float32) g[h//2, w//2] = 1 g = cv2.GaussianBlur(g, (-1, -1), 2.0) g /= g.max() self.G = cv2.dft(g, flags=cv2.DFT_COMPLEX_OUTPUT) self.H1 = np.zeros_like(self.G) self.H2 = np.zeros_like(self.G) for i in xrange(128): a = self.preprocess(rnd_warp(img)) A = cv2.dft(a, flags=cv2.DFT_COMPLEX_OUTPUT) self.H1 += cv2.mulSpectrums(self.G, A, 0, conjB=True) self.H2 += cv2.mulSpectrums( A, A, 0, conjB=True) self.update_kernel() self.update(frame)
Example #19
Source File: smoke_video_dataset_cp.py From deep-smoke-machine with BSD 3-Clause "New" or "Revised" License | 6 votes |
def load_frames(file_path, resize_to=224.0): # Saved numpy files should be read in with format (time, height, width, channel) frames = np.load(file_path) t, h, w, c = frames.shape # Resize and scale images for the network structure #TODO: maybe use opencv to normalize the image #frames = cv.normalize(frames, None, alpha=0, beta=1, norm_type=cv.NORM_MINMAX, dtype=cv.CV_32F) frames_out = [] need_resize = False if w < resize_to or h < resize_to: d = resize_to - min(w, h) sc = 1 + d / min(w, h) need_resize = True for i in range(t): img = frames[i, :, :, :] if need_resize: img = cv.resize(img, dsize=(0, 0), fx=sc, fy=sc) img = (img / 255.) * 2 - 1 frames_out.append(img) return np.asarray(frames_out, dtype=np.float32)
Example #20
Source File: gabor_threads.py From PyCV-time with MIT License | 5 votes |
def build_filters(): filters = [] ksize = 31 for theta in np.arange(0, np.pi, np.pi / 16): kern = cv2.getGaborKernel((ksize, ksize), 4.0, theta, 10.0, 0.5, 0, ktype=cv2.CV_32F) kern /= 1.5*kern.sum() filters.append(kern) return filters
Example #21
Source File: eigenfaces.py From python-examples-cv with GNU Lesser General Public License v3.0 | 5 votes |
def performPCA(images): # Allocate space for all images in one data matrix. The size of the data matrix is # ( w * h * c, numImages ) where, w = width of an image in the dataset. # h = height of an image in the dataset. c is for the number of color # channels. numImages = len(images) sz = images[0].shape channels = 1 # grayescale data = np.zeros((numImages, sz[0] * sz[1] * channels), dtype=np.float32) # store images as floating point vectors normalized 0 -> 1 for i in range(0, numImages): image = np.float32(images[i]) / 255.0 data[i, :] = image.flatten() # N.B. data is stored as rows # compute the eigenvectors from the stack of image vectors created mean, eigenVectors = cv2.PCACompute( data, mean=None, maxComponents=args.eigenfaces) # use the eigenvectors to project the set of images to the new PCA space # representation coefficients = cv2.PCAProject(data, mean, eigenVectors) # calculate the covariance and mean of the PCA space representation of the images # (skipping the first N eigenfaces that often contain just illumination variance, default N=3 ) covariance_coeffs, mean_coeffs = cv2.calcCovarMatrix( coefficients[:, args.eigenfaces_to_skip:args.eigenfaces], mean=None, flags=cv2.COVAR_NORMAL | cv2.COVAR_ROWS, ctype=cv2.CV_32F) return (mean, eigenVectors, coefficients, mean_coeffs, covariance_coeffs) ########################################################################## # return index of best matching face from set of all PCA projcted coefficients # based on miniumum Mahalanobis (M) distance and this minimum M distance
Example #22
Source File: gabor_threads.py From PyCV-time with MIT License | 5 votes |
def build_filters(): filters = [] ksize = 31 for theta in np.arange(0, np.pi, np.pi / 16): kern = cv2.getGaborKernel((ksize, ksize), 4.0, theta, 10.0, 0.5, 0, ktype=cv2.CV_32F) kern /= 1.5*kern.sum() filters.append(kern) return filters
Example #23
Source File: dataset_utils.py From EVDodgeNet with BSD 3-Clause "New" or "Revised" License | 5 votes |
def compute_gradient(img, use_scharr=True): if use_scharr: norm_factor = 32 gradx = cv2.Scharr(img, cv2.CV_32F, 1, 0, scale=1.0/norm_factor) grady = cv2.Scharr(img, cv2.CV_32F, 0, 1, scale=1.0/norm_factor) else: kx = cv2.getDerivKernels(1, 0, ksize=1, normalize=True) ky = cv2.getDerivKernels(0, 1, ksize=1, normalize=True) gradx = cv2.sepFilter2D(img, cv2.CV_32F, kx[0], kx[1]) grady = cv2.sepFilter2D(img, cv2.CV_32F, ky[0], ky[1]) gradient = np.dstack([gradx, grady]) return gradient
Example #24
Source File: distance_transform.py From plantcv with MIT License | 5 votes |
def distance_transform(bin_img, distance_type, mask_size): """Creates an image where for each object pixel, a number is assigned that corresponds to the distance to the nearest background pixel. Inputs: img = Binary image data distance_type = Type of distance. It can be CV_DIST_L1, CV_DIST_L2 , or CV_DIST_C which are 1, 2 and 3, respectively. mask_size = Size of the distance transform mask. It can be 3, 5, or CV_DIST_MASK_PRECISE (the latter option is only supported by the first function). In case of the CV_DIST_L1 or CV_DIST_C distance type, the parameter is forced to 3 because a 3 by 3 mask gives the same result as 5 by 5 or any larger aperture. Returns: norm_image = grayscale distance-transformed image normalized between [0, 1] :param bin_img: numpy.ndarray :param distance_type: int :param mask_size: int :return norm_image: numpy.ndarray """ params.device += 1 dist = cv2.distanceTransform(src=bin_img, distanceType=distance_type, maskSize=mask_size) norm_image = cv2.normalize(src=dist, dst=dist, alpha=0, beta=1, norm_type=cv2.NORM_MINMAX, dtype=cv2.CV_32F) if params.debug == 'print': print_image(norm_image, os.path.join(params.debug, str(params.device) + '_distance_transform.png')) elif params.debug == 'plot': plot_image(norm_image, cmap='gray') return norm_image
Example #25
Source File: pySaliencyMap.py From aim with MIT License | 5 votes |
def OFMGetFM(self, src): # Creating a Gaussian pyramid GaussianI = self.FMCreateGaussianPyr(src) # Convoluting a Gabor filter with an intensity image to extract oriemtation features GaborOutput0 = [np.empty((1, 1)), np.empty((1, 1))] # dummy data: any kinds of np.array()s are OK GaborOutput45 = [np.empty((1, 1)), np.empty((1, 1))] GaborOutput90 = [np.empty((1, 1)), np.empty((1, 1))] GaborOutput135 = [np.empty((1, 1)), np.empty((1, 1))] for j in range(2, 9): GaborOutput0.append(cv2.filter2D(GaussianI[j], cv2.CV_32F, self.GaborKernel0)) GaborOutput45.append(cv2.filter2D(GaussianI[j], cv2.CV_32F, self.GaborKernel45)) GaborOutput90.append(cv2.filter2D(GaussianI[j], cv2.CV_32F, self.GaborKernel90)) GaborOutput135.append(cv2.filter2D(GaussianI[j], cv2.CV_32F, self.GaborKernel135)) # Calculating center-surround differences for every oriantation CSD0 = self.FMCenterSurroundDiff(GaborOutput0) CSD45 = self.FMCenterSurroundDiff(GaborOutput45) CSD90 = self.FMCenterSurroundDiff(GaborOutput90) CSD135 = self.FMCenterSurroundDiff(GaborOutput135) # Concatenate dst = list(CSD0) dst.extend(CSD45) dst.extend(CSD90) dst.extend(CSD135) return dst # Motion feature maps
Example #26
Source File: blockmatchers.py From StereoVision with GNU General Public License v3.0 | 5 votes |
def get_disparity(self, pair): """ Compute disparity from image pair (left, right). First, convert images to grayscale if needed. Then pass to the ``_block_matcher`` for stereo matching. """ gray = [] if pair[0].ndim == 3: for side in pair: gray.append(cv2.cvtColor(side, cv2.COLOR_BGR2GRAY)) else: gray = pair return self._block_matcher.compute(gray[0], gray[1], disptype=cv2.CV_32F)
Example #27
Source File: agent.py From PEDRA with MIT License | 5 votes |
def get_state(self): camera_image = get_MonocularImageRGB(self.client, self.vehicle_name) self.iter = self.iter + 1 state = cv2.resize(camera_image, (self.input_size, self.input_size), cv2.INTER_LINEAR) state = cv2.normalize(state, state, 0, 1, cv2.NORM_MINMAX, cv2.CV_32F) state_rgb = [] state_rgb.append(state[:, :, 0:3]) state_rgb = np.array(state_rgb) state_rgb = state_rgb.astype('float32') return state_rgb
Example #28
Source File: trainer.py From SalsaNext with MIT License | 5 votes |
def make_log_img(depth, mask, pred, gt, color_fn): # input should be [depth, pred, gt] # make range image (normalized to 0,1 for saving) depth = (cv2.normalize(depth, None, alpha=0, beta=1, norm_type=cv2.NORM_MINMAX, dtype=cv2.CV_32F) * 255.0).astype(np.uint8) out_img = cv2.applyColorMap( depth, Trainer.get_mpl_colormap('viridis')) * mask[..., None] # make label prediction pred_color = color_fn((pred * mask).astype(np.int32)) out_img = np.concatenate([out_img, pred_color], axis=0) # make label gt gt_color = color_fn(gt) out_img = np.concatenate([out_img, gt_color], axis=0) return (out_img).astype(np.uint8)
Example #29
Source File: niblack_thresholding.py From lpr with Apache License 2.0 | 5 votes |
def niBlackThreshold( src, blockSize, k, binarizationMethod= 0 ): mean = cv2.boxFilter(src,cv2.CV_32F,(blockSize, blockSize),borderType=cv2.BORDER_REPLICATE) sqmean = cv2.sqrBoxFilter(src, cv2.CV_32F, (blockSize, blockSize), borderType = cv2.BORDER_REPLICATE) variance = sqmean - (mean*mean) stddev = np.sqrt(variance) thresh = mean + stddev * float(-k) thresh = thresh.astype(src.dtype) k = (src>thresh)*255 k = k.astype(np.uint8) return k # cv2.imshow()
Example #30
Source File: cut_part.py From 2019-CCF-BDCI-OCR-MCZJ-OCR-IdentificationIDElement with MIT License | 5 votes |
def gradient_and_binary(img_blurred, image_name='1.jpg', save_path='./'): # 将灰度图二值化,后面两个参数调试用 """ 求取梯度,二值化 :param img_blurred: 滤波后的图片 :param image_name: 图片名,测试用 :param save_path: 保存路径,测试用 :return: 二值化后的图片 """ gradX = cv2.Sobel(img_blurred, ddepth=cv2.CV_32F, dx=1, dy=0) gradY = cv2.Sobel(img_blurred, ddepth=cv2.CV_32F, dx=0, dy=1) img_gradient = cv2.subtract(gradX, gradY) img_gradient = cv2.convertScaleAbs(img_gradient) # sobel算子,计算梯度, 也可以用canny算子替代 # 这里改进成自适应阈值,貌似没用 img_thresh = cv2.adaptiveThreshold(img_gradient, 255, cv2.ADAPTIVE_THRESH_MEAN_C, cv2.THRESH_BINARY, 3, -3) # cv2.imwrite(os.path.join(save_path, img_name + '_binary.jpg'), img_thresh) # 二值化 阈值未调整好 kernel = cv2.getStructuringElement(cv2.MORPH_ELLIPSE, (5, 5)) img_closed = cv2.morphologyEx(img_thresh, cv2.MORPH_CLOSE, kernel) img_closed = cv2.morphologyEx(img_closed, cv2.MORPH_OPEN, kernel) img_closed = cv2.erode(img_closed, None, iterations=9) img_closed = cv2.dilate(img_closed, None, iterations=9) # 腐蚀膨胀 # 这里调整了kernel大小(减小),腐蚀膨胀次数后(增大),出错的概率大幅减小 return img_closed