Python caffe2.proto.caffe2_pb2.CUDA Examples
The following are 30
code examples of caffe2.proto.caffe2_pb2.CUDA().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
caffe2.proto.caffe2_pb2
, or try the search function
.
Example #1
Source File: test_batch_permutation_op.py From masktextspotter.caffe2 with Apache License 2.0 | 6 votes |
def _run_op_test(self, X, I, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('BatchPermutation', ['X', 'I'], ['Y']) workspace.FeedBlob('X', X) workspace.FeedBlob('I', I) workspace.RunOperatorOnce(op) Y = workspace.FetchBlob('Y') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.1, threshold=0.001, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [X, I], 0, [0]) self.assertTrue(res, 'Grad check failed') Y_ref = X[I] np.testing.assert_allclose(Y, Y_ref, rtol=1e-5, atol=1e-08)
Example #2
Source File: test_spatial_narrow_as_op.py From Detectron-DA-Faster-RCNN with Apache License 2.0 | 6 votes |
def _run_test(self, A, B, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('SpatialNarrowAs', ['A', 'B'], ['C']) workspace.FeedBlob('A', A) workspace.FeedBlob('B', B) workspace.RunOperatorOnce(op) C = workspace.FetchBlob('C') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [A, B], 0, [0]) self.assertTrue(res, 'Grad check failed') dims = C.shape C_ref = A[:dims[0], :dims[1], :dims[2], :dims[3]] np.testing.assert_allclose(C, C_ref, rtol=1e-5, atol=1e-08)
Example #3
Source File: caffe2_graph.py From tensorboardX with MIT License | 6 votes |
def _tf_device(device_option): ''' Handle the devices. Args: device_option (caffe2_pb2.DeviceOption): DeviceOption protobuf, associated to an operator, that contains information such as device_type (optional), cuda_gpu_id (optional), node_name (optional, tells which node the operator should execute on). See caffe2.proto in caffe2/proto for the full list. Returns: Formatted string representing device information contained in device_option. ''' if not device_option.HasField("device_type"): return "" if device_option.device_type == caffe2_pb2.CPU or device_option.device_type == caffe2_pb2.MKLDNN: return "/cpu:*" if device_option.device_type == caffe2_pb2.CUDA: return "/gpu:{}".format(device_option.device_id) raise Exception("Unhandled device", device_option)
Example #4
Source File: common_caffe2.py From optimized-models with Apache License 2.0 | 6 votes |
def UpdateDeviceOption(dev_opt, net_def): """update device options in net_def""" # net_def.device_option.CopyFrom(dev_opt) # gpufallbackop=['GenerateProposals', 'BoxWithNMSLimit', 'BBoxTransform', # 'PackedInt8BGRANHWCToNCHWCStylizerPreprocess', 'BRGNCHWCToPackedInt8BGRAStylizerDeprocess'] gpufallbackop = ['GenerateProposals', 'BoxWithNMSLimit', 'BBoxTransform'] # gpufallbackop=[] ideepfallbackop = [] from caffe2.proto import caffe2_pb2 for eop in net_def.op: if (eop.type in gpufallbackop and dev_opt.device_type == caffe2_pb2.CUDA) or ( eop.type in ideepfallbackop and dev_opt.device_type == caffe2_pb2.IDEEP): eop.device_option.device_type = caffe2_pb2.CPU elif ( eop.device_option and eop.device_option.device_type != dev_opt.device_type ): eop.device_option.device_type = dev_opt.device_type
Example #5
Source File: test_spatial_narrow_as_op.py From KL-Loss with Apache License 2.0 | 6 votes |
def _run_test(self, A, B, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('SpatialNarrowAs', ['A', 'B'], ['C']) workspace.FeedBlob('A', A) workspace.FeedBlob('B', B) workspace.RunOperatorOnce(op) C = workspace.FetchBlob('C') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [A, B], 0, [0]) self.assertTrue(res, 'Grad check failed') dims = C.shape C_ref = A[:dims[0], :dims[1], :dims[2], :dims[3]] np.testing.assert_allclose(C, C_ref, rtol=1e-5, atol=1e-08)
Example #6
Source File: caffe2_graph.py From tensorboardX with MIT License | 6 votes |
def _tf_device(device_option): ''' Handle the devices. Args: device_option (caffe2_pb2.DeviceOption): DeviceOption protobuf, associated to an operator, that contains information such as device_type (optional), cuda_gpu_id (optional), node_name (optional, tells which node the operator should execute on). See caffe2.proto in caffe2/proto for the full list. Returns: Formatted string representing device information contained in device_option. ''' if not device_option.HasField("device_type"): return "" if device_option.device_type == caffe2_pb2.CPU or device_option.device_type == caffe2_pb2.MKLDNN: return "/cpu:*" if device_option.device_type == caffe2_pb2.CUDA: return "/gpu:{}".format(device_option.device_id) raise Exception("Unhandled device", device_option)
Example #7
Source File: bn_helper.py From video-long-term-feature-banks with Apache License 2.0 | 6 votes |
def _update_bn_stats_gpu(self): """ Copy to GPU. Note: the actual blobs used at test time are "rm" and "riv" """ num_gpus = cfg.NUM_GPUS root_gpu_id = cfg.ROOT_GPU_ID for i in range(root_gpu_id, root_gpu_id + num_gpus): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, i)): for bn_layer in self._bn_layers: workspace.FeedBlob( 'gpu_{}/'.format(i) + bn_layer + '_bn_rm', np.array(self._meanX_dict[bn_layer], dtype=np.float32), ) """ Note: riv is acutally running var (not running inv var)!!!! """ workspace.FeedBlob( 'gpu_{}/'.format(i) + bn_layer + '_bn_riv', np.array(self._var_dict[bn_layer], dtype=np.float32), )
Example #8
Source File: dataloader.py From video-long-term-feature-banks with Apache License 2.0 | 6 votes |
def enqueue_blobs( self, gpu_id, enqueue_blobs_names, blob_values, ): enqueue_blobs_names = [ 'gpu_{}/{}'.format( gpu_id, enqueue_blob_name ) for enqueue_blob_name in enqueue_blobs_names ] deviceOption = core.DeviceOption(caffe2_pb2.CUDA, gpu_id) for (blob_name, blob) in zip(enqueue_blobs_names, blob_values): workspace.FeedBlob(blob_name, blob, device_option=deviceOption) queue_name = 'gpu_{}/{}'.format(gpu_id, self._blobs_queue_name) workspace.RunOperatorOnce( core.CreateOperator( 'SafeEnqueueBlobs', [queue_name] + enqueue_blobs_names, enqueue_blobs_names + [queue_name + '_enqueue_status'], device_option=deviceOption, ) )
Example #9
Source File: test_spatial_narrow_as_op.py From Clustered-Object-Detection-in-Aerial-Image with Apache License 2.0 | 6 votes |
def _run_test(self, A, B, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('SpatialNarrowAs', ['A', 'B'], ['C']) workspace.FeedBlob('A', A) workspace.FeedBlob('B', B) workspace.RunOperatorOnce(op) C = workspace.FetchBlob('C') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [A, B], 0, [0]) self.assertTrue(res, 'Grad check failed') dims = C.shape C_ref = A[:dims[0], :dims[1], :dims[2], :dims[3]] np.testing.assert_allclose(C, C_ref, rtol=1e-5, atol=1e-08)
Example #10
Source File: caffe2_network.py From deep500 with BSD 3-Clause "New" or "Revised" License | 6 votes |
def __init__(self, device_option: DeviceOption): super(Caffe2Network, self).__init__() self.device_option = device_option self.train_model = model_helper.ModelHelper(name="train_default_net") self.test_model = model_helper.ModelHelper(name="test_default_net", init_params=False) self.train_net = self.train_model.net self.test_net = self.test_model.net self.train_init_net = self.train_model.param_init_net self.test_init_net = self.test_model.param_init_net self.workspace = workspace self.output_dict = {} self.param_names = None # dict that helps us remember that we already added the gradients to the graph for a given loss self.gradients_by_loss = {} self.is_cuda = (device_option.device_type == caffe2_pb2.CUDA)
Example #11
Source File: caffe2_graph_executor.py From deep500 with BSD 3-Clause "New" or "Revised" License | 6 votes |
def inference(self, inputs: Dict[str, np.ndarray]) -> Dict[str, np.ndarray]: self.network._setup() for event in self.events: event.before_executor(inputs) if self.network.device_option.device_type == caffe2_pb2.CUDA: self.network.test_model.net.RunAllOnGPU(use_cudnn=True) for node_name, value in inputs.items(): self.network.workspace.FeedBlob(node_name, value, self.device_option) self.network.workspace.RunNetOnce(self.network.test_model.net) for key in self.network.output_dict.keys(): self.network.output_dict[key] = self.network.workspace.FetchBlob(key) for event in self.events: event.after_inference(self.network.output_dict) return self.network.output_dict
Example #12
Source File: caffe2_graph_executor.py From deep500 with BSD 3-Clause "New" or "Revised" License | 6 votes |
def inference_and_backprop(self, inputs: Dict[str, np.ndarray], y: str = 'loss') -> Dict[str, np.ndarray]: self.network._setup() self.network.add_loss(y) for event in self.events: event.before_executor(inputs) if self.device_option.device_type == caffe2_pb2.CUDA: self.network.train_model.param_init_net.RunAllOnGPU(use_cudnn=True) self.network.train_model.net.RunAllOnGPU(use_cudnn=True) for node_name, value in inputs.items(): self.network.workspace.FeedBlob(node_name, value, self.device_option) self.network.workspace.RunNetOnce(self.network.train_model.net) for key in self.network.output_dict.keys(): self.network.output_dict[key] = self.network.workspace.FetchBlob(key) for event in self.events: event.after_backprop(self.network.output_dict) return self.network.output_dict
Example #13
Source File: test_spatial_narrow_as_op.py From Detectron with Apache License 2.0 | 6 votes |
def _run_test(self, A, B, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('SpatialNarrowAs', ['A', 'B'], ['C']) workspace.FeedBlob('A', A) workspace.FeedBlob('B', B) workspace.RunOperatorOnce(op) C = workspace.FetchBlob('C') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [A, B], 0, [0]) self.assertTrue(res, 'Grad check failed') dims = C.shape C_ref = A[:dims[0], :dims[1], :dims[2], :dims[3]] np.testing.assert_allclose(C, C_ref, rtol=1e-5, atol=1e-08)
Example #14
Source File: test_spatial_narrow_as_op.py From seg_every_thing with Apache License 2.0 | 6 votes |
def _run_test(self, A, B, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('SpatialNarrowAs', ['A', 'B'], ['C']) workspace.FeedBlob('A', A) workspace.FeedBlob('B', B) workspace.RunOperatorOnce(op) C = workspace.FetchBlob('C') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [A, B], 0, [0]) self.assertTrue(res, 'Grad check failed') dims = C.shape C_ref = A[:dims[0], :dims[1], :dims[2], :dims[3]] np.testing.assert_allclose(C, C_ref, rtol=1e-5, atol=1e-08)
Example #15
Source File: test_batch_permutation_op.py From seg_every_thing with Apache License 2.0 | 6 votes |
def _run_op_test(self, X, I, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('BatchPermutation', ['X', 'I'], ['Y']) workspace.FeedBlob('X', X) workspace.FeedBlob('I', I) workspace.RunOperatorOnce(op) Y = workspace.FetchBlob('Y') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.1, threshold=0.001, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [X, I], 0, [0]) self.assertTrue(res, 'Grad check failed') Y_ref = X[I] np.testing.assert_allclose(Y, Y_ref, rtol=1e-5, atol=1e-08)
Example #16
Source File: test_spatial_narrow_as_op.py From Detectron-Cascade-RCNN with Apache License 2.0 | 6 votes |
def _run_test(self, A, B, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('SpatialNarrowAs', ['A', 'B'], ['C']) workspace.FeedBlob('A', A) workspace.FeedBlob('B', B) workspace.RunOperatorOnce(op) C = workspace.FetchBlob('C') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [A, B], 0, [0]) self.assertTrue(res, 'Grad check failed') dims = C.shape C_ref = A[:dims[0], :dims[1], :dims[2], :dims[3]] np.testing.assert_allclose(C, C_ref, rtol=1e-5, atol=1e-08)
Example #17
Source File: test_spatial_narrow_as_op.py From masktextspotter.caffe2 with Apache License 2.0 | 6 votes |
def _run_test(self, A, B, check_grad=False): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('SpatialNarrowAs', ['A', 'B'], ['C']) workspace.FeedBlob('A', A) workspace.FeedBlob('B', B) workspace.RunOperatorOnce(op) C = workspace.FetchBlob('C') if check_grad: gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple(op, [A, B], 0, [0]) self.assertTrue(res, 'Grad check failed') dims = C.shape C_ref = A[:dims[0], :dims[1], :dims[2], :dims[3]] np.testing.assert_allclose(C, C_ref, rtol=1e-5, atol=1e-08)
Example #18
Source File: model_loader.py From VMZ with Apache License 2.0 | 6 votes |
def BroacastParameters(model, src_gpu, gpus): log.info("Broadcasting parameters from gpu {} to gpu: {}".format( src_gpu, ','.join([str(g) for g in gpus])) ) for param in model.params: if 'gpu_{}'.format(gpus[0]) in str(param): for i in gpus: blob = workspace.FetchBlob(str(param)) target_blob_name = str(param).replace( 'gpu_{}'.format(src_gpu), 'gpu_{}'.format(i) ) log.info('broadcast {} -> {}'.format( str(param), target_blob_name) ) workspace.FetchBlob(str(param)) with core.DeviceScope( core.DeviceOption(caffe2_pb2.CUDA, i)): workspace.FeedBlob(target_blob_name, blob)
Example #19
Source File: c2.py From Detectron-Cascade-RCNN with Apache License 2.0 | 5 votes |
def CudaScope(gpu_id): """Create a CUDA device scope for GPU device `gpu_id`.""" gpu_dev = CudaDevice(gpu_id) with core.DeviceScope(gpu_dev): yield
Example #20
Source File: c2.py From Detectron-Cascade-RCNN with Apache License 2.0 | 5 votes |
def NamedCudaScope(gpu_id): """Creates a GPU name scope and CUDA device scope. This function is provided to reduce `with ...` nesting levels.""" with GpuNameScope(gpu_id): with CudaScope(gpu_id): yield
Example #21
Source File: c2.py From masktextspotter.caffe2 with Apache License 2.0 | 5 votes |
def CudaDevice(gpu_id): """Create a Cuda device.""" return core.DeviceOption(caffe2_pb2.CUDA, gpu_id)
Example #22
Source File: c2.py From masktextspotter.caffe2 with Apache License 2.0 | 5 votes |
def CudaScope(gpu_id): """Create a CUDA device scope for GPU device `gpu_id`.""" gpu_dev = CudaDevice(gpu_id) with core.DeviceScope(gpu_dev): yield
Example #23
Source File: c2.py From CBNet with Apache License 2.0 | 5 votes |
def CudaDevice(gpu_id): """Create a Cuda device.""" return core.DeviceOption(caffe2_pb2.CUDA, gpu_id)
Example #24
Source File: c2.py From CBNet with Apache License 2.0 | 5 votes |
def NamedCudaScope(gpu_id): """Creates a GPU name scope and CUDA device scope. This function is provided to reduce `with ...` nesting levels.""" with GpuNameScope(gpu_id): with CudaScope(gpu_id): yield
Example #25
Source File: test_batch_permutation_op.py From Detectron-Cascade-RCNN with Apache License 2.0 | 5 votes |
def test_size_exceptions(self): A = np.random.randn(2, 256, 42, 86).astype(np.float32) I = np.array(np.random.permutation(10), dtype=np.int32) with self.assertRaises(RuntimeError): self._run_op_test(A, I) # See doc string in _run_speed_test # def test_perf(self): # with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): # self._run_speed_test()
Example #26
Source File: test_zero_even_op.py From Detectron-Cascade-RCNN with Apache License 2.0 | 5 votes |
def _run_zero_even_op_gpu(self, X): with core.DeviceScope(core.DeviceOption(caffe2_pb2.CUDA, 0)): op = core.CreateOperator('ZeroEven', ['X'], ['Y']) workspace.FeedBlob('X', X) workspace.RunOperatorOnce(op) Y = workspace.FetchBlob('Y') return Y
Example #27
Source File: test_loader.py From Detectron-Cascade-RCNN with Apache License 2.0 | 5 votes |
def run_net(net): workspace.RunNetOnce(net) gpu_dev = core.DeviceOption(caffe2_pb2.CUDA, 0) name_scope = 'gpu_{}'.format(0) with core.NameScope(name_scope): with core.DeviceScope(gpu_dev): data = workspace.FetchBlob(core.ScopedName('data')) return data
Example #28
Source File: test_smooth_l1_loss_op.py From Detectron-Cascade-RCNN with Apache License 2.0 | 5 votes |
def test_forward_and_gradient(self): Y = np.random.randn(128, 4 * 21).astype(np.float32) Y_hat = np.random.randn(128, 4 * 21).astype(np.float32) inside_weights = np.random.randn(128, 4 * 21).astype(np.float32) inside_weights[inside_weights < 0] = 0 outside_weights = np.random.randn(128, 4 * 21).astype(np.float32) outside_weights[outside_weights < 0] = 0 scale = np.random.random() beta = np.random.random() op = core.CreateOperator( 'SmoothL1Loss', ['Y_hat', 'Y', 'inside_weights', 'outside_weights'], ['loss'], scale=scale, beta=beta ) gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple( op, [Y_hat, Y, inside_weights, outside_weights], 0, [0] ) self.assertTrue( grad.shape == grad_estimated.shape, 'Fail check: grad.shape != grad_estimated.shape' ) # To inspect the gradient and estimated gradient: # np.set_printoptions(precision=3, suppress=True) # print('grad:') # print(grad) # print('grad_estimated:') # print(grad_estimated) self.assertTrue(res)
Example #29
Source File: test_smooth_l1_loss_op.py From Detectron-DA-Faster-RCNN with Apache License 2.0 | 5 votes |
def test_forward_and_gradient(self): Y = np.random.randn(128, 4 * 21).astype(np.float32) Y_hat = np.random.randn(128, 4 * 21).astype(np.float32) inside_weights = np.random.randn(128, 4 * 21).astype(np.float32) inside_weights[inside_weights < 0] = 0 outside_weights = np.random.randn(128, 4 * 21).astype(np.float32) outside_weights[outside_weights < 0] = 0 scale = np.random.random() beta = np.random.random() op = core.CreateOperator( 'SmoothL1Loss', ['Y_hat', 'Y', 'inside_weights', 'outside_weights'], ['loss'], scale=scale, beta=beta ) gc = gradient_checker.GradientChecker( stepsize=0.005, threshold=0.005, device_option=core.DeviceOption(caffe2_pb2.CUDA, 0) ) res, grad, grad_estimated = gc.CheckSimple( op, [Y_hat, Y, inside_weights, outside_weights], 0, [0] ) self.assertTrue( grad.shape == grad_estimated.shape, 'Fail check: grad.shape != grad_estimated.shape' ) # To inspect the gradient and estimated gradient: # np.set_printoptions(precision=3, suppress=True) # print('grad:') # print(grad) # print('grad_estimated:') # print(grad_estimated) self.assertTrue(res)
Example #30
Source File: model_convert_utils.py From Detectron-Cascade-RCNN with Apache License 2.0 | 5 votes |
def get_device_option_cuda(gpu_id=0): device_option = caffe2_pb2.DeviceOption() device_option.device_type = caffe2_pb2.CUDA device_option.cuda_gpu_id = gpu_id return device_option