Python keras.models() Examples
The following are 30
code examples of keras.models().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
keras
, or try the search function
.
Example #1
Source File: shufflenetv2b.py From imgclsmob with MIT License | 6 votes |
def shufflenetv2b_wd2(**kwargs): """ ShuffleNetV2(b) 0.5x model from 'ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design,' https://arxiv.org/abs/1807.11164. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. Returns ------- functor Functor for model graph creation with extra fields. """ return get_shufflenetv2b( width_scale=(12.0 / 29.0), model_name="shufflenetv2b_wd2", **kwargs)
Example #2
Source File: example.py From residual_block_keras with GNU General Public License v3.0 | 6 votes |
def get_residual_model(is_mnist=True, img_channels=1, img_rows=28, img_cols=28): model = keras.models.Sequential() first_layer_channel = 128 if is_mnist: # size to be changed to 32,32 model.add(ZeroPadding2D((2,2), input_shape=(img_channels, img_rows, img_cols))) # resize (28,28)-->(32,32) # the first conv model.add(Convolution2D(first_layer_channel, 3, 3, border_mode='same')) else: model.add(Convolution2D(first_layer_channel, 3, 3, border_mode='same', input_shape=(img_channels, img_rows, img_cols))) model.add(Activation('relu')) # [residual-based Conv layers] residual_blocks = design_for_residual_blocks(num_channel_input=first_layer_channel) model.add(residual_blocks) model.add(BatchNormalization(axis=1)) model.add(Activation('relu')) # [Classifier] model.add(Flatten()) model.add(Dense(nb_classes)) model.add(Activation('softmax')) # [END] return model
Example #3
Source File: seq2seq_class.py From Deep_Learning_Weather_Forecasting with Apache License 2.0 | 6 votes |
def __init__(self, model_save_path='../models', model_structure_name='seq2seq_model_demo', model_weights_name='seq2seq_model_demo', model_name=None): super().__init__() self.model_save_path = model_save_path self.model_structure_name=model_structure_name + self.model_name_format_str +'.json' self.model_weights_name=model_weights_name + self.model_name_format_str +'.h5' print('model_structure_name:', self.model_structure_name) print('model_weights_name:', self.model_weights_name) self.pred_result = None # Predicted mean value self.pred_var_result = None # Predicted variance value self.current_mean_val_loss = None self.EARLY_STOP=False self.val_loss_list=[] self.train_loss_list=[] self.pred_var_result = []
Example #4
Source File: prepare_model_yaml.py From models with MIT License | 6 votes |
def make_secondary_dl_yaml(template_yaml, model_json, output_yaml_path): with open(template_yaml, 'r') as f: model_yaml = yaml.load(f) # # get the model config: json_file = open(model_json, 'r') loaded_model_json = json_file.read() json_file.close() loaded_model = keras.models.model_from_json(loaded_model_json) # model_yaml["output_schema"]["targets"] = [] for oname, oshape in zip(loaded_model.output_names, loaded_model.output_shape): append_el ={"name":oname , "shape":str(oshape)#replace("None,", "") , "doc":"Methylation probability for %s"%oname} model_yaml["output_schema"]["targets"].append(append_el) # with open(output_yaml_path, 'w') as f: yaml.dump(model_yaml, f, default_flow_style=False)
Example #5
Source File: prepare_model_yaml.py From models with MIT License | 6 votes |
def make_model_yaml(template_yaml, model_json, output_yaml_path): # with open(template_yaml, 'r') as f: model_yaml = yaml.load(f) # # get the model config: json_file = open(model_json, 'r') loaded_model_json = json_file.read() json_file.close() loaded_model = keras.models.model_from_json(loaded_model_json) # model_yaml["schema"]["targets"] = [] for oname, oshape in zip(loaded_model.output_names, loaded_model.output_shape): append_el ={"name":oname , "shape":str(oshape)#replace("None,", "") , "doc":"Methylation probability for %s"%oname} model_yaml["schema"]["targets"].append(append_el) # with open(output_yaml_path, 'w') as f: yaml.dump(model_yaml, f, default_flow_style=False)
Example #6
Source File: ELMoBiLSTM.py From elmo-bilstm-cnn-crf with Apache License 2.0 | 6 votes |
def computeF1(self, modelName, sentences): labelKey = self.labelKeys[modelName] model = self.models[modelName] idx2Label = self.idx2Labels[modelName] correctLabels = [sentences[idx][labelKey] for idx in range(len(sentences))] predLabels = self.predictLabels(model, sentences) labelKey = self.labelKeys[modelName] encodingScheme = labelKey[labelKey.index('_')+1:] pre, rec, f1 = BIOF1Validation.compute_f1(predLabels, correctLabels, idx2Label, 'O', encodingScheme) pre_b, rec_b, f1_b = BIOF1Validation.compute_f1(predLabels, correctLabels, idx2Label, 'B', encodingScheme) if f1_b > f1: logging.debug("Setting wrong tags to B- improves from %.4f to %.4f" % (f1, f1_b)) pre, rec, f1 = pre_b, rec_b, f1_b return pre, rec, f1
Example #7
Source File: model.py From dataiku-contrib with Apache License 2.0 | 6 votes |
def compute_backbone_shapes(config, image_shape): """Computes the width and height of each stage of the backbone network. Returns: [N, (height, width)]. Where N is the number of stages """ if callable(config.BACKBONE): return config.COMPUTE_BACKBONE_SHAPE(image_shape) # Currently supports ResNet only assert config.BACKBONE in ["resnet50", "resnet101"] return np.array( [[int(math.ceil(image_shape[0] / stride)), int(math.ceil(image_shape[1] / stride))] for stride in config.BACKBONE_STRIDES]) ############################################################ # Resnet Graph ############################################################ # Code adopted from: # https://github.com/fchollet/deep-learning-models/blob/master/resnet50.py
Example #8
Source File: ELMoBiLSTM.py From elmo-bilstm-cnn-crf with Apache License 2.0 | 6 votes |
def saveModel(self, modelName, epoch, dev_score, test_score): import json import h5py if self.modelSavePath == None: raise ValueError('modelSavePath not specified.') savePath = self.modelSavePath.replace("[DevScore]", "%.4f" % dev_score).replace("[TestScore]", "%.4f" % test_score).replace("[Epoch]", str(epoch+1)).replace("[ModelName]", modelName) directory = os.path.dirname(savePath) if not os.path.exists(directory): os.makedirs(directory) if os.path.isfile(savePath): logging.info("Model "+savePath+" already exists. Model will be overwritten") self.models[modelName].save(savePath, True) with h5py.File(savePath, 'a') as h5file: h5file.attrs['mappings'] = json.dumps(self.mappings) h5file.attrs['params'] = json.dumps(self.params) h5file.attrs['modelName'] = modelName h5file.attrs['labelKey'] = self.datasets[modelName]['label']
Example #9
Source File: qrnn.py From typhon with MIT License | 6 votes |
def predict(self, x): r""" Predict quantiles of the conditional distribution P(y|x). Forward propagates the inputs in `x` through the network to obtain the predicted quantiles `y`. Arguments: x(np.array): Array of shape `(n, m)` containing `n` m-dimensional inputs for which to predict the conditional quantiles. Returns: Array of shape `(n, k)` with the columns corresponding to the k quantiles of the network. """ predictions = np.stack( [m.predict((x - self.x_mean) / self.x_sigma) for m in self.models]) return np.mean(predictions, axis=0)
Example #10
Source File: ELMoBiLSTM.py From elmo-bilstm-cnn-crf with Apache License 2.0 | 6 votes |
def tagSentences(self, sentences): # Pad characters if 'characters' in self.params['featureNames']: self.padCharacters(sentences) labels = {} for modelName, model in self.models.items(): paddedPredLabels = self.predictLabels(model, sentences) predLabels = [] for idx in range(len(sentences)): unpaddedPredLabels = [] for tokenIdx in range(len(sentences[idx]['tokens'])): if sentences[idx]['tokens'][tokenIdx] != 0: # Skip padding tokens unpaddedPredLabels.append(paddedPredLabels[idx][tokenIdx]) predLabels.append(unpaddedPredLabels) idx2Label = self.idx2Labels[modelName] labels[modelName] = [[idx2Label[tag] for tag in tagSentence] for tagSentence in predLabels] return labels
Example #11
Source File: test_shap.py From AIX360 with Apache License 2.0 | 6 votes |
def test_ShapGradientExplainer(self): # model = VGG16(weights='imagenet', include_top=True) # X, y = shap.datasets.imagenet50() # to_explain = X[[39, 41]] # # url = "https://s3.amazonaws.com/deep-learning-models/image-models/imagenet_class_index.json" # fname = shap.datasets.cache(url) # with open(fname) as f: # class_names = json.load(f) # # def map2layer(x, layer): # feed_dict = dict(zip([model.layers[0].input], [preprocess_input(x.copy())])) # return K.get_session().run(model.layers[layer].input, feed_dict) # # e = GradientExplainer((model.layers[7].input, model.layers[-1].output), # map2layer(preprocess_input(X.copy()), 7)) # shap_values, indexes = e.explain_instance(map2layer(to_explain, 7), ranked_outputs=2) # print("Skipped Shap GradientExplainer")
Example #12
Source File: model.py From PanopticSegmentation with MIT License | 6 votes |
def compute_backbone_shapes(config, image_shape): """Computes the width and height of each stage of the backbone network. Returns: [N, (height, width)]. Where N is the number of stages """ if callable(config.BACKBONE): return config.COMPUTE_BACKBONE_SHAPE(image_shape) # Currently supports ResNet only assert config.BACKBONE in ["resnet50", "resnet101"] return np.array( [[int(math.ceil(image_shape[0] / stride)), int(math.ceil(image_shape[1] / stride))] for stride in config.BACKBONE_STRIDES]) ############################################################ # Resnet Graph ############################################################ # Code adopted from: # https://github.com/fchollet/deep-learning-models/blob/master/resnet50.py
Example #13
Source File: keras_policy.py From rasa_wechat with Apache License 2.0 | 6 votes |
def _build_model(self, num_features, num_actions, max_history_len): """Build a keras model and return a compiled model. :param max_history_len: The maximum number of historical turns used to decide on next action """ from keras.layers import LSTM, Activation, Masking, Dense from keras.models import Sequential n_hidden = 32 # Neural Net and training params batch_shape = (None, max_history_len, num_features) # Build Model model = Sequential() model.add(Masking(-1, batch_input_shape=batch_shape)) model.add(LSTM(n_hidden, batch_input_shape=batch_shape)) model.add(Dense(input_dim=n_hidden, units=num_actions)) model.add(Activation('softmax')) model.compile(loss='categorical_crossentropy', optimizer='rmsprop', metrics=['accuracy']) logger.debug(model.summary()) return model
Example #14
Source File: shufflenetv2b.py From imgclsmob with MIT License | 6 votes |
def shufflenetv2b_w2(**kwargs): """ ShuffleNetV2(b) 2x model from 'ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design,' https://arxiv.org/abs/1807.11164. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. Returns ------- functor Functor for model graph creation with extra fields. """ return get_shufflenetv2b( width_scale=(61.0 / 29.0), model_name="shufflenetv2b_w2", **kwargs)
Example #15
Source File: shufflenetv2b.py From imgclsmob with MIT License | 6 votes |
def shufflenetv2b_w3d2(**kwargs): """ ShuffleNetV2(b) 1.5x model from 'ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design,' https://arxiv.org/abs/1807.11164. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. Returns ------- functor Functor for model graph creation with extra fields. """ return get_shufflenetv2b( width_scale=(44.0 / 29.0), model_name="shufflenetv2b_w3d2", **kwargs)
Example #16
Source File: shufflenetv2b.py From imgclsmob with MIT License | 6 votes |
def shufflenetv2b_w1(**kwargs): """ ShuffleNetV2(b) 1x model from 'ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design,' https://arxiv.org/abs/1807.11164. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. Returns ------- functor Functor for model graph creation with extra fields. """ return get_shufflenetv2b( width_scale=1.0, model_name="shufflenetv2b_w1", **kwargs)
Example #17
Source File: squeezenet.py From imgclsmob with MIT License | 5 votes |
def squeezenet_v1_0(**kwargs): """ SqueezeNet 'vanilla' model from 'SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size,' https://arxiv.org/abs/1602.07360. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_squeezenet(version="1.0", residual=False, model_name="squeezenet_v1_0", **kwargs)
Example #18
Source File: mobilenetv2.py From imgclsmob with MIT License | 5 votes |
def mobilenetv2_w3d4(**kwargs): """ 0.75 MobileNetV2-224 model from 'MobileNetV2: Inverted Residuals and Linear Bottlenecks,' https://arxiv.org/abs/1801.04381. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_mobilenetv2(width_scale=0.75, model_name="mobilenetv2_w3d4", **kwargs)
Example #19
Source File: mobilenetv2.py From imgclsmob with MIT License | 5 votes |
def mobilenetv2_w1(**kwargs): """ 1.0 MobileNetV2-224 model from 'MobileNetV2: Inverted Residuals and Linear Bottlenecks,' https://arxiv.org/abs/1801.04381. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_mobilenetv2(width_scale=1.0, model_name="mobilenetv2_w1", **kwargs)
Example #20
Source File: mnasnet.py From imgclsmob with MIT License | 5 votes |
def mnasnet_a1(**kwargs): """ MnasNet-A1 model from 'MnasNet: Platform-Aware Neural Architecture Search for Mobile,' https://arxiv.org/abs/1807.11626. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_mnasnet(version="a1", width_scale=1.0, model_name="mnasnet_a1", **kwargs)
Example #21
Source File: squeezenet.py From imgclsmob with MIT License | 5 votes |
def squeezeresnet_v1_1(**kwargs): """ SqueezeNet v1.1 model with residual connections from 'SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size,' https://arxiv.org/abs/1602.07360. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_squeezenet(version="1.1", residual=True, model_name="squeezeresnet_v1_1", **kwargs)
Example #22
Source File: squeezenet.py From imgclsmob with MIT License | 5 votes |
def squeezenet_v1_1(**kwargs): """ SqueezeNet v1.1 model from 'SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size,' https://arxiv.org/abs/1602.07360. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_squeezenet(version="1.1", residual=False, model_name="squeezenet_v1_1", **kwargs)
Example #23
Source File: squeezenet.py From imgclsmob with MIT License | 5 votes |
def _test(): import numpy as np import keras pretrained = False models = [ squeezenet_v1_0, squeezenet_v1_1, squeezeresnet_v1_0, squeezeresnet_v1_1, ] for model in models: net = model(pretrained=pretrained) # net.summary() weight_count = keras.utils.layer_utils.count_params(net.trainable_weights) print("m={}, {}".format(model.__name__, weight_count)) assert (model != squeezenet_v1_0 or weight_count == 1248424) assert (model != squeezenet_v1_1 or weight_count == 1235496) assert (model != squeezeresnet_v1_0 or weight_count == 1248424) assert (model != squeezeresnet_v1_1 or weight_count == 1235496) if is_channels_first(): x = np.zeros((1, 3, 224, 224), np.float32) else: x = np.zeros((1, 224, 224, 3), np.float32) y = net.predict(x) assert (y.shape == (1, 1000))
Example #24
Source File: squeezenet.py From imgclsmob with MIT License | 5 votes |
def squeezeresnet_v1_0(**kwargs): """ SqueezeNet model with residual connections from 'SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size,' https://arxiv.org/abs/1602.07360. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_squeezenet(version="1.0", residual=True, model_name="squeezeresnet_v1_0", **kwargs)
Example #25
Source File: mobilenetv2.py From imgclsmob with MIT License | 5 votes |
def mobilenetv2_wd4(**kwargs): """ 0.25 MobileNetV2-224 model from 'MobileNetV2: Inverted Residuals and Linear Bottlenecks,' https://arxiv.org/abs/1801.04381. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_mobilenetv2(width_scale=0.25, model_name="mobilenetv2_wd4", **kwargs)
Example #26
Source File: mobilenetv2.py From imgclsmob with MIT License | 5 votes |
def _test(): import numpy as np import keras pretrained = False models = [ mobilenetv2_w1, mobilenetv2_w3d4, mobilenetv2_wd2, mobilenetv2_wd4, ] for model in models: net = model(pretrained=pretrained) # net.summary() weight_count = keras.utils.layer_utils.count_params(net.trainable_weights) print("m={}, {}".format(model.__name__, weight_count)) assert (model != mobilenetv2_w1 or weight_count == 3504960) assert (model != mobilenetv2_w3d4 or weight_count == 2627592) assert (model != mobilenetv2_wd2 or weight_count == 1964736) assert (model != mobilenetv2_wd4 or weight_count == 1516392) if is_channels_first(): x = np.zeros((1, 3, 224, 224), np.float32) else: x = np.zeros((1, 224, 224, 3), np.float32) y = net.predict(x) assert (y.shape == (1, 1000))
Example #27
Source File: resnext.py From imgclsmob with MIT License | 5 votes |
def resnext101_64x4d(**kwargs): """ ResNeXt-101 (64x4d) model from 'Aggregated Residual Transformations for Deep Neural Networks,' http://arxiv.org/abs/1611.05431. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_resnext(blocks=101, cardinality=64, bottleneck_width=4, model_name="resnext101_64x4d", **kwargs)
Example #28
Source File: resnext.py From imgclsmob with MIT License | 5 votes |
def resnext101_32x4d(**kwargs): """ ResNeXt-101 (32x4d) model from 'Aggregated Residual Transformations for Deep Neural Networks,' http://arxiv.org/abs/1611.05431. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_resnext(blocks=101, cardinality=32, bottleneck_width=4, model_name="resnext101_32x4d", **kwargs)
Example #29
Source File: resnext.py From imgclsmob with MIT License | 5 votes |
def resnext50_32x4d(**kwargs): """ ResNeXt-50 (32x4d) model from 'Aggregated Residual Transformations for Deep Neural Networks,' http://arxiv.org/abs/1611.05431. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_resnext(blocks=50, cardinality=32, bottleneck_width=4, model_name="resnext50_32x4d", **kwargs)
Example #30
Source File: mobilenetv2.py From imgclsmob with MIT License | 5 votes |
def mobilenetv2_wd2(**kwargs): """ 0.5 MobileNetV2-224 model from 'MobileNetV2: Inverted Residuals and Linear Bottlenecks,' https://arxiv.org/abs/1801.04381. Parameters: ---------- pretrained : bool, default False Whether to load the pretrained weights for model. root : str, default '~/.keras/models' Location for keeping the model parameters. """ return get_mobilenetv2(width_scale=0.5, model_name="mobilenetv2_wd2", **kwargs)