Python keras.initializers.get() Examples
The following are 30
code examples of keras.initializers.get().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
keras.initializers
, or try the search function
.
Example #1
Source File: attention.py From deephlapan with GNU General Public License v2.0 | 6 votes |
def __init__(self, W_regularizer=None, b_regularizer=None, W_constraint=None, b_constraint=None, bias=True, return_attention=False, **kwargs): self.supports_masking = True self.return_attention = return_attention self.init = initializers.get('glorot_uniform') self.W_regularizer = regularizers.get(W_regularizer) self.b_regularizer = regularizers.get(b_regularizer) self.W_constraint = constraints.get(W_constraint) self.b_constraint = constraints.get(b_constraint) self.bias = bias super(Attention, self).__init__(**kwargs)
Example #2
Source File: ChainCRF.py From elmo-bilstm-cnn-crf with Apache License 2.0 | 6 votes |
def __init__(self, init='glorot_uniform', U_regularizer=None, b_start_regularizer=None, b_end_regularizer=None, U_constraint=None, b_start_constraint=None, b_end_constraint=None, weights=None, **kwargs): super(ChainCRF, self).__init__(**kwargs) self.init = initializers.get(init) self.U_regularizer = regularizers.get(U_regularizer) self.b_start_regularizer = regularizers.get(b_start_regularizer) self.b_end_regularizer = regularizers.get(b_end_regularizer) self.U_constraint = constraints.get(U_constraint) self.b_start_constraint = constraints.get(b_start_constraint) self.b_end_constraint = constraints.get(b_end_constraint) self.initial_weights = weights self.supports_masking = True self.uses_learning_phase = True self.input_spec = [InputSpec(ndim=3)]
Example #3
Source File: models.py From DigiX_HuaWei_Population_Age_Attribution_Predict with MIT License | 6 votes |
def on_epoch_end(self, epoch, logs=None): logs = logs or {} self.epochs_since_last_save += 1 if self.epochs_since_last_save >= self.period: self.epochs_since_last_save = 0 #filepath = self.filepath.format(epoch=epoch + 1, **logs) current = logs.get(self.monitor) if current is None: warnings.warn('Can pick best model only with %s available, ' 'skipping.' % (self.monitor), RuntimeWarning) else: if self.monitor_op(current, self.best): if self.verbose > 0: print('\nEpoch %05d: %s improved from %0.5f to %0.5f,' ' storing weights.' % (epoch + 1, self.monitor, self.best, current)) self.best = current self.best_epochs = epoch + 1 self.best_weights = self.model.get_weights() else: if self.verbose > 0: print('\nEpoch %05d: %s did not improve' % (epoch + 1, self.monitor))
Example #4
Source File: rnn_feature.py From DigiX_HuaWei_Population_Age_Attribution_Predict with MIT License | 6 votes |
def __init__(self, W_regularizer=None, u_regularizer=None, b_regularizer=None, W_constraint=None, u_constraint=None, b_constraint=None, bias=True, **kwargs): self.supports_masking = True self.init = initializers.get('glorot_uniform') self.W_regularizer = regularizers.get(W_regularizer) self.u_regularizer = regularizers.get(u_regularizer) self.b_regularizer = regularizers.get(b_regularizer) self.W_constraint = constraints.get(W_constraint) self.u_constraint = constraints.get(u_constraint) self.b_constraint = constraints.get(b_constraint) self.bias = bias super(AttentionWithContext, self).__init__(**kwargs)
Example #5
Source File: layers.py From bert4keras with Apache License 2.0 | 6 votes |
def __init__( self, heads, head_size, key_size=None, use_bias=True, attention_scale=True, kernel_initializer='glorot_uniform', **kwargs ): super(MultiHeadAttention, self).__init__(**kwargs) self.heads = heads self.head_size = head_size self.out_dim = heads * head_size self.key_size = key_size or head_size self.use_bias = use_bias self.attention_scale = attention_scale self.kernel_initializer = initializers.get(kernel_initializer)
Example #6
Source File: attention_with_context.py From DeepResearch with MIT License | 6 votes |
def __init__(self, W_regularizer=None, u_regularizer=None, b_regularizer=None, W_constraint=None, u_constraint=None, b_constraint=None, bias=True, **kwargs): self.supports_masking = True self.init = initializers.get('glorot_uniform') self.W_regularizer = regularizers.get(W_regularizer) self.u_regularizer = regularizers.get(u_regularizer) self.b_regularizer = regularizers.get(b_regularizer) self.W_constraint = constraints.get(W_constraint) self.u_constraint = constraints.get(u_constraint) self.b_constraint = constraints.get(b_constraint) self.bias = bias super(AttentionWithContext, self).__init__(**kwargs)
Example #7
Source File: layers.py From bert4keras with Apache License 2.0 | 6 votes |
def __init__( self, center=True, scale=True, epsilon=None, conditional=False, hidden_units=None, hidden_activation='linear', hidden_initializer='glorot_uniform', **kwargs ): super(LayerNormalization, self).__init__(**kwargs) self.center = center self.scale = scale self.conditional = conditional self.hidden_units = hidden_units self.hidden_activation = activations.get(hidden_activation) self.hidden_initializer = initializers.get(hidden_initializer) self.epsilon = epsilon or 1e-12
Example #8
Source File: models.py From DigiX_HuaWei_Population_Age_Attribution_Predict with MIT License | 6 votes |
def __init__(self, W_regularizer=None, u_regularizer=None, b_regularizer=None, W_constraint=None, u_constraint=None, b_constraint=None, bias=True, **kwargs): self.supports_masking = True self.init = initializers.get('glorot_uniform') self.W_regularizer = regularizers.get(W_regularizer) self.u_regularizer = regularizers.get(u_regularizer) self.b_regularizer = regularizers.get(b_regularizer) self.W_constraint = constraints.get(W_constraint) self.u_constraint = constraints.get(u_constraint) self.b_constraint = constraints.get(b_constraint) self.bias = bias super(AttentionWithContext, self).__init__(**kwargs)
Example #9
Source File: models.py From DigiX_HuaWei_Population_Age_Attribution_Predict with MIT License | 6 votes |
def __init__(self, W_regularizer=None, u_regularizer=None, b_regularizer=None, W_constraint=None, u_constraint=None, b_constraint=None, bias=True, **kwargs): self.supports_masking = True self.init = initializers.get('glorot_uniform') self.W_regularizer = regularizers.get(W_regularizer) self.u_regularizer = regularizers.get(u_regularizer) self.b_regularizer = regularizers.get(b_regularizer) self.W_constraint = constraints.get(W_constraint) self.u_constraint = constraints.get(u_constraint) self.b_constraint = constraints.get(b_constraint) self.bias = bias super(AttentionWithContext, self).__init__(**kwargs)
Example #10
Source File: instance_normalization.py From Coloring-greyscale-images with MIT License | 6 votes |
def __init__(self, axis=None, epsilon=1e-3, center=True, scale=True, beta_initializer='zeros', gamma_initializer='ones', beta_regularizer=None, gamma_regularizer=None, beta_constraint=None, gamma_constraint=None, **kwargs): super(InstanceNormalization, self).__init__(**kwargs) self.supports_masking = True self.axis = axis self.epsilon = epsilon self.center = center self.scale = scale self.beta_initializer = initializers.get(beta_initializer) self.gamma_initializer = initializers.get(gamma_initializer) self.beta_regularizer = regularizers.get(beta_regularizer) self.gamma_regularizer = regularizers.get(gamma_regularizer) self.beta_constraint = constraints.get(beta_constraint) self.gamma_constraint = constraints.get(gamma_constraint)
Example #11
Source File: attention.py From Document-Classifier-LSTM with MIT License | 6 votes |
def __init__(self, W_regularizer=None, u_regularizer=None, b_regularizer=None, W_constraint=None, u_constraint=None, b_constraint=None, bias=True, **kwargs): self.init = initializers.get('glorot_uniform') self.W_regularizer = regularizers.get(W_regularizer) self.u_regularizer = regularizers.get(u_regularizer) self.b_regularizer = regularizers.get(b_regularizer) self.W_constraint = constraints.get(W_constraint) self.u_constraint = constraints.get(u_constraint) self.b_constraint = constraints.get(b_constraint) self.bias = bias super(AttentionWithContext, self).__init__(**kwargs)
Example #12
Source File: normalizations.py From se_relativisticgan with MIT License | 6 votes |
def __init__(self, axis=None, epsilon=1e-3, center=True, scale=True, beta_initializer='zeros', gamma_initializer='ones', beta_regularizer=None, gamma_regularizer=None, beta_constraint=None, gamma_constraint=None, **kwargs): super(InstanceNormalization, self).__init__(**kwargs) self.supports_masking = True self.axis = axis self.epsilon = epsilon self.center = center self.scale = scale self.beta_initializer = initializers.get(beta_initializer) self.gamma_initializer = initializers.get(gamma_initializer) self.beta_regularizer = regularizers.get(beta_regularizer) self.gamma_regularizer = regularizers.get(gamma_regularizer) self.beta_constraint = constraints.get(beta_constraint) self.gamma_constraint = constraints.get(gamma_constraint)
Example #13
Source File: FFNN.py From dts with MIT License | 6 votes |
def evaluate(self, inputs, fn_inverse=None, fn_plot=None): try: X, y = inputs inputs = X except: X, conditions, y = inputs inputs = [X, conditions] y_hat = self.predict(inputs) if fn_inverse is not None: y_hat = fn_inverse(y_hat) y = fn_inverse(y) if fn_plot is not None: fn_plot([y, y_hat]) results = [] for m in self.model.metrics: if isinstance(m, str): results.append(K.eval(K.mean(get(m)(y, y_hat)))) else: results.append(K.eval(K.mean(m(y, y_hat)))) return results
Example #14
Source File: normalizations.py From se_relativisticgan with MIT License | 6 votes |
def __init__(self, axis=-1, momentum=0.99, center=True, scale=True, epsilon=1e-3, r_max_value=3., d_max_value=5., t_delta=1e-3, weights=None, beta_initializer='zero', gamma_initializer='one', moving_mean_initializer='zeros', moving_variance_initializer='ones', gamma_regularizer=None, beta_regularizer=None, beta_constraint=None, gamma_constraint=None, **kwargs): self.supports_masking = True self.axis = axis self.epsilon = epsilon self.center = center self.scale = scale self.momentum = momentum self.gamma_regularizer = regularizers.get(gamma_regularizer) self.beta_regularizer = regularizers.get(beta_regularizer) self.initial_weights = weights self.r_max_value = r_max_value self.d_max_value = d_max_value self.t_delta = t_delta self.beta_initializer = initializers.get(beta_initializer) self.gamma_initializer = initializers.get(gamma_initializer) self.moving_mean_initializer = initializers.get(moving_mean_initializer) self.moving_variance_initializer = initializers.get(moving_variance_initializer) self.beta_constraint = constraints.get(beta_constraint) self.gamma_constraint = constraints.get(gamma_constraint) super(BatchRenormalization, self).__init__(**kwargs)
Example #15
Source File: keras2_emitter.py From MMdnn with MIT License | 6 votes |
def emit_Affine(self, IR_node, in_scope=False): if in_scope: raise NotImplementedError else: self.used_layers.add('Affine') if IR_node.layer.attr.get('beta', None) is None: bias = None else: bias = IR_node.layer.attr['beta'].f code = "{:<15} = Affine(name='{}', scale={}, bias={})({})".format( IR_node.variable_name, IR_node.name, IR_node.layer.attr['gamma'].f, bias, self.parent_variable_name(IR_node)) return code
Example #16
Source File: keras2_emitter.py From MMdnn with MIT License | 6 votes |
def _emit_h_zero(self, IR_node): if not self.layers_codes.get(IR_node.pattern, None): class_code = ''' class my_h_zero(keras.layers.Layer): def __init__(self, **kwargs): super(my_h_zero, self).__init__(**kwargs) def call(self, dummy): {:<15} = K.constant(np.full((1, {}), {})) return {} '''.format(IR_node.variable_name, IR_node.get_attr('fill_size'), IR_node.get_attr('fill_value'), IR_node.variable_name) self.layers_codes[IR_node.pattern] = class_code code = "{:<15} = my_h_zero()({})".format(IR_node.variable_name, self.parent_variable_name(IR_node)) return code
Example #17
Source File: layers.py From keras-utilities with MIT License | 6 votes |
def __init__(self, W_regularizer=None, u_regularizer=None, b_regularizer=None, W_constraint=None, u_constraint=None, b_constraint=None, bias=True, return_attention=False, **kwargs): self.supports_masking = True self.return_attention = return_attention self.init = initializers.get('glorot_uniform') self.W_regularizer = regularizers.get(W_regularizer) self.u_regularizer = regularizers.get(u_regularizer) self.b_regularizer = regularizers.get(b_regularizer) self.W_constraint = constraints.get(W_constraint) self.u_constraint = constraints.get(u_constraint) self.b_constraint = constraints.get(b_constraint) self.bias = bias super(AttentionWithContext, self).__init__(**kwargs)
Example #18
Source File: norm.py From deep_complex_networks with MIT License | 6 votes |
def __init__(self, epsilon=1e-4, axis=-1, beta_init='zeros', gamma_init='ones', gamma_regularizer=None, beta_regularizer=None, **kwargs): self.supports_masking = True self.beta_init = initializers.get(beta_init) self.gamma_init = initializers.get(gamma_init) self.epsilon = epsilon self.axis = axis self.gamma_regularizer = regularizers.get(gamma_regularizer) self.beta_regularizer = regularizers.get(beta_regularizer) super(LayerNormalization, self).__init__(**kwargs)
Example #19
Source File: word_vectors.py From keras-image-captioning with MIT License | 5 votes |
def vectorize_words(self, words): vectors = [] for word in words: vector = self._word_vector_of.get(word) vectors.append(vector) num_unknowns = len(filter(lambda x: x is None, vectors)) inits = self._initializer(shape=(num_unknowns, self._embedding_size)) inits = K.get_session().run(inits) inits = iter(inits) for i in range(len(vectors)): if vectors[i] is None: vectors[i] = next(inits) return np.array(vectors)
Example #20
Source File: bert.py From keras-bert-ner with MIT License | 5 votes |
def __init__(self, input_dim, output_dim, merge_mode="add", embeddings_initializer="zeros", **kwargs): super(PositionEmbedding, self).__init__(**kwargs) self.input_dim = input_dim self.output_dim = output_dim self.merge_mode = merge_mode self.embeddings_initializer = initializers.get(embeddings_initializer)
Example #21
Source File: keras2_emitter.py From MMdnn with MIT License | 5 votes |
def _emit_activation(self, IR_node, op, in_scope=False): if in_scope: code = "{:<15} = keras.activations.get('{}')({})".format( IR_node.variable_name, op, self.parent_variable_name(IR_node)) else: code = "{:<15} = layers.Activation(name='{}', activation='{}')({})".format( IR_node.variable_name, IR_node.name, op, self.parent_variable_name(IR_node)) return code
Example #22
Source File: extra_layers.py From MMdnn with MIT License | 5 votes |
def __init__(self, axis=-1, center=True, scale=True, beta_initializer='zeros', gamma_initializer='ones', **kwargs): super(Scale, self).__init__(**kwargs) self.supports_masking = True self.axis = axis self.center = center self.scale = scale self.beta_initializer = initializers.get(beta_initializer) self.gamma_initializer = initializers.get(gamma_initializer)
Example #23
Source File: capsule.py From Keras-TextClassification with MIT License | 5 votes |
def __init__(self, num_capsule, dim_capsule, routings=3, kernel_initializer='glorot_uniform', **kwargs): super(CapsuleLayer, self).__init__(**kwargs) self.num_capsule = num_capsule self.dim_capsule = dim_capsule self.routings = routings self.kernel_initializer = initializers.get(kernel_initializer)
Example #24
Source File: bert.py From keras-bert-ner with MIT License | 5 votes |
def __init__(self, heads, head_size, key_size=None, kernel_initializer="glorot_uniform", **kwargs): super(MultiHeadAttention, self).__init__(**kwargs) self.heads = heads self.head_size = head_size self.out_dim = heads * head_size self.key_size = key_size if key_size else head_size self.kernel_initializer = initializers.get(kernel_initializer)
Example #25
Source File: capsulelayers.py From Multi-level-DCNet with GNU General Public License v3.0 | 5 votes |
def __init__(self, num_capsule, dim_capsule, routings=3, kernel_initializer='glorot_uniform', **kwargs): super(CapsuleLayer, self).__init__(**kwargs) self.num_capsule = num_capsule self.dim_capsule = dim_capsule self.routings = routings self.kernel_initializer = initializers.get(kernel_initializer)
Example #26
Source File: word_vectors.py From keras-image-captioning with MIT License | 5 votes |
def __init__(self, vocab_words, initializer): self._vocab_words = set(vocab_words) self._word_vector_of = dict() self._initializer = initializers.get(initializer)
Example #27
Source File: layers.py From keras-utilities with MIT License | 5 votes |
def __init__(self, W_regularizer=None, b_regularizer=None, W_constraint=None, b_constraint=None, bias=True, return_attention=False, **kwargs): """ Keras Layer that implements an Attention mechanism for temporal data. Supports Masking. Follows the work of Raffel et al. [https://arxiv.org/abs/1512.08756] # Input shape 3D tensor with shape: `(samples, steps, features)`. # Output shape 2D tensor with shape: `(samples, features)`. :param kwargs: Just put it on top of an RNN Layer (GRU/LSTM/SimpleRNN) with return_sequences=True. The dimensions are inferred based on the output shape of the RNN. Note: The layer has been tested with Keras 1.x Example: # 1 model.add(LSTM(64, return_sequences=True)) model.add(Attention()) # next add a Dense layer (for classification/regression) or whatever... # 2 - Get the attention scores hidden = LSTM(64, return_sequences=True)(words) sentence, word_scores = Attention(return_attention=True)(hidden) """ self.supports_masking = True self.return_attention = return_attention self.init = initializers.get('glorot_uniform') self.W_regularizer = regularizers.get(W_regularizer) self.b_regularizer = regularizers.get(b_regularizer) self.W_constraint = constraints.get(W_constraint) self.b_constraint = constraints.get(b_constraint) self.bias = bias super(Attention, self).__init__(**kwargs)
Example #28
Source File: attention_layer.py From text-classifier with Apache License 2.0 | 5 votes |
def __init__(self, attention_dim): self.init = initializers.get('normal') self.supports_masking = True self.attention_dim = attention_dim super(AttLayer, self).__init__()
Example #29
Source File: capsulelayers.py From textcaps with MIT License | 5 votes |
def __init__(self, num_capsule, dim_capsule,channels, routings=3, kernel_initializer='glorot_uniform', **kwargs): super(CapsuleLayer, self).__init__(**kwargs) self.num_capsule = num_capsule self.dim_capsule = dim_capsule self.routings = routings self.channels = channels self.kernel_initializer = initializers.get(kernel_initializer)
Example #30
Source File: contrib.py From steppy-toolkit with MIT License | 5 votes |
def __init__(self, return_attention=False, **kwargs): self.init = initializers.get('uniform') self.supports_masking = True self.return_attention = return_attention super(AttentionWeightedAverage, self).__init__(**kwargs)