Secure your code as it's written. Use Snyk Code to scan source code in minutes - no build needed - and fix issues immediately.
gamma_regularizer, prior_mean, prior_stddev),
recurrent_alpha_regularizer=rank1_utils.make_regularizer(
alpha_regularizer, prior_mean, prior_stddev),
recurrent_gamma_regularizer=rank1_utils.make_regularizer(
gamma_regularizer, prior_mean, prior_stddev),
kernel_regularizer=tf.keras.regularizers.l2(l2),
recurrent_regularizer=tf.keras.regularizers.l2(l2),
bias_regularizer=tf.keras.regularizers.l2(l2),
use_additive_perturbation=use_additive_perturbation,
ensemble_size=ensemble_size)
cells.append(lstm_cell)
self.rnn_layer = tf.keras.layers.RNN(cells, return_sequences=False)
# 2. Affine layer on combination of RNN output and context features.
if self.hidden_layer_dim > 0:
self.hidden_layer = rank1_bnn_layers.DenseRank1(
self.hidden_layer_dim,
activation=tf.nn.relu6,
alpha_initializer=rank1_utils.make_initializer(
alpha_initializer, random_sign_init, dropout_rate),
gamma_initializer=rank1_utils.make_initializer(
gamma_initializer, random_sign_init, dropout_rate),
kernel_initializer="he_normal",
alpha_regularizer=rank1_utils.make_regularizer(
alpha_regularizer, prior_mean, prior_stddev),
gamma_regularizer=rank1_utils.make_regularizer(
gamma_regularizer, prior_mean, prior_stddev),
kernel_regularizer=tf.keras.regularizers.l2(l2),
bias_regularizer=tf.keras.regularizers.l2(l2),
use_additive_perturbation=use_additive_perturbation,
ensemble_size=ensemble_size)