@@ -12,10 +12,8 @@ def __init__(self, num_layers, num_units, batch_size, input_size, keep_prob=1.0,
1212 self .dropout_mask = []
1313 for layer in range (num_layers ):
1414 input_size_ = input_size if layer == 0 else 2 * num_units
15- gru_fw = tf .contrib .cudnn_rnn .CudnnGRU (
16- 1 , num_units , kernel_initializer = tf .random_normal_initializer (stddev = 0.1 ))
17- gru_bw = tf .contrib .cudnn_rnn .CudnnGRU (
18- 1 , num_units , kernel_initializer = tf .random_normal_initializer (stddev = 0.1 ))
15+ gru_fw = tf .contrib .cudnn_rnn .CudnnGRU (1 , num_units )
16+ gru_bw = tf .contrib .cudnn_rnn .CudnnGRU (1 , num_units )
1917 init_fw = tf .tile (tf .Variable (
2018 tf .zeros ([1 , 1 , num_units ])), [1 , batch_size , 1 ])
2119 init_bw = tf .tile (tf .Variable (
@@ -62,10 +60,8 @@ def __init__(self, num_layers, num_units, batch_size, input_size, keep_prob=1.0,
6260 self .scope = scope
6361 for layer in range (num_layers ):
6462 input_size_ = input_size if layer == 0 else 2 * num_units
65- gru_fw = tf .contrib .rnn .GRUCell (
66- num_units , kernel_initializer = tf .random_normal_initializer (stddev = 0.1 ))
67- gru_bw = tf .contrib .rnn .GRUCell (
68- num_units , kernel_initializer = tf .random_normal_initializer (stddev = 0.1 ))
63+ gru_fw = tf .contrib .rnn .GRUCell (num_units )
64+ gru_bw = tf .contrib .rnn .GRUCell (num_units )
6965 init_fw = tf .tile (tf .Variable (
7066 tf .zeros ([1 , num_units ])), [batch_size , 1 ])
7167 init_bw = tf .tile (tf .Variable (
0 commit comments