Home
last modified time | relevance | path

Searched refs:vocabulary_size (Results 1 – 17 of 17) sorted by relevance

/external/tensorflow/tensorflow/contrib/seq2seq/python/kernel_tests/
Dbasic_decoder_test.py133 vocabulary_size = 7
134 cell_depth = vocabulary_size # cell's logits must match vocabulary size
136 start_tokens = np.random.randint(0, vocabulary_size, size=batch_size)
140 embeddings = np.random.randn(vocabulary_size,
142 cell = rnn_cell.LSTMCell(vocabulary_size)
205 vocabulary_size = 7
206 cell_depth = vocabulary_size # cell's logits must match vocabulary size
209 start_tokens = np.random.randint(0, vocabulary_size, size=batch_size)
216 embeddings = np.random.randn(vocabulary_size,
218 cell = rnn_cell.LSTMCell(vocabulary_size)
[all …]
/external/tensorflow/tensorflow/python/estimator/
Dwarm_starting_util_test.py378 "sc_vocab", vocabulary_file=vocab_path, vocabulary_size=4)
415 "sc_vocab", vocabulary_file=vocab_path, vocabulary_size=4)
461 "sc_vocab", vocabulary_file=current_vocab_path, vocabulary_size=2)
484 new_vocab_size=sc_vocab.vocabulary_size,
546 "sc_vocab", vocabulary_file=vocab_path, vocabulary_size=4)
606 new_vocab_size=sc_vocab.vocabulary_size,
640 "sc_vocab", vocabulary_file=new_vocab_path, vocabulary_size=6)
668 new_vocab_size=sc_vocab.vocabulary_size,
709 "sc_vocab", vocabulary_file=new_vocab_path, vocabulary_size=6)
732 new_vocab_size=sc_vocab.vocabulary_size,
[all …]
/external/tensorflow/tensorflow/examples/tutorials/word2vec/
Dword2vec_basic.py89 vocabulary_size = 50000 variable
118 vocabulary, vocabulary_size)
192 tf.random_uniform([vocabulary_size, embedding_size], -1.0, 1.0))
199 [vocabulary_size, embedding_size],
202 nce_biases = tf.Variable(tf.zeros([vocabulary_size]))
217 num_classes=vocabulary_size))
302 for i in xrange(vocabulary_size):
/external/tensorflow/tensorflow/python/feature_column/
Dfeature_column_test.py2324 key='aaa', vocabulary_file='path_to_file', vocabulary_size=3)
2335 key='aaa', vocabulary_file='path_to_file', vocabulary_size=3,
2344 key='aaa', vocabulary_file='path_to_file', vocabulary_size=3,
2356 key='aaa', vocabulary_file=None, vocabulary_size=3)
2361 key='aaa', vocabulary_file='', vocabulary_size=3)
2365 key='aaa', vocabulary_file='file_does_not_exist', vocabulary_size=10)
2379 vocabulary_size=-1)
2383 vocabulary_size=0)
2389 vocabulary_size=self._wire_vocabulary_size + 1)
2402 key='aaa', vocabulary_file='path', vocabulary_size=3,
[all …]
Dfeature_column.py1043 vocabulary_size=None, argument
1129 if vocabulary_size is None:
1134 vocabulary_size = sum(1 for _ in f)
1137 'in the vocabulary_file %s.', vocabulary_size, key, vocabulary_file)
1140 if vocabulary_size < 1:
1154 vocabulary_size=vocabulary_size,
2499 vocab_size=self.vocabulary_size,
2507 return self.vocabulary_size + self.num_oov_buckets
/external/tensorflow/tensorflow/python/keras/_impl/keras/preprocessing/
Dsequence_test.py85 np.arange(3), vocabulary_size=3)
91 np.arange(5), vocabulary_size=5, window_size=1, categorical=True)
Dsequence.py145 vocabulary_size, argument
215 random.randint(1, vocabulary_size - 1)]
/external/tensorflow/tensorflow/examples/udacity/
D6_lstm.ipynb300 "vocabulary_size = len(string.ascii_lowercase) + 1 # [a-z] + ' '\n",
395 " batch = np.zeros(shape=(self._batch_size, vocabulary_size), dtype=np.float)\n",
479 " p = np.zeros(shape=[1, vocabulary_size], dtype=np.float)\n",
485 " b = np.random.uniform(0.0, 1.0, size=[1, vocabulary_size])\n",
522 " ix = tf.Variable(tf.truncated_normal([vocabulary_size, num_nodes], -0.1, 0.1))\n",
526 " fx = tf.Variable(tf.truncated_normal([vocabulary_size, num_nodes], -0.1, 0.1))\n",
530 " cx = tf.Variable(tf.truncated_normal([vocabulary_size, num_nodes], -0.1, 0.1))\n",
534 " ox = tf.Variable(tf.truncated_normal([vocabulary_size, num_nodes], -0.1, 0.1))\n",
541 " w = tf.Variable(tf.truncated_normal([num_nodes, vocabulary_size], -0.1, 0.1))\n",
542 " b = tf.Variable(tf.zeros([vocabulary_size]))\n",
[all …]
D5_word2vec.ipynb249 "vocabulary_size = 50000\n",
253 " count.extend(collections.Counter(words).most_common(vocabulary_size - 1))\n",
434 " tf.random_uniform([vocabulary_size, embedding_size], -1.0, 1.0))\n",
436 " tf.truncated_normal([vocabulary_size, embedding_size],\n",
438 " softmax_biases = tf.Variable(tf.zeros([vocabulary_size]))\n",
446 … labels=train_labels, num_sampled=num_sampled, num_classes=vocabulary_size))\n",
/external/tensorflow/tensorflow/tools/api/golden/
Dtensorflow.keras.preprocessing.sequence.pbtxt13 …argspec: "args=[\'sequence\', \'vocabulary_size\', \'window_size\', \'negative_samples\', \'shuffl…
Dtensorflow.feature_column.pbtxt17 …argspec: "args=[\'key\', \'vocabulary_file\', \'vocabulary_size\', \'num_oov_buckets\', \'default_…
/external/tensorflow/tensorflow/docs_src/tutorials/
Dword2vec.md256 tf.random_uniform([vocabulary_size, embedding_size], -1.0, 1.0))
266 tf.truncated_normal([vocabulary_size, embedding_size],
268 nce_biases = tf.Variable(tf.zeros([vocabulary_size]))
304 num_classes=vocabulary_size))
Drecurrent.md148 # embedding_matrix is a tensor of shape [vocabulary_size, embedding size]
/external/tensorflow/tensorflow/python/estimator/canned/
Ddnn_testing_utils.py864 vocabulary_size=len(vocab_list)),
891 vocabulary_size=len(new_vocab_list)),
897 new_vocab_size=new_occupation.categorical_column.vocabulary_size,
900 old_vocab_size=occupation.categorical_column.vocabulary_size,
Dlinear_testing_utils.py1992 vocabulary_size=len(vocab_list))
2016 vocabulary_size=len(new_vocab_list))
2021 new_vocab_size=new_occupation.vocabulary_size,
2024 old_vocab_size=occupation.vocabulary_size,
/external/tensorflow/tensorflow/docs_src/programmers_guide/
Dembedding.md65 [vocabulary_size, embedding_size])
/external/tensorflow/tensorflow/docs_src/get_started/
Dfeature_columns.md264 vocabulary_size=3)