/external/tensorflow/tensorflow/python/training/ |
D | learning_rate_decay.py | 29 def exponential_decay(learning_rate, argument 159 def polynomial_decay(learning_rate, argument 259 def natural_exp_decay(learning_rate, argument 343 def inverse_time_decay(learning_rate, argument 429 def cosine_decay(learning_rate, global_step, decay_steps, alpha=0.0, name=None): argument 488 def cosine_decay_restarts(learning_rate, argument 561 def linear_cosine_decay(learning_rate, argument 643 def noisy_linear_cosine_decay(learning_rate, argument
|
D | adagrad_test.py | 53 learning_rate = lambda: 3.0 function 326 learning_rate = lambda: 3.0 function
|
D | gradient_descent.py | 34 def __init__(self, learning_rate, use_locking=False, name="GradientDescent"): argument
|
D | proximal_gradient_descent.py | 38 def __init__(self, learning_rate, l1_regularization_strength=0.0, argument
|
/external/tensorflow/tensorflow/contrib/timeseries/python/timeseries/ |
D | test_utils.py | 104 learning_rate=0.1, ignore_params_fn=lambda _: (), argument 172 learning_rate=0.1, rtol=0.2, atol=0.1, train_loss_tolerance_coeff=0.99, argument 257 learning_rate=0.1, argument
|
/external/tensorflow/tensorflow/contrib/opt/python/training/ |
D | addsign_test.py | 63 learning_rate=0.1, argument 166 learning_rate=0.1, argument
|
D | powersign_test.py | 64 learning_rate=0.1, argument 171 learning_rate=0.1, argument
|
D | weight_decay_optimizers.py | 304 def __init__(self, weight_decay, learning_rate, momentum, argument 359 def __init__(self, weight_decay, learning_rate=0.001, beta1=0.9, beta2=0.999, argument 400 learning_rate=1.0, argument
|
D | reg_adagrad_optimizer.py | 47 learning_rate, argument
|
/external/tensorflow/tensorflow/python/keras/mixed_precision/experimental/ |
D | loss_scale_optimizer.py | 101 def learning_rate(self): member in LossScaleOptimizer 105 def learning_rate(self, lr): member in LossScaleOptimizer
|
/external/tensorflow/tensorflow/contrib/training/python/training/ |
D | sgdr_learning_rate_decay.py | 28 def sgdr_decay(learning_rate, global_step, initial_period_steps, argument
|
/external/tensorflow/tensorflow/contrib/boosted_trees/proto/ |
D | learner.proto | 42 float learning_rate = 1; field 88 float learning_rate = 3; field
|
/external/tensorflow/tensorflow/examples/tutorials/mnist/ |
D | mnist.py | 100 def training(loss, learning_rate): argument
|
/external/tensorflow/tensorflow/python/keras/optimizer_v2/ |
D | adamax.py | 45 learning_rate=0.001, argument
|
D | ftrl.py | 57 learning_rate, argument
|
D | adadelta.py | 60 learning_rate=0.001, argument
|
D | nadam.py | 64 learning_rate=0.001, argument
|
D | adam.py | 48 learning_rate=0.001, argument
|
D | gradient_descent.py | 63 learning_rate=0.001, argument
|
D | rmsprop.py | 64 learning_rate=0.001, argument
|
/external/tensorflow/tensorflow/compiler/tests/ |
D | powersign_test.py | 61 learning_rate=0.1, argument
|
D | addsign_test.py | 60 learning_rate=0.1, argument
|
/external/tensorflow/tensorflow/python/tpu/ |
D | tpu_embedding.py | 122 def __init__(self, learning_rate, use_gradient_accumulation): argument 130 def __init__(self, learning_rate, initial_accumulator=0.1, argument 152 def __init__(self, learning_rate, argument 203 def __init__(self, learning_rate): argument
|
/external/tensorflow/tensorflow/contrib/optimizer_v2/ |
D | gradient_descent.py | 30 def __init__(self, learning_rate, use_locking=False, name="GradientDescent"): argument
|
/external/tensorflow/tensorflow/contrib/learn/python/learn/estimators/ |
D | dynamic_rnn_estimator.py | 394 learning_rate=None, argument 564 learning_rate=0.1, argument
|