Home
last modified time | relevance | path

Searched refs:grad_loss (Results 1 – 25 of 57) sorted by relevance

123

/external/tensorflow/tensorflow/contrib/constrained_optimization/python/
Dconstrained_optimizer.py67 grad_loss=None): argument
99 grad_loss=None): argument
131 grad_loss=grad_loss)
152 grad_loss=None): argument
184 grad_loss=grad_loss)
206 grad_loss=None): argument
249 grad_loss=grad_loss)
261 grad_loss=grad_loss)
Dexternal_regret_optimizer.py202 grad_loss=None): argument
273 grad_loss=grad_loss)
288 grad_loss=grad_loss)
Dswap_regret_optimizer.py307 grad_loss=None): argument
388 grad_loss=grad_loss)
403 grad_loss=grad_loss)
/external/tensorflow/tensorflow/contrib/mixed_precision/python/
Dloss_scale_optimizer.py120 grad_loss=None): argument
141 grad_loss=grad_loss)
/external/tensorflow/tensorflow/contrib/optimizer_v2/
Doptimizer_v2.py661 grad_loss=None, argument
712 grad_loss=grad_loss,
731 grad_loss=None, argument
786 grads = tape.gradient(loss_value, var_list, grad_loss)
803 if grad_loss is not None:
804 self._assert_valid_dtypes([grad_loss])
821 grad_ys=grad_loss,
Doptimizer_v2_test.py93 grad_loss = constant_op.constant([42, -42], dtype=dtype)
98 cost, global_step, [var0, var1], grad_loss=grad_loss)
/external/tensorflow/tensorflow/python/training/
Doptimizer.py357 grad_loss=None): argument
402 grad_loss=grad_loss)
418 grad_loss=None): argument
468 grads = tape.gradient(loss_value, var_list, grad_loss)
483 if grad_loss is not None:
484 self._assert_valid_dtypes([grad_loss])
499 loss, var_refs, grad_ys=grad_loss,
Doptimizer_test.py98 grad_loss = constant_op.constant([42, -42], dtype=dtype)
103 cost, global_step, [var0, var1], grad_loss=grad_loss)
/external/tensorflow/tensorflow/tools/api/golden/v1/
Dtensorflow.train.-optimizer.pbtxt28 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
44 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-gradient-descent-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-proximal-adagrad-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-adagrad-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-proximal-gradient-descent-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-momentum-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-adadelta-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-adagrad-d-a-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-r-m-s-prop-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-adam-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-ftrl-optimizer.pbtxt29 …gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None…
45 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
Dtensorflow.train.-sync-replicas-optimizer.pbtxt57 …', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None…
/external/tensorflow/tensorflow/python/keras/mixed_precision/experimental/
Dloss_scale_optimizer.py72 def _compute_gradients(self, loss, var_list, grad_loss=None): argument
75 grad_loss)
/external/tensorflow/tensorflow/contrib/opt/python/training/
Dweight_decay_optimizers.py99 name=None, grad_loss=None, decay_var_list=None): argument
135 grad_loss=grad_loss)
/external/tensorflow/tensorflow/python/keras/optimizer_v2/
Doptimizer_v2.py268 def minimize(self, loss, var_list, grad_loss=None, name=None): argument
298 loss, var_list=var_list, grad_loss=grad_loss)
302 def _compute_gradients(self, loss, var_list, grad_loss=None): argument
331 grads = tape.gradient(loss_value, var_list, grad_loss)
/external/tensorflow/tensorflow/cc/gradients/
Dnn_grad.cc90 auto grad_loss = grad_inputs[0]; in SoftmaxCrossEntropyWithLogitsGrad() local
93 auto grad = BroadcastMul(scope, grad_loss, softmax_grad); in SoftmaxCrossEntropyWithLogitsGrad()
110 grad_outputs->push_back(BroadcastMul(scope, grad_loss, minus_log_softmax)); in SoftmaxCrossEntropyWithLogitsGrad()
/external/tensorflow/tensorflow/python/ops/
Dctc_ops.py182 def _CTCLossGrad(op, grad_loss, _): argument
204 return [_BroadcastMul(grad_loss, grad_without_gradient), None, None, None]
576 def _ctc_loss_grad(op, grad_loss, _): argument
578 grad = [array_ops.reshape(grad_loss, [1, -1, 1]) * grad]

123