Searched refs:_opt (Results 1 – 15 of 15) sorted by relevance
/external/tensorflow/tensorflow/python/tpu/ |
D | tpu_optimizer.py | 57 self._opt = opt 140 return self._opt.compute_gradients(loss, var_list=var_list, **kwargs) 171 return self._opt.apply_gradients(summed_grads_and_vars, global_step, name) 185 return self._opt.get_slot(*args, **kwargs) 199 return self._opt.get_slot_names(*args, **kwargs) 203 return self._opt.variables()
|
/external/tensorflow/tensorflow/contrib/opt/python/training/ |
D | multitask_optimizer_wrapper.py | 103 self._opt = opt 107 fn = getattr(self._opt, name) 108 wrapper = _get_wrapper(fn, self._opt) 109 setattr(self._opt, name, wrapper) 112 return getattr(self._opt, name)
|
D | drop_stale_gradient_optimizer.py | 60 self._opt = opt 70 return self._opt.compute_gradients(loss, *args, **kwargs) 73 return self._opt.get_slot(*args, **kwargs) 76 return self._opt.get_slot_names(*args, **kwargs) 89 [self._opt.apply_gradients(
|
D | variable_clipping_optimizer.py | 76 self._opt = opt 84 return self._opt.compute_gradients(*args, **kwargs) 87 return self._opt.get_slot(*args, **kwargs) 90 return self._opt.get_slot_names(*args, **kwargs) 94 update_op = self._opt.apply_gradients(
|
D | model_average_optimizer.py | 133 self._opt = opt 147 self._opt._prepare() # pylint:disable=protected-access 161 return self._opt.compute_gradients(*args, **kwargs) 215 apply_updates = self._opt.apply_gradients(grads_and_vars)
|
D | agn_optimizer.py | 122 self._opt = optimizer 132 self._opt._prepare() 166 local_update_op = self._opt.apply_gradients(grads_and_vars) 185 apply_global_op = self._opt.apply_gradients(
|
D | elastic_average_optimizer.py | 190 self._opt = opt 202 self._rho = self._moving_rate / self._opt._learning_rate 211 self._opt._prepare() 263 return self._opt.compute_gradients(total_loss, var_list, gate_gradients, 290 apply_updates = self._opt.apply_gradients(grads_and_vars)
|
/external/tensorflow/tensorflow/python/training/ |
D | sync_replicas_optimizer.py | 189 self._opt = opt 224 return self._opt.compute_gradients(*args, **kwargs) 310 update_op = self._opt.apply_gradients(aggregated_grads_and_vars, 393 return self._opt.get_slot(*args, **kwargs) 404 return self._opt.variables() 418 return self._opt.get_slot_names(*args, **kwargs)
|
D | sync_replicas_optimizer_test.py | 295 beta1_power, beta2_power = opt._opt._get_beta_accumulators()
|
/external/tensorflow/tensorflow/contrib/mixed_precision/python/ |
D | loss_scale_optimizer.py | 111 self._opt = opt 135 grads_and_vars = self._opt.compute_gradients( 155 return self._opt.apply_gradients(grads_and_vars, global_step, name)
|
/external/newfs_msdos/ |
D | newfs_msdos.c | 275 #define AOPT(_opt, _type, _name, _min, _desc) { _opt, _desc }, in usage() argument
|
D | mkfs_msdos.h | 64 #define AOPT(_opt, _type, _name, _min, _desc) _type _name; argument
|
/external/tensorflow/tensorflow/contrib/tpu/python/tpu/ |
D | keras_support.py | 342 self._opt = opt 343 logging.info('KerasCrossShard: %s %s', self._opt, self._opt.weights) 346 self._opt.get_gradients = self.get_gradients 347 return self._opt.get_updates(loss, params) 355 return self._opt.get_weights() 358 return self._opt.get_config() 362 return getattr(self._opt, key)
|
/external/brotli/research/ |
D | README.md | 7 ### find\_opt\_references
|
/external/capstone/bindings/ocaml/ |
D | ocaml.c | 862 CAMLprim value ocaml_option(value _handle, value _opt, value _value) in ocaml_option() argument 864 CAMLparam3(_handle, _opt, _value); in ocaml_option() 868 switch (Int_val(_opt)) { in ocaml_option()
|