/external/libcxx/include/ |
D | __split_buffer | 424 __split_buffer<value_type, __alloc_rr&> __t(__n, 0, __alloc()); 425 __t.__construct_at_end(move_iterator<pointer>(__begin_), 427 _VSTD::swap(__first_, __t.__first_); 428 _VSTD::swap(__begin_, __t.__begin_); 429 _VSTD::swap(__end_, __t.__end_); 430 _VSTD::swap(__end_cap(), __t.__end_cap()); 444 __split_buffer<value_type, __alloc_rr&> __t(size(), 0, __alloc()); 445 __t.__construct_at_end(move_iterator<pointer>(__begin_), 447 __t.__end_ = __t.__begin_ + (__end_ - __begin_); 448 _VSTD::swap(__first_, __t.__first_); [all …]
|
D | tuple | 219 explicit __tuple_leaf(_Tp&& __t) _NOEXCEPT_((is_nothrow_constructible<_Hp, _Tp>::value)) 220 : value(_VSTD::forward<_Tp>(__t)) 235 explicit __tuple_leaf(integral_constant<int, 0>, const _Alloc&, _Tp&& __t) 236 : value(_VSTD::forward<_Tp>(__t)) 249 explicit __tuple_leaf(integral_constant<int, 1>, const _Alloc& __a, _Tp&& __t) 250 : value(allocator_arg_t(), __a, _VSTD::forward<_Tp>(__t)) 263 explicit __tuple_leaf(integral_constant<int, 2>, const _Alloc& __a, _Tp&& __t) 264 : value(_VSTD::forward<_Tp>(__t), __a) 275 __tuple_leaf(const __tuple_leaf& __t) = default; 276 __tuple_leaf(__tuple_leaf&& __t) = default; [all …]
|
D | thread | 291 thread(thread&& __t) _NOEXCEPT : __t_(__t.__t_) {__t.__t_ = 0;} 292 thread& operator=(thread&& __t) _NOEXCEPT; 296 void swap(thread& __t) _NOEXCEPT {_VSTD::swap(__t_, __t.__t_);} 335 __thread_execute(tuple<_Fp, _Args...>& __t, __tuple_indices<_Indices...>) 337 __invoke(_VSTD::move(_VSTD::get<0>(__t)), _VSTD::move(_VSTD::get<_Indices>(__t))...); 395 thread::operator=(thread&& __t) _NOEXCEPT 399 __t_ = __t.__t_; 400 __t.__t_ = 0; 437 sleep_until(const chrono::time_point<_Clock, _Duration>& __t) 443 while (_Clock::now() < __t) [all …]
|
D | numeric | 117 typename iterator_traits<_InputIterator>::value_type __t(*__first); 118 *__result = __t; 121 __t = __t + *__first; 122 *__result = __t; 136 typename iterator_traits<_InputIterator>::value_type __t(*__first); 137 *__result = __t; 140 __t = __binary_op(__t, *__first); 141 *__result = __t;
|
D | __tree | 660 {__tree_iterator __t(*this); ++(*this); return __t;} 670 {__tree_iterator __t(*this); --(*this); return __t;} 756 {__tree_const_iterator __t(*this); ++(*this); return __t;} 775 {__tree_const_iterator __t(*this); --(*this); return __t;} 899 __tree(const __tree& __t); 900 __tree& operator=(const __tree& __t); 906 __tree(__tree&& __t) 910 __tree(__tree&& __t, const allocator_type& __a); 911 __tree& operator=(__tree&& __t) 935 void swap(__tree& __t) [all …]
|
D | __mutex_base | 122 unique_lock(mutex_type& __m, const chrono::time_point<_Clock, _Duration>& __t) 123 : __m_(&__m), __owns_(__m.try_lock_until(__t)) {} 165 bool try_lock_until(const chrono::time_point<_Clock, _Duration>& __t); 233 unique_lock<_Mutex>::try_lock_until(const chrono::time_point<_Clock, _Duration>& __t) 239 __owns_ = __m_->try_lock_until(__t); 294 const chrono::time_point<_Clock, _Duration>& __t); 299 const chrono::time_point<_Clock, _Duration>& __t, 350 const chrono::time_point<_Clock, _Duration>& __t) 353 wait_for(__lk, __t - _Clock::now()); 354 return _Clock::now() < __t ? cv_status::no_timeout : cv_status::timeout; [all …]
|
D | functional | 506 auto operator()(_T1&& __t, _T2&& __u) const 507 _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u))) 508 -> decltype (_VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u)) 509 { return _VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u); } 533 auto operator()(_T1&& __t, _T2&& __u) const 534 _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u))) 535 -> decltype (_VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u)) 536 { return _VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u); } 560 auto operator()(_T1&& __t, _T2&& __u) const 561 _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) * _VSTD::forward<_T2>(__u))) [all …]
|
D | exception | 188 _LIBCPP_INLINE_VISIBILITY explicit __nested(const _Tp& __t) : _Tp(__t) {} 195 throw_with_nested(_Tp&& __t, typename enable_if< 201 throw_with_nested (_Tp& __t, typename enable_if< 207 throw __nested<typename remove_reference<_Tp>::type>(_VSTD::forward<_Tp>(__t)); 215 throw_with_nested(_Tp&& __t, typename enable_if< 221 throw_with_nested (_Tp& __t, typename enable_if< 227 throw _VSTD::forward<_Tp>(__t);
|
D | valarray | 463 explicit __scalar_expr(const value_type& __t, size_t __s) : __t_(__t), __s_(__s) {} 1220 value_type* __t = __vp_; 1222 for (size_t __n = __size_; __n; --__n, __t += __stride_, __s += __sa.__stride_) 1223 *__t = *__s; 1237 value_type* __t = __vp_; 1238 for (size_t __i = 0; __i < __size_; ++__i, __t += __stride_) 1239 *__t = __v[__i]; 1252 value_type* __t = __vp_; 1253 for (size_t __i = 0; __i < __size_; ++__i, __t += __stride_) 1254 *__t *= __v[__i]; [all …]
|
D | condition_variable | 140 const chrono::time_point<_Clock, _Duration>& __t); 145 const chrono::time_point<_Clock, _Duration>& __t, 210 const chrono::time_point<_Clock, _Duration>& __t) 217 return __cv_.wait_until(__lk, __t); 224 const chrono::time_point<_Clock, _Duration>& __t, 228 if (wait_until(__lock, __t) == cv_status::timeout)
|
D | __functional_base | 72 auto operator()(_T1&& __t, _T2&& __u) const 73 _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u))) 74 -> decltype (_VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u)) 75 { return _VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u); } 567 ref(_Tp& __t) _NOEXCEPT 569 return reference_wrapper<_Tp>(__t); 575 ref(reference_wrapper<_Tp> __t) _NOEXCEPT 577 return ref(__t.get()); 583 cref(const _Tp& __t) _NOEXCEPT 585 return reference_wrapper<const _Tp>(__t); [all …]
|
D | mutex | 237 bool try_lock_until(const chrono::time_point<_Clock, _Duration>& __t); 243 timed_mutex::try_lock_until(const chrono::time_point<_Clock, _Duration>& __t) 247 bool no_timeout = _Clock::now() < __t; 249 no_timeout = __cv_.wait_until(__lk, __t) == cv_status::no_timeout; 280 bool try_lock_until(const chrono::time_point<_Clock, _Duration>& __t); 286 recursive_timed_mutex::try_lock_until(const chrono::time_point<_Clock, _Duration>& __t) 298 bool no_timeout = _Clock::now() < __t; 300 no_timeout = __cv_.wait_until(lk, __t) == cv_status::no_timeout;
|
D | complex | 530 complex<_Tp> __t(__x); 531 __t += __y; 532 return __t; 540 complex<_Tp> __t(__x); 541 __t += __y; 542 return __t; 550 complex<_Tp> __t(__y); 551 __t += __x; 552 return __t; 560 complex<_Tp> __t(__x); [all …]
|
D | forward_list | 268 __forward_list_iterator __t(*this); 270 return __t; 342 __forward_list_const_iterator __t(*this); 344 return __t; 1489 __node_pointer __t = __f2; 1490 while (__t->__next_ != nullptr && 1491 __comp(__t->__next_->__value_, __f1->__value_)) 1492 __t = __t->__next_; 1494 __f2 = __t->__next_; 1495 __t->__next_ = __f1; [all …]
|
D | random | 1997 _UIntType __t; 1998 __is >> __t; 2000 __x.__x_ = __t; 2013 _UIntType __b, size_t __t, _UIntType __c, size_t __l, _UIntType __f> 2055 _UIntType __b, size_t __t, _UIntType __c, size_t __l, _UIntType __f> 2074 static_assert(__t <= __w, "mersenne_twister_engine invalid parameters"); 2097 static _LIBCPP_CONSTEXPR const size_t tempering_t = __t; 2221 _UIntType __b, size_t __t, _UIntType __c, size_t __l, _UIntType __f> 2223 …mersenne_twister_engine<_UIntType, __w, __n, __m, __r, __a, __u, __d, __s, __b, __t, __c, __l, __f… 2227 _UIntType __b, size_t __t, _UIntType __c, size_t __l, _UIntType __f> [all …]
|
D | algorithm | 2247 typename iterator_traits<_InputIterator>::value_type __t(*__first); 2248 *__result = __t; 2252 if (!__pred(__t, *__first)) 2254 __t = *__first; 2255 *__result = __t; 2435 _Integral __t = __x % __y; 2437 __y = __t; 2459 value_type __t(_VSTD::move(*--__p)); 2472 *__p1 = _VSTD::move(__t); 2599 min(initializer_list<_Tp> __t, _Compare __comp) [all …]
|
D | __bit_reference | 89 bool __t = __x; 91 __y = __t; 99 bool __t = __x; 101 __y = __t; 109 bool __t = __x; 111 __y = __t; 119 bool __t = __x; 121 __y = __t; 1180 __bit_iterator __t(*this); 1181 __t += __n; [all …]
|
D | bitset | 228 __storage_type __t[sizeof(unsigned long long) / sizeof(__storage_type)]; 229 for (size_t __i = 0; __i < sizeof(__t)/sizeof(__t[0]); ++__i, __v >>= __bits_per_word) 230 __t[__i] = static_cast<__storage_type>(__v); 231 _VSTD::copy(__t, __t + sizeof(__t)/sizeof(__t[0]), __first_); 232 …_VSTD::fill(__first_ + sizeof(__t)/sizeof(__t[0]), __first_ + sizeof(__first_)/sizeof(__first_[0]),
|
D | __functional_03 | 657 __base* __t = (__base*)&__tempbuf; 658 __f_->__clone(__t); 665 __t->__clone((__base*)&__f.__buf_); 666 __t->destroy(); 954 __base* __t = (__base*)&__tempbuf; 955 __f_->__clone(__t); 962 __t->__clone((__base*)&__f.__buf_); 963 __t->destroy(); 1251 __base* __t = (__base*)&__tempbuf; 1252 __f_->__clone(__t); [all …]
|
/external/libcxx/include/experimental/ |
D | tuple | 60 decltype(auto) __apply_tuple_impl(_Fn && __f, _Tuple && __t, 64 _VSTD::get<_Id>(_VSTD::forward<_Tuple>(__t))... 70 decltype(auto) apply(_Fn && __f, _Tuple && __t) { 72 _VSTD::forward<_Fn>(__f), _VSTD::forward<_Tuple>(__t),
|
/external/valgrind/drd/tests/ |
D | std_thread.cpp | 30 std::thread::_Impl_base* __t = static_cast<std::thread::_Impl_base*>(__p); in execute_native_thread_routine() local 32 __local.swap(__t->_M_this_ptr); in execute_native_thread_routine() 35 __t->_M_run(); in execute_native_thread_routine()
|
D | std_thread2.cpp | 35 std::thread::_Impl_base* __t = static_cast<std::thread::_Impl_base*>(__p); in execute_native_thread_routine() local 37 __local.swap(__t->_M_this_ptr); in execute_native_thread_routine() 40 __t->_M_run(); in execute_native_thread_routine()
|
/external/clang/test/CodeGen/ |
D | mips-constraint-regs.c | 14 int __t; in main() local 17 : "=c" (__t) in main()
|
/external/clang/www/ |
D | libstdc++4.4-clang0x.patch | 28 - forward(typename std::identity<_Tp>::type&& __t) 29 + forward(typename std::remove_reference<_Tp>::type& __t) 31 + { return static_cast<_Tp&&>(__t); } 33 { return __t; } 38 + forward(typename std::remove_reference<_Tp>::type&& __t) 44 + return static_cast<_Tp&&>(__t); 46 + return __t; 52 move(_Tp&& __t) 54 + { return static_cast<typename std::remove_reference<_Tp>::type&&>(__t); } 56 { return __t; } [all …]
|
/external/blktrace/ |
D | blkparse.c | 801 struct trace *__t; in trace_rb_insert() local 806 __t = rb_entry(parent, struct trace, rb_node); in trace_rb_insert() 808 if (t->bit->time < __t->bit->time) in trace_rb_insert() 810 else if (t->bit->time > __t->bit->time) in trace_rb_insert() 812 else if (t->bit->device < __t->bit->device) in trace_rb_insert() 814 else if (t->bit->device > __t->bit->device) in trace_rb_insert() 816 else if (t->bit->sequence < __t->bit->sequence) in trace_rb_insert() 861 struct trace *__t; in trace_rb_find() local 864 __t = rb_entry(n, struct trace, rb_node); in trace_rb_find() 867 if (device < __t->bit->device) in trace_rb_find() [all …]
|