/external/compiler-rt/lib/sanitizer_common/ |
D | sanitizer_syscall_linux_aarch64.inc | 16 static uptr __internal_syscall(u64 nr) { 17 register u64 x8 asm("x8") = nr; 18 register u64 x0 asm("x0"); 28 static uptr __internal_syscall(u64 nr, u64 arg1) { 29 register u64 x8 asm("x8") = nr; 30 register u64 x0 asm("x0") = arg1; 38 (__internal_syscall)(n, (u64)(a1)) 40 static uptr __internal_syscall(u64 nr, u64 arg1, long arg2) { 41 register u64 x8 asm("x8") = nr; 42 register u64 x0 asm("x0") = arg1; [all …]
|
D | sanitizer_syscall_linux_x86_64.inc | 16 static uptr internal_syscall(u64 nr) { 17 u64 retval; 24 static uptr internal_syscall(u64 nr, T1 arg1) { 25 u64 retval; 26 asm volatile("syscall" : "=a"(retval) : "a"(nr), "D"((u64)arg1) : 32 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2) { 33 u64 retval; 34 asm volatile("syscall" : "=a"(retval) : "a"(nr), "D"((u64)arg1), 35 "S"((u64)arg2) : "rcx", "r11", "memory", "cc"); 40 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2, T3 arg3) { [all …]
|
D | sanitizer_lfstack.h | 37 u64 cmp = atomic_load(&head_, memory_order_relaxed); in Push() 39 u64 cnt = (cmp & kCounterMask) + kCounterInc; in Push() 40 u64 xch = (u64)(uptr)p | cnt; in Push() 49 u64 cmp = atomic_load(&head_, memory_order_acquire); in Pop() 55 u64 cnt = (cmp & kCounterMask); in Pop() 56 u64 xch = (u64)(uptr)nxt | cnt; in Pop() 65 static const u64 kPtrMask = ((u64)-1) >> kCounterBits; 66 static const u64 kCounterMask = ~kPtrMask; 67 static const u64 kCounterInc = kPtrMask + 1;
|
D | sanitizer_allocator.cc | 90 const u64 kBlockMagic = 0x6A6CB03ABCEBC041ull; 93 if (size + sizeof(u64) < size) in InternalAlloc() 95 void *p = RawInternalAlloc(size + sizeof(u64), cache); in InternalAlloc() 98 ((u64*)p)[0] = kBlockMagic; in InternalAlloc() 99 return (char*)p + sizeof(u64); in InternalAlloc() 105 addr = (char*)addr - sizeof(u64); in InternalFree() 106 CHECK_EQ(kBlockMagic, ((u64*)addr)[0]); in InternalFree() 107 ((u64*)addr)[0] = 0; in InternalFree()
|
/external/compiler-rt/lib/tsan/rtl/ |
D | tsan_rtl.h | 81 u64 v1, u64 v2); 83 const u64 kShadowRodata = (u64)-1; // .rodata shadow marker 93 FastState(u64 tid, u64 epoch) { in FastState() 101 explicit FastState(u64 x) in FastState() 105 u64 raw() const { in raw() 109 u64 tid() const { in tid() 110 u64 res = (x_ & ~kIgnoreBit) >> kTidShift; in tid() 114 u64 TidWithIgnore() const { in TidWithIgnore() 115 u64 res = x_ >> kTidShift; in TidWithIgnore() 119 u64 epoch() const { in epoch() [all …]
|
D | tsan_mutexset.h | 27 u64 id; 28 u64 epoch; 35 void Add(u64 id, bool write, u64 epoch); 36 void Del(u64 id, bool write); 37 void Remove(u64 id); // Removes the mutex completely (if it's destroyed). 60 void MutexSet::Add(u64 id, bool write, u64 epoch) {} in Add() 61 void MutexSet::Del(u64 id, bool write) {} in Del() 62 void MutexSet::Remove(u64 id) {} in Remove()
|
D | tsan_mutexset.cc | 25 void MutexSet::Add(u64 id, bool write, u64 epoch) { in Add() 36 u64 minepoch = (u64)-1; in Add() 37 u64 mini = (u64)-1; in Add() 55 void MutexSet::Del(u64 id, bool write) { in Del() 65 void MutexSet::Remove(u64 id) { in Remove()
|
D | tsan_sync.h | 33 u64 uid; // Globally unique id. 36 u64 last_lock; 49 void Init(ThreadState *thr, uptr pc, uptr addr, u64 uid); 52 u64 GetId() const { in GetId() 54 return GetLsb((u64)addr | (uid << 47), 61); in GetId() 56 bool CheckId(u64 uid) const { in CheckId() 60 static uptr SplitId(u64 id, u64 *uid) { in SplitId()
|
D | tsan_rtl.cc | 110 ThreadState::ThreadState(Context *ctx, int tid, int unique_id, u64 epoch, in ThreadState() 151 const u64 kMs2Ns = 1000 * 1000; in BackgroundThread() 172 u64 last_flush = NanoTime(); in BackgroundThread() 178 u64 now = NanoTime(); in BackgroundThread() 194 (u64)rss >> 20, (u64)last_rss >> 20, (u64)limit >> 20); in BackgroundThread() 199 VPrintf(1, "ThreadSanitizer: memory flushed RSS=%llu\n", (u64)rss>>20); in BackgroundThread() 210 u64 last = atomic_load(&ctx->last_symbolize_time_ns, in BackgroundThread() 538 Shadow LoadShadow(u64 *p) { in LoadShadow() 539 u64 raw = atomic_load((atomic_uint64_t*)p, memory_order_relaxed); in LoadShadow() 544 void StoreShadow(u64 *sp, u64 s) { in StoreShadow() [all …]
|
D | tsan_clock.h | 22 u64 epoch : kClkBits; 23 u64 reused : 64 - kClkBits; 53 u64 get(unsigned tid) const { in get() 87 u64 get(unsigned tid) const { in get() 92 void set(unsigned tid, u64 v); 94 void set(u64 v) { in set() 119 u64 last_acquire_;
|
D | tsan_defs.h | 124 T RoundUp(T p, u64 align) { in RoundUp() 126 return (T)(((u64)p + align - 1) & ~(align - 1)); in RoundUp() 130 T RoundDown(T p, u64 align) { in RoundDown() 132 return (T)((u64)p & ~(align - 1)); in RoundDown() 138 return (T)((u64)v & ((1ull << bits) - 1)); in GetLsb() 142 u64 hash[2]; 157 u64 siz;
|
/external/compiler-rt/lib/tsan/tests/unit/ |
D | tsan_sync_test.cc | 22 u64 block[1] = {}; // fake malloc block in TEST() 23 m->AllocBlock(thr, 0, (uptr)&block[0], 1 * sizeof(u64)); in TEST() 26 EXPECT_EQ(mb->siz, 1 * sizeof(u64)); in TEST() 29 EXPECT_EQ(sz, 1 * sizeof(u64)); in TEST() 37 u64 block[4] = {}; // fake malloc block in TEST() 38 m->AllocBlock(thr, 0, (uptr)&block[0], 1 * sizeof(u64)); in TEST() 39 m->AllocBlock(thr, 0, (uptr)&block[1], 3 * sizeof(u64)); in TEST() 41 EXPECT_EQ(mb1->siz, 1 * sizeof(u64)); in TEST() 43 EXPECT_EQ(mb2->siz, 3 * sizeof(u64)); in TEST() 44 m->FreeRange(thr, 0, (uptr)&block[0], 4 * sizeof(u64)); in TEST() [all …]
|
D | tsan_shadow_test.cc | 21 EXPECT_EQ(s.tid(), (u64)11); in TEST() 22 EXPECT_EQ(s.epoch(), (u64)22); in TEST() 26 EXPECT_EQ(s.addr0(), (u64)0); in TEST() 27 EXPECT_EQ(s.size(), (u64)1); in TEST() 31 EXPECT_EQ(s.epoch(), (u64)23); in TEST() 33 EXPECT_EQ(s.epoch(), (u64)24); in TEST() 65 u64 aligned_data[4]; in TEST()
|
/external/squashfs-tools/kernel/fs/squashfs/ |
D | squashfs.h | 36 extern int squashfs_read_data(struct super_block *, void **, u64, int, u64 *, 43 struct squashfs_cache *, u64, int); 46 extern int squashfs_read_metadata(struct super_block *, void *, u64 *, 49 u64, int); 51 u64, int); 52 extern int squashfs_read_table(struct super_block *, void *, u64, int); 55 extern __le64 *squashfs_read_inode_lookup_table(struct super_block *, u64, 59 extern int squashfs_frag_lookup(struct super_block *, unsigned int, u64 *); 61 u64, unsigned int); 65 extern __le64 *squashfs_read_id_index_table(struct super_block *, u64,
|
D | squashfs_fs_i.h | 27 u64 start; 31 u64 fragment_block; 34 u64 block_list_start; 37 u64 dir_idx_start;
|
/external/compiler-rt/lib/asan/ |
D | asan_mapping.h | 113 static const u64 kDefaultShadowScale = 3; 114 static const u64 kDefaultShadowOffset32 = 1ULL << 29; // 0x20000000 115 static const u64 kDefaultShadowOffset64 = 1ULL << 44; 116 static const u64 kDefaultShort64bitShadowOffset = 0x7FFF8000; // < 2G. 117 static const u64 kIosShadowOffset32 = 1ULL << 30; // 0x40000000 118 static const u64 kIosShadowOffset64 = 0x130000000; 119 static const u64 kIosSimShadowOffset32 = 1ULL << 30; 120 static const u64 kIosSimShadowOffset64 = kDefaultShadowOffset64; 121 static const u64 kAArch64_ShadowOffset64 = 1ULL << 36; 122 static const u64 kMIPS32_ShadowOffset32 = 0x0aaa0000; [all …]
|
/external/v8/test/unittests/base/ |
D | division-by-constant-unittest.cc | 43 static M64 u64(uint64_t d) { return UnsignedDivisionByConstant<uint64_t>(d); } in u64() function 116 EXPECT_EQ(M64(0x0000000000000000ULL, 0, true), u64(1)); in TEST() 118 EXPECT_EQ(M64(1ULL << (64 - k), 0, false), u64(1ULL << k)); in TEST() 120 EXPECT_EQ(M64(0xAAAAAAAAAAAAAAABULL, 1, false), u64(3)); in TEST() 121 EXPECT_EQ(M64(0xCCCCCCCCCCCCCCCDULL, 2, false), u64(5)); in TEST() 122 EXPECT_EQ(M64(0xAAAAAAAAAAAAAAABULL, 2, false), u64(6)); in TEST() 123 EXPECT_EQ(M64(0x2492492492492493ULL, 3, true), u64(7)); in TEST() 124 EXPECT_EQ(M64(0xE38E38E38E38E38FULL, 3, false), u64(9)); in TEST() 125 EXPECT_EQ(M64(0xCCCCCCCCCCCCCCCDULL, 3, false), u64(10)); in TEST() 126 EXPECT_EQ(M64(0x2E8BA2E8BA2E8BA3ULL, 1, false), u64(11)); in TEST() [all …]
|
/external/wpa_supplicant_8/src/utils/ |
D | common.h | 80 typedef UINT64 u64; typedef 92 typedef unsigned long long u64; typedef 109 typedef uint64_t u64; typedef 173 #define le_to_host64(n) ((__force u64) (le64) (n)) 174 #define host_to_le64(n) ((__force le64) (u64) (n)) 175 #define be_to_host64(n) bswap_64((__force u64) (be64) (n)) 263 static inline u64 WPA_GET_BE64(const u8 *a) in WPA_GET_BE64() 265 return (((u64) a[0]) << 56) | (((u64) a[1]) << 48) | in WPA_GET_BE64() 266 (((u64) a[2]) << 40) | (((u64) a[3]) << 32) | in WPA_GET_BE64() 267 (((u64) a[4]) << 24) | (((u64) a[5]) << 16) | in WPA_GET_BE64() [all …]
|
/external/llvm/test/CodeGen/NVPTX/ |
D | local-stack-frame.ll | 10 ; PTX64: mov.u64 %SPL, __local_depot{{[0-9]+}}; 11 ; PTX64: cvta.local.u64 %SP, %SPL; 25 ; PTX64: mov.u64 %SPL, __local_depot{{[0-9]+}}; 26 ; PTX64: cvta.local.u64 %SP, %SPL; 28 ; PTX64: add.u64 %rd[[SP_REG:[0-9]+]], %SPL, 0; 47 ; PTX64: mov.u64 %SPL, __local_depot{{[0-9]+}}; 48 ; PTX64-NOT: cvta.local.u64 %SP, %SPL; 50 ; PTX64: add.u64 %rd{{[0-9]+}}, %SPL, 0; 67 ; PTX64: cvta.local.u64 %SP, %SPL; 68 ; PTX64: add.u64 {{%rd[0-9]+}}, %SP, 0; [all …]
|
/external/llvm/test/MC/ARM/ |
D | neon-shift-encoding.s | 7 vshl.u64 d16, d17, d16 15 vshl.u64 q8, q9, q8 24 @ CHECK: vshl.u64 d16, d17, d16 @ encoding: [0xa1,0x04,0x70,0xf3] 32 @ CHECK: vshl.u64 q8, q9, q8 @ encoding: [0xe2,0x04,0x70,0xf3] 42 vshr.u64 d16, d16, #63 46 vshr.u64 q8, q8, #63 59 @ CHECK: vshr.u64 d16, d16, #63 @ encoding: [0xb0,0x00,0xc1,0xf3] 63 @ CHECK: vshr.u64 q8, q8, #63 @ encoding: [0xf0,0x00,0xc1,0xf3] 77 vshr.u64 d16, #63 81 vshr.u64 q8, #63 [all …]
|
D | neon-satshift-encoding.s | 17 @ CHECK: vqshl.u64 d16, d16, d17 @ encoding: [0xb0,0x04,0x71,0xf3] 18 vqshl.u64 d16, d16, d17 33 @ CHECK: vqshl.u64 q8, q8, q9 @ encoding: [0xf0,0x04,0x72,0xf3] 34 vqshl.u64 q8, q8, q9 49 @ CHECK: vqshl.u64 d16, d16, #63 @ encoding: [0xb0,0x07,0xff,0xf3] 50 vqshl.u64 d16, d16, #63 73 @ CHECK: vqshl.u64 q8, q8, #63 @ encoding: [0xf0,0x07,0xff,0xf3] 74 vqshl.u64 q8, q8, #63 97 @ CHECK: vqrshl.u64 d16, d16, d17 @ encoding: [0xb0,0x05,0x71,0xf3] 98 vqrshl.u64 d16, d16, d17 [all …]
|
D | neont2-shift-encoding.s | 11 @ CHECK: vshl.u64 d16, d17, d16 @ encoding: [0x70,0xff,0xa1,0x04] 12 vshl.u64 d16, d17, d16 27 @ CHECK: vshl.u64 q8, q9, q8 @ encoding: [0x70,0xff,0xe2,0x04] 28 vshl.u64 q8, q9, q8 43 @ CHECK: vshr.u64 d16, d16, #64 @ encoding: [0xc0,0xff,0xb0,0x00] 44 vshr.u64 d16, d16, #64 51 @ CHECK: vshr.u64 q8, q8, #64 @ encoding: [0xc0,0xff,0xf0,0x00] 52 vshr.u64 q8, q8, #64 107 @ CHECK: vrshl.u64 d16, d17, d16 @ encoding: [0x70,0xff,0xa1,0x05] 108 vrshl.u64 d16, d17, d16 [all …]
|
D | neont2-satshift-encoding.s | 19 @ CHECK: vqshl.u64 d16, d16, d17 @ encoding: [0x71,0xff,0xb0,0x04] 20 vqshl.u64 d16, d16, d17 35 @ CHECK: vqshl.u64 q8, q8, q9 @ encoding: [0x72,0xff,0xf0,0x04] 36 vqshl.u64 q8, q8, q9 51 @ CHECK: vqshl.u64 d16, d16, #63 @ encoding: [0xff,0xff,0xb0,0x07] 52 vqshl.u64 d16, d16, #63 75 @ CHECK: vqshl.u64 q8, q8, #63 @ encoding: [0xff,0xff,0xf0,0x07] 76 vqshl.u64 q8, q8, #63 99 @ CHECK: vqrshl.u64 d16, d16, d17 @ encoding: [0x71,0xff,0xb0,0x05] 100 vqrshl.u64 d16, d16, d17 [all …]
|
/external/valgrind/memcheck/tests/vbit-test/ |
D | vbits.c | 59 case 64: return v.bits.u64; in get_bits64() 76 case 64: fprintf(fp, "%016"PRIx64, v.bits.u64); break; in print_vbits() 116 case 64: new.bits.u64 = ~0ull; break; in undefined_vbits() 143 case 64: new.bits.u64 = 0x0; break; in defined_vbits() 170 case 64: return v1.bits.u64 == v2.bits.u64; in equal_vbits() 216 case 64: new.bits.u64 = bits & ~0ll; break; in truncate_vbits() 264 case 64: new.bits.u64 = bits & ~0ll; break; in left_vbits() 390 case 64: new.bits.u64 = v1.bits.u64 | v2.bits.u64; break; in or_vbits() 418 case 64: new.bits.u64 = v1.bits.u64 & v2.bits.u64; break; in and_vbits() 447 case 32: new.bits.u64 = v1.bits.u32; in concat_vbits() [all …]
|
/external/boringssl/src/crypto/ec/ |
D | p256-64.c | 38 typedef uint64_t u64; typedef 92 typedef u64 smallfelem[NLIMBS]; 95 static const u64 kPrime[4] = {0xfffffffffffffffful, 0xffffffff, 0, 97 static const u64 bottom63bits = 0x7ffffffffffffffful; 102 out[0] = *((u64 *)&in[0]); in bin32_to_felem() 103 out[1] = *((u64 *)&in[8]); in bin32_to_felem() 104 out[2] = *((u64 *)&in[16]); in bin32_to_felem() 105 out[3] = *((u64 *)&in[24]); in bin32_to_felem() 111 *((u64 *)&out[0]) = in[0]; in smallfelem_to_bin32() 112 *((u64 *)&out[8]) = in[1]; in smallfelem_to_bin32() [all …]
|