/art/runtime/ |
D | oat_file-inl.h | 36 return 0u; in GetOatQuickMethodHeaderOffset() 44 return 0u; in GetQuickCodeSize() 52 return 0u; in GetQuickCodeSizeOffset() 60 return 0u; in GetFrameSizeInBytes() 68 return 0u; in GetCoreSpillMask() 76 return 0u; in GetFpSpillMask() 87 if (UNLIKELY(offset == 0u)) { in GetGcMap() 95 return static_cast<uint32_t>(gc_map != nullptr ? gc_map - begin_ : 0u); in GetGcMapOffset() 101 return 0u; in GetGcMapOffsetOffset() 108 return static_cast<uint32_t>(mapping_table != nullptr ? mapping_table - begin_ : 0u); in GetMappingTableOffset() [all …]
|
D | leb128.h | 130 while (extra_bits != 0u) { in EncodeSignedLeb128() 171 while (extra_bits != 0u) { in PushBackSigned()
|
D | vmap_table.h | 111 DCHECK_NE(spill_mask, 0u); in ComputeRegister()
|
/art/compiler/dex/ |
D | mir_field_info.h | 43 return (flags_ & kFlagIsStatic) != 0u; in IsStatic() 63 return (flags_ & kFlagIsVolatile) != 0u; in IsVolatile() 78 declaring_field_idx_(0u), in MirFieldInfo() 79 declaring_class_idx_(0u), in MirFieldInfo() 112 field_offset_(0u) { in MirIFieldLoweringInfo() 116 return (flags_ & kFlagFastGet) != 0u; in FastGet() 120 return (flags_ & kFlagFastPut) != 0u; in FastPut() 159 field_offset_(0u), in MirSFieldLoweringInfo() 164 return (flags_ & kFlagFastGet) != 0u; in FastGet() 168 return (flags_ & kFlagFastPut) != 0u; in FastPut() [all …]
|
D | mir_method_info.h | 39 return (flags_ & kFlagIsStatic) != 0u; in IsStatic() 69 declaring_method_idx_(0u), in MirMethodInfo() 70 declaring_class_idx_(0u), in MirMethodInfo() 103 ((type == kStatic) ? kFlagIsStatic : 0u) | in MirMethodLoweringInfo() 106 direct_code_(0u), in MirMethodLoweringInfo() 107 direct_method_(0u), in MirMethodLoweringInfo() 109 target_method_idx_(0u), in MirMethodLoweringInfo() 110 vtable_idx_(0u), in MirMethodLoweringInfo() 116 DCHECK_EQ(target_method_idx_, 0u); in SetDevirtualizationTarget() 123 return (flags_ & kFlagFastPath) != 0u; in FastPath() [all …]
|
D | reg_storage.h | 109 DCHECK_CONSTEXPR(rs_kind != k64BitPair, , 0u) in RegStorage() 110 DCHECK_CONSTEXPR((rs_kind & ~kShapeMask) == 0, , 0u) in RegStorage() 115 DCHECK_CONSTEXPR(rs_kind == k64BitPair, << rs_kind, 0u) in RegStorage() 117 << low_reg << ", " << high_reg, 0u) in RegStorage()
|
D | dataflow_iterator-inl.h | 141 while (loop_head_stack_->Size() != 0u && in Next() 166 if (loop_ends_->Get(idx) != 0u) { in Next()
|
D | pass_me.h | 70 unsigned int flags = 0u, const char* dump = "")
|
D | dataflow_iterator.h | 419 DCHECK_EQ(loop_head_stack_->Size(), 0u); in LoopRepeatingTopologicalSortIterator() 423 DCHECK_EQ(loop_head_stack_->Size(), 0u); in ~LoopRepeatingTopologicalSortIterator()
|
/art/compiler/dex/quick/ |
D | resource_mask.h | 59 : masks_ { 0u, 0u } { in ResourceMask() 76 return ResourceMask(bit >= 64u ? 0u : UINT64_C(1) << bit, in Bit() 77 bit >= 64u ? UINT64_C(1) << (bit - 64u) : 0u); in Bit() 83 DCHECK_CONSTEXPR((start_bit & 1u) == 0u, << start_bit << " isn't even", Bit(0)) in TwoBits() 84 ResourceMask(start_bit >= 64u ? 0u : UINT64_C(3) << start_bit, in TwoBits() 85 start_bit >= 64u ? UINT64_C(3) << (start_bit - 64u) : 0u); in TwoBits() 113 return (masks_[0] & other.masks_[0]) != 0u || (masks_[1] & other.masks_[1]) != 0u; in Intersects() 122 return (masks_[bit / 64u] & (UINT64_C(1) << (bit & 63u))) != 0u; in HasBit()
|
D | local_optimizations.cc | 24 #define LOAD_STORE_CHECK_REG_DEP(mask, check) (mask.Intersects(*check->u.m.def_mask)) 27 #define CHECK_REG_DEP(use, def, check) (def.Intersects(*check->u.m.use_mask)) || \ 28 (use.Union(def).Intersects(*check->u.m.def_mask)) 177 ResourceMask this_mem_mask = kEncodeMem.Intersection(this_lir->u.m.use_mask->Union( in ApplyLoadStoreElimination() 178 *this_lir->u.m.def_mask)); in ApplyLoadStoreElimination() 187 if (this_lir->u.m.def_mask->Intersects(*this_lir->u.m.use_mask)) { in ApplyLoadStoreElimination() 191 ResourceMask stop_def_reg_mask = this_lir->u.m.def_mask->Without(kEncodeMem); in ApplyLoadStoreElimination() 192 ResourceMask stop_use_reg_mask = this_lir->u.m.use_mask->Without(kEncodeMem); in ApplyLoadStoreElimination() 196 if (uses_pc.Intersects(this_lir->u.m.use_mask->Union(*this_lir->u.m.def_mask))) { in ApplyLoadStoreElimination() 214 if (uses_pc.Intersects(check_lir->u.m.use_mask->Union(*check_lir->u.m.def_mask))) { in ApplyLoadStoreElimination() [all …]
|
D | mir_to_lir-inl.h | 60 insn->u.m.use_mask = insn->u.m.def_mask = &kEncodeAll; in RawLIR() 165 lir->u.m.use_mask = lir->u.m.def_mask = &kEncodeNone; in SetupResourceMasks() 207 lir->u.m.def_mask = lir->u.m.use_mask = &kEncodeAll; in SetupResourceMasks() 254 lir->u.m.use_mask = mask_cache_.GetMask(use_mask); in SetupResourceMasks() 255 lir->u.m.def_mask = mask_cache_.GetMask(def_mask); in SetupResourceMasks()
|
D | codegen_util.cc | 84 inst->u.m.def_mask = &kEncodeAll; in MarkSafepointPC() 86 DCHECK(safepoint_pc->u.m.def_mask->Equals(kEncodeAll)); in MarkSafepointPC() 91 after->u.m.def_mask = &kEncodeAll; in MarkSafepointPCAfter() 100 DCHECK(safepoint_pc->u.m.def_mask->Equals(kEncodeAll)); in MarkSafepointPCAfter() 138 mask_ptr = &lir->u.m.use_mask; in SetMemRefType() 140 mask_ptr = &lir->u.m.def_mask; in SetMemRefType() 173 DCHECK((is_load ? lir->u.m.use_mask : lir->u.m.def_mask)->Intersection(kEncodeMem).Equals( in AnnotateDalvikRegAccess() 268 if (lir->u.m.use_mask && (!lir->flags.is_nop || dump_nop)) { in DumpLIRInsn() 269 DUMP_RESOURCE_MASK(DumpResourceMask(lir, *lir->u.m.use_mask, "use")); in DumpLIRInsn() 271 if (lir->u.m.def_mask && (!lir->flags.is_nop || dump_nop)) { in DumpLIRInsn() [all …]
|
/art/runtime/quick/ |
D | quick_method_frame_info.h | 29 : frame_size_in_bytes_(0u), in QuickMethodFrameInfo() 30 core_spill_mask_(0u), in QuickMethodFrameInfo() 31 fp_spill_mask_(0u) { in QuickMethodFrameInfo()
|
/art/compiler/utils/ |
D | array_ref.h | 61 : array_(nullptr), size_(0u) { in ArrayRef() 128 bool empty() const { return size() == 0u; } in empty() 143 DCHECK_NE(size_, 0u); in front() 148 DCHECK_NE(size_, 0u); in front() 153 DCHECK_NE(size_, 0u); in back() 158 DCHECK_NE(size_, 0u); in back()
|
D | debug_stack.h | 48 size_t IncrementRefCount() { return 0u; } in IncrementRefCount() 50 size_t GetRefCount() const { return 0u; } in GetRefCount() 77 DebugStackRefCounterImpl() : ref_count_(0u) { } in DebugStackRefCounterImpl() 82 void CheckNoRefs() const { CHECK_EQ(ref_count_, 0u); } in CheckNoRefs()
|
D | arena_allocator.h | 75 size_t NumAllocations() const { return 0u; } in NumAllocations() 76 size_t BytesAllocated() const { return 0u; } in BytesAllocated()
|
/art/compiler/ |
D | gc_map_builder.h | 31 : entries_(entries), references_width_(entries != 0u ? references_width : 0u), in GcMapBuilder() 34 : 0u), in GcMapBuilder()
|
/art/runtime/mirror/ |
D | art_method-inl.h | 149 return 0u; in GetCodeSize() 263 if (UNLIKELY(offset == 0u)) { in GetMappingTable() 282 if (UNLIKELY(offset == 0u)) { in GetVmapTable() 301 if (UNLIKELY(offset == 0u)) { in GetNativeGcMap() 363 return QuickMethodFrameInfo(kStackAlignment, 0u, 0u); in GetQuickFrameInfo()
|
/art/runtime/arch/x86/ |
D | quick_method_frame_info_x86.h | 45 0u); in X86CalleeSaveMethodFrameInfo()
|
/art/runtime/base/ |
D | bit_vector.h | 90 bit_index_(FindIndex(0u)) { } in IndexIterator() 110 while (word == 0u) { in FindIndex()
|
/art/runtime/arch/mips/ |
D | quick_method_frame_info_mips.h | 50 0u); in MipsCalleeSaveMethodFrameInfo()
|
/art/compiler/dex/quick/arm64/ |
D | assemble_arm64.cc | 629 new_lir->u.a.pcrel_next = orig_lir->u.a.pcrel_next; in ReplaceFixup() 633 prev_lir->u.a.pcrel_next = new_lir; in ReplaceFixup() 640 new_lir->u.a.pcrel_next = orig_lir; in InsertFixupBefore() 644 DCHECK(prev_lir->u.a.pcrel_next == orig_lir); in InsertFixupBefore() 645 prev_lir->u.a.pcrel_next = new_lir; in InsertFixupBefore() 909 lir = lir->u.a.pcrel_next; in AssembleLIR() 978 lir->u.a.pcrel_next = NULL; in LinkFixupInsns() 982 last_fixup->u.a.pcrel_next = lir; in LinkFixupInsns()
|
/art/compiler/dex/quick/arm/ |
D | assemble_arm.cc | 1042 new_lir->u.a.pcrel_next = orig_lir->u.a.pcrel_next; in ReplaceFixup() 1046 prev_lir->u.a.pcrel_next = new_lir; in ReplaceFixup() 1053 new_lir->u.a.pcrel_next = orig_lir; in InsertFixupBefore() 1057 DCHECK(prev_lir->u.a.pcrel_next == orig_lir); in InsertFixupBefore() 1058 prev_lir->u.a.pcrel_next = new_lir; in InsertFixupBefore() 1323 lir = new_adr->u.a.pcrel_next; in AssembleLIR() 1374 lir = new_inst->u.a.pcrel_next; in AssembleLIR() 1587 lir = lir->u.a.pcrel_next; in AssembleLIR() 1656 lir->u.a.pcrel_next = NULL; in LinkFixupInsns() 1660 last_fixup->u.a.pcrel_next = lir; in LinkFixupInsns()
|
/art/runtime/gc/ |
D | heap-inl.h | 105 DCHECK_GT(bytes_allocated, 0u); in AllocObjectWithAllocator() 106 DCHECK_GT(usable_size, 0u); in AllocObjectWithAllocator()
|