/art/runtime/gc/ |
D | task_processor_test.cc | 32 RecursiveTask(TaskProcessor* task_processor, Atomic<size_t>* counter, size_t max_recursion) in RecursiveTask() argument 33 : HeapTask(NanoTime() + MsToNs(10)), task_processor_(task_processor), counter_(counter), in RecursiveTask() 71 Atomic<size_t> counter(0); in TEST_F() local 72 task_processor.AddTask(self, new RecursiveTask(&task_processor, &counter, kRecursion)); in TEST_F() 79 while (counter.LoadSequentiallyConsistent() != kRecursion) { in TEST_F() 90 counter.StoreSequentiallyConsistent(0); in TEST_F() 95 task_processor.AddTask(self, new RecursiveTask(&task_processor, &counter, kRecursion)); in TEST_F() 100 ASSERT_EQ(counter.LoadSequentiallyConsistent(), kRecursion); in TEST_F() 105 explicit TestOrderTask(uint64_t expected_time, size_t expected_counter, size_t* counter) in TestOrderTask() argument 106 : HeapTask(expected_time), expected_counter_(expected_counter), counter_(counter) { in TestOrderTask() [all …]
|
D | heap.cc | 1725 InstanceCounter counter(classes, use_is_assignable_from, counts); in CountInstances() local 1726 VisitObjects(InstanceCounter::Callback, &counter); in CountInstances()
|
/art/test/135-MirandaDispatch/src/ |
D | Main.java | 20 static int counter = 0; field in Main 31 ++counter; in m() 46 if (counter != loopIterations * loopIterations) { in main() 47 System.out.println("Expected " + loopIterations * loopIterations + " got " + counter); in main()
|
/art/runtime/base/ |
D | debug_stack.h | 57 explicit DebugStackReferenceImpl(DebugStackRefCounterImpl<false>* counter) { UNUSED(counter); } in DebugStackReferenceImpl() argument 91 explicit DebugStackReferenceImpl(DebugStackRefCounterImpl<kIsDebug>* counter) in DebugStackReferenceImpl() argument 92 : counter_(counter), ref_count_(counter->IncrementRefCount()) { in DebugStackReferenceImpl()
|
D | variant_map.h | 76 static size_t counter = 0; in AllocateCounter() local 77 counter++; in AllocateCounter() 79 return counter; in AllocateCounter()
|
/art/test/441-checker-inliner/src/ |
D | Main.java | 197 private static int counter = 42; field in Main 200 return ++counter; in incCounter()
|
/art/test/114-ParallelGC/src/ |
D | Main.java | 33 private final static AtomicInteger counter = new AtomicInteger(); field in Main 112 int number = counter.incrementAndGet(); in work()
|
/art/runtime/gc/accounting/ |
D | space_bitmap_test.cc | 92 explicit SimpleCounter(size_t* counter) : count_(counter) {} in SimpleCounter() argument
|
/art/runtime/ |
D | oat.cc | 416 ssize_t counter = static_cast<ssize_t>(index); in GetStoreKeyValuePairByIndex() local 418 while (ptr < end && counter >= 0) { in GetStoreKeyValuePairByIndex() 425 if (counter == 0) { in GetStoreKeyValuePairByIndex() 430 counter--; in GetStoreKeyValuePairByIndex()
|
D | debugger.cc | 3198 size_t* counter = GetReferenceCounterForEvent(req.InstrumentationEvent()); in RequestDeoptimizationLocked() local 3199 CHECK(counter != nullptr) << StringPrintf("No counter for instrumentation event 0x%x", in RequestDeoptimizationLocked() 3201 if (*counter == 0) { in RequestDeoptimizationLocked() 3206 *counter = *counter + 1; in RequestDeoptimizationLocked() 3211 size_t* counter = GetReferenceCounterForEvent(req.InstrumentationEvent()); in RequestDeoptimizationLocked() local 3212 CHECK(counter != nullptr) << StringPrintf("No counter for instrumentation event 0x%x", in RequestDeoptimizationLocked() 3214 *counter = *counter - 1; in RequestDeoptimizationLocked() 3215 if (*counter == 0) { in RequestDeoptimizationLocked()
|
/art/compiler/optimizing/ |
D | intrinsics_x86_64.cc | 906 CpuRegister counter = locations->GetTemp(0).AsRegister<CpuRegister>(); in GenerateStringIndexOf() local 913 DCHECK_EQ(counter.AsRegister(), RCX); in GenerateStringIndexOf() 954 __ movl(counter, string_length); in GenerateStringIndexOf() 966 __ xorl(counter, counter); in GenerateStringIndexOf() 968 __ cmov(kGreater, counter, start_index, false); // 32-bit copy is enough. in GenerateStringIndexOf() 971 __ leaq(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_2, value_offset)); in GenerateStringIndexOf() 974 __ negq(counter); // Needs to be 64-bit negation, as the address computation is 64-bit. in GenerateStringIndexOf() 975 __ leaq(counter, Address(string_length, counter, ScaleFactor::TIMES_1, 0)); in GenerateStringIndexOf() 987 __ subl(string_length, counter); in GenerateStringIndexOf()
|
D | intrinsics_x86.cc | 981 Register counter = locations->GetTemp(0).AsRegister<Register>(); in GenerateStringIndexOf() local 988 DCHECK_EQ(counter, ECX); in GenerateStringIndexOf() 1029 __ movl(counter, string_length); in GenerateStringIndexOf() 1041 __ xorl(counter, counter); in GenerateStringIndexOf() 1043 __ cmovl(kGreater, counter, start_index); in GenerateStringIndexOf() 1046 __ leal(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_2, value_offset)); in GenerateStringIndexOf() 1050 __ negl(counter); in GenerateStringIndexOf() 1051 __ leal(counter, Address(string_length, counter, ScaleFactor::TIMES_1, 0)); in GenerateStringIndexOf() 1063 __ subl(string_length, counter); in GenerateStringIndexOf()
|
D | bounds_check_elimination.cc | 1794 size_t counter = 0; in AddComparesWithDeoptimization() local 1810 counter++; in AddComparesWithDeoptimization() 1813 if (counter >= kThresholdForAddingDeoptimize && in AddComparesWithDeoptimization()
|
/art/runtime/arch/arm64/ |
D | quick_entrypoints_arm64.S | 767 .macro LOADREG counter size register return 769 add \counter, \counter, 12
|