Lines Matching refs:self
39 inline mirror::Object* Heap::AllocObjectWithAllocator(Thread* self, mirror::Class* klass, in AllocObjectWithAllocator() argument
46 CHECK_EQ(self->GetState(), kRunnable); in AllocObjectWithAllocator()
47 self->AssertThreadSuspensionIsAllowable(); in AllocObjectWithAllocator()
53 obj = AllocLargeObject<kInstrumented, PreFenceVisitor>(self, &klass, byte_count, in AllocObjectWithAllocator()
59 self->ClearException(); in AllocObjectWithAllocator()
73 if (allocator == kAllocatorTypeTLAB && byte_count <= self->TlabSize()) { in AllocObjectWithAllocator()
74 obj = self->AllocTlab(byte_count); in AllocObjectWithAllocator()
88 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated, in AllocObjectWithAllocator()
92 obj = AllocateInternalWithGc(self, allocator, byte_count, &bytes_allocated, &usable_size, in AllocObjectWithAllocator()
98 if (!self->IsExceptionPending() && is_current_allocator && !after_is_current_allocator) { in AllocObjectWithAllocator()
100 return AllocObject<kInstrumented>(self, klass, byte_count, pre_fence_visitor); in AllocObjectWithAllocator()
137 RuntimeStats* thread_stats = self->GetStats(); in AllocObjectWithAllocator()
148 PushOnAllocationStack(self, &obj); in AllocObjectWithAllocator()
162 CheckConcurrentGC(self, new_num_bytes_allocated, &obj); in AllocObjectWithAllocator()
165 self->VerifyStack(); in AllocObjectWithAllocator()
172 inline void Heap::PushOnAllocationStack(Thread* self, mirror::Object** obj) { in PushOnAllocationStack() argument
174 if (UNLIKELY(!self->PushOnThreadLocalAllocationStack(*obj))) { in PushOnAllocationStack()
175 PushOnThreadLocalAllocationStackWithInternalGC(self, obj); in PushOnAllocationStack()
178 PushOnAllocationStackWithInternalGC(self, obj); in PushOnAllocationStack()
183 inline mirror::Object* Heap::AllocLargeObject(Thread* self, mirror::Class** klass, in AllocLargeObject() argument
187 StackHandleScope<1> hs(self); in AllocLargeObject()
189 return AllocObjectWithAllocator<kInstrumented, false, PreFenceVisitor>(self, *klass, byte_count, in AllocLargeObject()
195 inline mirror::Object* Heap::TryToAllocate(Thread* self, AllocatorType allocator_type, in TryToAllocate() argument
217 ret = rosalloc_space_->Alloc(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
220 ret = rosalloc_space_->AllocNonvirtual(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
227 ret = dlmalloc_space_->Alloc(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
230 ret = dlmalloc_space_->AllocNonvirtual(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
235 ret = non_moving_space_->Alloc(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
239 ret = large_object_space_->Alloc(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
248 if (UNLIKELY(self->TlabSize() < alloc_size)) { in TryToAllocate()
255 if (!bump_pointer_space_->AllocNewTlab(self, new_tlab_size)) { in TryToAllocate()
263 ret = self->AllocTlab(alloc_size); in TryToAllocate()
322 inline void Heap::CheckConcurrentGC(Thread* self, size_t new_num_bytes_allocated, in CheckConcurrentGC() argument
325 RequestConcurrentGCAndSaveObject(self, obj); in CheckConcurrentGC()