/external/v8/src/mips64/ |
D | builtins-mips64.cc | 50 __ dsll(s1, s0, kPointerSizeLog2); in Generate_Adaptor() 170 __ dsll(a0, a0, kPointerSizeLog2); in Generate_StringConstructCode() 207 a4, Operand(JSValue::kSize >> kPointerSizeLog2)); in Generate_StringConstructCode() 460 __ dsll(at, a0, kPointerSizeLog2); in Generate_JSConstructStubHelper() 464 __ dsll(at, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper() 479 __ dsll(a0, a0, kPointerSizeLog2); in Generate_JSConstructStubHelper() 495 __ dsll(at, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper() 564 __ dsll(a7, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper() 675 __ dsll(a4, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper() 738 __ SmiScale(a4, a1, kPointerSizeLog2); in Generate_JSConstructStubHelper() [all …]
|
D | code-stubs-mips64.cc | 1625 __ SmiScale(a7, a3, kPointerSizeLog2); in GenerateReadElement() 1639 __ SmiScale(a7, a3, kPointerSizeLog2); in GenerateReadElement() 1668 __ SmiScale(a7, a2, kPointerSizeLog2); in GenerateNewSloppySlow() 1707 __ SmiScale(t2, a2, kPointerSizeLog2); in GenerateNewSloppyFast() 1731 __ SmiScale(t1, a1, kPointerSizeLog2); in GenerateNewSloppyFast() 1736 __ SmiScale(t2, a2, kPointerSizeLog2); in GenerateNewSloppyFast() 1814 __ SmiScale(t2, a1, kPointerSizeLog2); in GenerateNewSloppyFast() 1833 __ SmiScale(t2, a6, kPointerSizeLog2); in GenerateNewSloppyFast() 1848 __ SmiScale(a5, a6, kPointerSizeLog2); in GenerateNewSloppyFast() 1871 __ SmiScale(t2, t1, kPointerSizeLog2); in GenerateNewSloppyFast() [all …]
|
/external/v8/src/mips/ |
D | builtins-mips.cc | 51 __ sll(s1, s0, kPointerSizeLog2); in Generate_Adaptor() 171 __ sll(a0, a0, kPointerSizeLog2); in Generate_StringConstructCode() 208 t0, Operand(JSValue::kSize >> kPointerSizeLog2)); in Generate_StringConstructCode() 453 __ sll(at, a0, kPointerSizeLog2); in Generate_JSConstructStubHelper() 457 __ sll(at, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper() 472 __ sll(a0, a0, kPointerSizeLog2); in Generate_JSConstructStubHelper() 488 __ sll(at, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper() 553 __ sll(t3, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper() 663 __ sll(t0, a3, kPointerSizeLog2 - kSmiTagSize); in Generate_JSConstructStubHelper() 726 __ sll(t0, a1, kPointerSizeLog2 - 1); in Generate_JSConstructStubHelper() [all …]
|
/external/v8/src/arm/ |
D | builtins-arm.cc | 164 __ ldr(r0, MemOperand(sp, r0, LSL, kPointerSizeLog2, PreIndex)); in Generate_StringConstructCode() 198 __ cmp(r4, Operand(JSValue::kSize >> kPointerSizeLog2)); in Generate_StringConstructCode() 442 __ add(r0, r5, Operand(r0, LSL, kPointerSizeLog2)); in Generate_JSConstructStubHelper() 445 __ add(ip, r4, Operand(r3, LSL, kPointerSizeLog2)); // End of object. in Generate_JSConstructStubHelper() 459 __ add(r0, r4, Operand(ip, LSL, kPointerSizeLog2)); // End of object. in Generate_JSConstructStubHelper() 472 __ add(r0, r4, Operand(r3, LSL, kPointerSizeLog2)); // End of object. in Generate_JSConstructStubHelper() 536 __ add(r6, r2, Operand(r3, LSL, kPointerSizeLog2)); // End of object. in Generate_JSConstructStubHelper() 642 __ ldr(ip, MemOperand(r2, r3, LSL, kPointerSizeLog2 - 1)); in Generate_JSConstructStubHelper() 707 __ add(sp, sp, Operand(r1, LSL, kPointerSizeLog2 - 1)); in Generate_JSConstructStubHelper() 756 __ add(r2, r4, Operand(r3, LSL, kPointerSizeLog2)); in Generate_JSEntryTrampolineHelper() [all …]
|
D | macro-assembler-arm.cc | 435 ldr(destination, MemOperand(kRootRegister, index << kPointerSizeLog2), cond); in LoadRoot() 442 str(source, MemOperand(kRootRegister, index << kPointerSizeLog2), cond); in StoreRoot() 483 tst(dst, Operand((1 << kPointerSizeLog2) - 1)); in RecordWriteField() 547 tst(dst, Operand((1 << kPointerSizeLog2) - 1)); in RecordWriteForMap() 1157 add(sp, sp, Operand(argument_count, LSL, kPointerSizeLog2)); in LeaveExitFrame() 1452 ldr(r2, MemOperand(r3, r2, LSL, kPointerSizeLog2)); // Smi-tagged offset. in JumpToHandlerEntry() 1691 add(t2, elements, Operand(t2, LSL, kPointerSizeLog2)); in LoadFromNumberDictionary() 1920 add(scratch2, result, Operand(object_size, LSL, kPointerSizeLog2), SetCC); in Allocate() 3138 Operand(scratch1, LSL, kPointerSizeLog2 + 1)); in LookupNumberStringCache() 3158 Operand(scratch, LSL, kPointerSizeLog2 + 1)); in LookupNumberStringCache() [all …]
|
/external/v8/src/ |
D | globals.h | 141 const int kPointerSizeLog2 = 3; variable 147 const int kPointerSizeLog2 = 2; variable 160 STATIC_ASSERT(kPointerSize == (1 << kPointerSizeLog2)); 225 const int kObjectAlignmentBits = kPointerSizeLog2; 230 const intptr_t kPointerAlignment = (1 << kPointerSizeLog2);
|
D | elements-kind.cc | 48 return kPointerSizeLog2; in ElementsKindToShiftSize()
|
/external/v8/src/ic/mips64/ |
D | ic-mips64.cc | 208 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); in GenerateFastArrayLoad() 209 __ SmiScale(at, key, kPointerSizeLog2); in GenerateFastArrayLoad() 331 __ dsll(scratch3, scratch3, kPointerSizeLog2); in GenerateMappedArgumentsLookup() 344 __ dsll(scratch3, scratch3, kPointerSizeLog2); in GenerateMappedArgumentsLookup() 368 __ dsll(scratch, scratch, kPointerSizeLog2); in GenerateUnmappedArgumentsLookup() 515 __ dsll(at, a3, kPointerSizeLog2 + 1); in GenerateGeneric() 544 __ dsll(at, a3, kPointerSizeLog2 - 1); in GenerateGeneric() 563 __ dsll(at, a6, kPointerSizeLog2); in GenerateGeneric() 574 __ dsll(v0, a5, kPointerSizeLog2); in GenerateGeneric() 652 __ SmiScale(at, key, kPointerSizeLog2); in KeyedStoreGenerateGenericHelper() [all …]
|
D | stub-cache-mips64.cc | 50 __ dsll(at, offset_scratch, kPointerSizeLog2); in ProbeTable()
|
/external/v8/src/arm64/ |
D | builtins-arm64.cc | 191 __ Cmp(x4, JSValue::kSize >> kPointerSizeLog2); in Generate_StringConstructCode() 447 Operand(prealloc_fields, LSL, kPointerSizeLog2)); in Generate_JSConstructStubHelper() 451 __ Add(obj_end, new_obj, Operand(obj_size, LSL, kPointerSizeLog2)); in Generate_JSConstructStubHelper() 465 __ Add(first_prop, new_obj, Operand(obj_size, LSL, kPointerSizeLog2)); in Generate_JSConstructStubHelper() 598 __ Add(x3, x2, Operand(argc, LSL, kPointerSizeLog2)); in Generate_JSConstructStubHelper() 726 __ Add(x10, argv, Operand(argc, LSL, kPointerSizeLog2)); in Generate_JSEntryTrampolineHelper() 1156 __ Add(scratch2, jssp, Operand(argc, LSL, kPointerSizeLog2)); in Generate_FunctionCall() 1248 __ Cmp(x10, Operand::UntagSmiAndScale(argc, kPointerSizeLog2)); in Generate_FunctionApply() 1389 __ Cmp(x10, Operand(x2, LSL, kPointerSizeLog2)); in ArgumentAdaptorStackCheck() 1452 __ Lsl(argc_expected, argc_expected, kPointerSizeLog2); in Generate_ArgumentsAdaptorTrampoline() [all …]
|
D | code-stubs-arm64.cc | 1068 __ Add(temp_argv, jssp, Operand(x0, LSL, kPointerSizeLog2)); in Generate() 1644 __ Add(x10, local_fp, Operand::UntagSmiAndScale(x10, kPointerSizeLog2)); in GenerateReadElement() 1676 __ Add(x10, caller_fp, Operand::UntagSmiAndScale(x11, kPointerSizeLog2)); in GenerateNewSloppySlow() 1733 __ Add(x10, caller_fp, Operand(arg_count, LSL, kPointerSizeLog2)); in GenerateNewSloppyFast() 1761 __ Mov(size, Operand(mapped_params, LSL, kPointerSizeLog2)); in GenerateNewSloppyFast() 1770 __ Add(size, size, Operand(arg_count, LSL, kPointerSizeLog2)); in GenerateNewSloppyFast() 1860 __ Add(x10, elements, Operand(mapped_params, LSL, kPointerSizeLog2)); in GenerateNewSloppyFast() 1901 __ Add(backing_store, elements, Operand(loop_count, LSL, kPointerSizeLog2)); in GenerateNewSloppyFast() 1908 __ Mov(x10, Operand(loop_count, LSL, kPointerSizeLog2)); in GenerateNewSloppyFast() 1935 __ Sub(recv_arg, recv_arg, Operand(x10, LSL, kPointerSizeLog2)); in GenerateNewSloppyFast() [all …]
|
D | macro-assembler-arm64.cc | 1400 Ldr(destination, MemOperand(root, index << kPointerSizeLog2)); in LoadRoot() 1406 Str(source, MemOperand(root, index << kPointerSizeLog2)); in StoreRoot() 1414 MemOperand(root, Heap::kTrueValueRootIndex << kPointerSizeLog2)); in LoadTrueFalseRoots() 1551 STATIC_ASSERT(StackHandler::kKindWidth < kPointerSizeLog2); in JumpToHandlerEntry() 1553 Ldr(scratch2, MemOperand(scratch1, scratch2, LSL, kPointerSizeLog2)); in JumpToHandlerEntry() 2326 Operand(scratch1, LSL, kPointerSizeLog2 + 1)); in LookupNumberStringCache() 2343 Operand(scratch, LSL, kPointerSizeLog2 + 1)); in LookupNumberStringCache() 2662 Sub(field_ptr, dst, Operand(field_ptr, LSL, kPointerSizeLog2)); in FillFields() 3465 Adds(scratch3, result, Operand(object_size, LSL, kPointerSizeLog2)); in Allocate() 4196 Add(scratch2, elements, Operand(scratch2, LSL, kPointerSizeLog2)); in LoadFromNumberDictionary() [all …]
|
/external/v8/src/ic/mips/ |
D | ic-mips.cc | 209 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); in GenerateFastArrayLoad() 210 __ sll(at, key, kPointerSizeLog2 - kSmiTagSize); in GenerateFastArrayLoad() 515 __ sll(at, a3, kPointerSizeLog2 + 1); in GenerateGeneric() 542 __ sll(at, a3, kPointerSizeLog2); in GenerateGeneric() 558 __ sll(at, t2, kPointerSizeLog2); in GenerateGeneric() 569 __ sll(v0, t1, kPointerSizeLog2); in GenerateGeneric() 647 __ sll(at, key, kPointerSizeLog2 - kSmiTagSize); in KeyedStoreGenerateGenericHelper() 668 __ sll(scratch_value, key, kPointerSizeLog2 - kSmiTagSize); in KeyedStoreGenerateGenericHelper() 686 __ sll(scratch_value, key, kPointerSizeLog2 - kSmiTagSize); in KeyedStoreGenerateGenericHelper() 708 __ sll(at, key, kPointerSizeLog2); in KeyedStoreGenerateGenericHelper()
|
D | stub-cache-mips.cc | 50 __ sll(at, offset_scratch, kPointerSizeLog2); in ProbeTable()
|
/external/v8/src/ic/arm64/ |
D | ic-arm64.cc | 201 __ Ldr(scratch2, MemOperand(scratch1, scratch2, LSL, kPointerSizeLog2)); in GenerateFastArrayLoad() 292 __ Ldr(scratch1, MemOperand(scratch1, scratch2, LSL, kPointerSizeLog2)); in GenerateMappedArgumentsLookup() 298 __ Lsl(scratch1, scratch1, kPointerSizeLog2); in GenerateMappedArgumentsLookup() 330 return MemOperand(backing_store, scratch, LSL, kPointerSizeLog2); in GenerateUnmappedArgumentsLookup() 522 __ Add(scratch3, scratch3, Operand(scratch2, LSL, kPointerSizeLog2 + 1)); in GenerateKeyedLoadWithNameKey() 570 __ Ldr(result, MemOperand(receiver, scratch5, LSL, kPointerSizeLog2)); in GenerateKeyedLoadWithNameKey() 579 __ Ldr(result, MemOperand(scratch1, scratch4, LSL, kPointerSizeLog2)); in GenerateKeyedLoadWithNameKey() 696 __ Add(x10, x10, Operand::UntagSmiAndScale(key, kPointerSizeLog2)); in KeyedStoreGenerateGenericHelper() 717 __ Add(address, address, Operand::UntagSmiAndScale(key, kPointerSizeLog2)); in KeyedStoreGenerateGenericHelper() 743 __ Add(x10, x10, Operand::UntagSmiAndScale(key, kPointerSizeLog2)); in KeyedStoreGenerateGenericHelper()
|
D | stub-cache-arm64.cc | 52 __ Add(scratch, scratch, Operand(scratch3, LSL, kPointerSizeLog2)); in ProbeTable()
|
/external/v8/src/ic/x64/ |
D | stub-cache-x64.cc | 27 ? kPointerSizeLog2 == StubCache::kCacheIndexShift + 1 in ProbeTable() 28 : kPointerSizeLog2 == StubCache::kCacheIndexShift); in ProbeTable()
|
/external/v8/src/heap/ |
D | store-buffer.cc | 550 int_addr >>= kPointerSizeLog2; in Compact() 555 int_addr & (Page::kPageAlignmentMask >> kPointerSizeLog2); in Compact() 575 *old_top_++ = reinterpret_cast<Address>(int_addr << kPointerSizeLog2); in Compact()
|
D | store-buffer.h | 67 static const int kStoreBufferOverflowBit = 1 << (14 + kPointerSizeLog2);
|
D | spaces.h | 152 static const size_t kLength = (1 << kPageSizeBits) >> (kPointerSizeLog2); 155 (1 << kPageSizeBits) >> (kPointerSizeLog2 + kBitsPerByteLog2); 587 return static_cast<uint32_t>(addr - this->address()) >> kPointerSizeLog2; in AddressToMarkbitIndex() 593 return static_cast<uint32_t>(offset) >> kPointerSizeLog2; in FastAddressToMarkbitIndex() 597 return this->address() + (index << kPointerSizeLog2); in MarkbitIndexToAddress() 2448 return static_cast<uint32_t>(addr - start_) >> kPointerSizeLog2; in INLINE() 2452 return reinterpret_cast<Address>(index << kPointerSizeLog2); in INLINE()
|
/external/v8/src/x64/ |
D | builtins-x64.cc | 194 __ shlp(rdi, Immediate(kPointerSizeLog2)); in Generate_JSConstructStubHelper() 465 SmiIndex index = masm->SmiToIndex(rbx, rbx, kPointerSizeLog2); in Generate_JSConstructStubHelper() 995 __ PositiveSmiTimesPowerOfTwoToInteger64(rdx, rax, kPointerSizeLog2); in Generate_FunctionApply() 1244 Immediate(JSValue::kSize >> kPointerSizeLog2)); in Generate_StringConstructCode() 1331 __ shlp(rdx, Immediate(kPointerSizeLog2)); in ArgumentsAdaptorStackCheck() 1366 SmiIndex index = masm->SmiToIndex(rbx, rbx, kPointerSizeLog2); in LeaveArgumentsAdaptorFrame()
|
/external/v8/src/ic/arm/ |
D | ic-arm.cc | 507 __ add(r4, r4, Operand(r3, LSL, kPointerSizeLog2 + 1)); in GenerateGeneric() 542 __ ldr(r5, MemOperand(r4, r3, LSL, kPointerSizeLog2)); in GenerateGeneric() 556 __ ldr(r0, MemOperand(receiver, r6, LSL, kPointerSizeLog2)); in GenerateGeneric() 565 __ ldr(r0, MemOperand(receiver, r5, LSL, kPointerSizeLog2)); in GenerateGeneric() 711 MemOperand(address, key, LSL, kPointerSizeLog2, PreIndex)); in KeyedStoreGenerateGenericHelper()
|
D | stub-cache-arm.cc | 49 __ add(base_addr, base_addr, Operand(offset_scratch, LSL, kPointerSizeLog2)); in ProbeTable()
|
/external/v8/src/ic/x87/ |
D | stub-cache-x87.cc | 159 DCHECK(kCacheIndexShift == kPointerSizeLog2); in GenerateProbe()
|
/external/v8/src/ic/ia32/ |
D | stub-cache-ia32.cc | 159 DCHECK(kCacheIndexShift == kPointerSizeLog2); in GenerateProbe()
|