/external/v8/test/cctest/ |
D | test-code-stubs-arm64.cc | 63 __ Mov(jssp, csp); in MakeConvertDToIFuncTrampoline() 64 __ SetStackPointer(jssp); in MakeConvertDToIFuncTrampoline() 68 __ Mov(source_reg, jssp); in MakeConvertDToIFuncTrampoline() 93 __ add(source_reg, jssp, Operand(source_reg_offset)); in MakeConvertDToIFuncTrampoline() 117 __ Mov(csp, jssp); in MakeConvertDToIFuncTrampoline() 166 Register source_registers[] = {jssp, x0, x1, x2, x3, x4, x5, x6, x7, x8, x9, in TEST()
|
D | test-disasm-arm64.cc | 386 COMPARE(adds(cp, jssp, Operand(fp, SXTX)), "adds cp, jssp, fp, sxtx"); in TEST_() 412 COMPARE(subs(cp, jssp, Operand(fp, SXTX)), "subs cp, jssp, fp, sxtx"); in TEST_() 879 COMPARE(ldr(w24, MemOperand(jssp)), "ldr w24, [jssp]"); in TEST_() 880 COMPARE(ldr(x25, MemOperand(jssp, 8)), "ldr x25, [jssp, #8]"); in TEST_() 881 COMPARE(str(w26, MemOperand(jssp, 4, PreIndex)), "str w26, [jssp, #4]!"); in TEST_() 882 COMPARE(str(cp, MemOperand(jssp, -8, PostIndex)), "str cp, [jssp], #-8"); in TEST_() 971 COMPARE(ldr(x0, MemOperand(jssp, wzr, SXTW)), "ldr x0, [jssp, wzr, sxtw]"); in TEST_() 972 COMPARE(str(x1, MemOperand(jssp, xzr)), "str x1, [jssp, xzr]"); in TEST_() 999 COMPARE(ldrb(w28, MemOperand(jssp, 3, PostIndex)), "ldrb w28, [jssp], #3"); in TEST_() 1000 COMPARE(strb(fp, MemOperand(jssp, -42, PreIndex)), "strb w29, [jssp, #-42]!"); in TEST_() [all …]
|
D | test-assembler-arm64.cc | 8251 CHECK(jssp.Bit() == (1UL << kJSSPCode)); in TEST() 8259 CHECK(jssp.Bit() == wjssp.Bit()); in TEST() 8277 __ SetStackPointer(jssp); in TEST() 8278 CHECK(jssp.Is(__ StackPointer())); in TEST() 8592 static RegList const allowed = ~(masm.TmpList()->list() | jssp.Bit()); in PushPopJsspSimpleHelper() 8611 __ Mov(jssp, __ StackPointer()); in PushPopJsspSimpleHelper() 8612 __ SetStackPointer(jssp); in PushPopJsspSimpleHelper() 8793 __ Mov(jssp, __ StackPointer()); in PushPopFPJsspSimpleHelper() 8794 __ SetStackPointer(jssp); in PushPopFPJsspSimpleHelper() 8940 ~(x8.Bit() | x9.Bit() | jssp.Bit() | xzr.Bit()); in PushPopJsspMixedMethodsHelper() [all …]
|
/external/v8/src/arm64/ |
D | builtins-arm64.cc | 162 __ Ldr(arg, MemOperand(jssp, 2 * kPointerSize, PostIndex)); in Generate_StringConstructCode() 1019 __ CompareRoot(jssp, Heap::kStackLimitRootIndex); in Generate_OsrAfterStackCheck() 1156 __ Add(scratch2, jssp, Operand(argc, LSL, kPointerSizeLog2)); in Generate_FunctionCall() 1162 __ Cmp(scratch1, jssp); in Generate_FunctionCall() 1246 __ Sub(x10, jssp, x10); in Generate_FunctionApply() 1387 __ Sub(x10, jssp, x10); in ArgumentAdaptorStackCheck() 1399 __ Add(fp, jssp, in EnterArgumentsAdaptorFrame() 1412 __ Mov(jssp, fp); in LeaveArgumentsAdaptorFrame() 1459 __ Mov(copy_to, jssp); in Generate_ArgumentsAdaptorTrampoline() 1498 __ Mov(copy_to, jssp); in Generate_ArgumentsAdaptorTrampoline() [all …]
|
D | macro-assembler-arm64.cc | 35 sp_(jssp), in MacroAssembler() 1589 DCHECK(jssp.Is(StackPointer())); in Throw() 1592 Ldr(jssp, MemOperand(scratch1)); in Throw() 1631 DCHECK(jssp.Is(StackPointer())); in ThrowUncatchable() 1634 Ldr(jssp, MemOperand(scratch1)); in ThrowUncatchable() 1640 Peek(jssp, StackHandlerConstants::kNextOffset); in ThrowUncatchable() 2988 Push(jssp, xzr); // Push xzr to maintain csp required 16-bytes alignment. in TruncateDoubleToI() 2989 Mov(jssp, csp); in TruncateDoubleToI() 2990 SetStackPointer(jssp); in TruncateDoubleToI() 2997 jssp, in TruncateDoubleToI() [all …]
|
D | code-stubs-arm64.cc | 132 DCHECK(jssp.Is(masm->StackPointer())); in Generate() 143 if (input.is(jssp)) double_offset += 2 * kPointerSize; in Generate() 147 if (input.is(jssp)) double_offset += 1 * kDoubleSize; in Generate() 1049 DCHECK(jssp.Is(__ StackPointer())); in Generate() 1068 __ Add(temp_argv, jssp, Operand(x0, LSL, kPointerSizeLog2)); in Generate() 1173 DCHECK(jssp.Is(__ StackPointer())); in Generate() 1217 __ SetStackPointer(jssp); in Generate() 1245 DCHECK(jssp.Is(__ StackPointer())); in Generate() 1261 __ Mov(jssp, csp); in Generate() 1262 __ SetStackPointer(jssp); in Generate() [all …]
|
D | deoptimizer-arm64.cc | 71 input_->SetRegister(jssp.code(), reinterpret_cast<intptr_t>(frame->sp())); in FillInputFrame()
|
D | simulator-arm64.h | 404 int64_t jssp() { return xreg(kJSSPCode, Reg31IsStackPointer); }
|
D | full-codegen-arm64.cc | 163 DCHECK(jssp.Is(__ StackPointer())); in Generate() 164 __ Sub(x10, jssp, locals_count * kPointerSize); in Generate() 307 DCHECK(jssp.Is(__ StackPointer())); in Generate() 308 __ CompareRoot(jssp, Heap::kStackLimitRootIndex); in Generate() 363 DCHECK(jssp.Is(__ StackPointer())); in EmitBackEdgeBookkeeping()
|
D | assembler-arm64.h | 391 ALIAS_REGISTER(Register, jssp, x28);
|
D | lithium-codegen-arm64.cc | 664 DCHECK(__ StackPointer().Is(jssp)); in GeneratePrologue() 1708 DCHECK(masm()->StackPointer().Is(jssp)); in DoArgumentsElements() 1709 __ Sub(result, jssp, 2 * kPointerSize); in DoArgumentsElements()
|
D | simulator-arm64.cc | 3465 cur = reinterpret_cast<int64_t*>(jssp()); in Debug()
|
/external/v8/src/compiler/arm64/ |
D | code-generator-arm64.cc | 654 __ SetStackPointer(jssp); in AssemblePrologue() 676 __ SetStackPointer(jssp); in AssemblePrologue() 709 __ Mov(jssp, fp); in AssembleReturn()
|
/external/v8/src/ |
D | isolate.cc | 2371 uintptr_t jssp = reinterpret_cast<uintptr_t>(jssp_address); in JsHasOverflowed() local 2372 if (jssp < stack_guard->real_jslimit()) return true; in JsHasOverflowed()
|