/external/vixl/src/vixl/a64/ |
D | macro-assembler-a64.cc | 1675 Stp(bottom_0, bottom_1, MemOperand(StackPointer(), -size, PreIndex)); in PushCPURegList() 1677 Str(bottom_0, MemOperand(StackPointer(), -size, PreIndex)); in PushCPURegList() 1685 Stp(src0, src1, MemOperand(StackPointer(), offset)); in PushCPURegList() 1687 Str(src0, MemOperand(StackPointer(), offset)); in PushCPURegList() 1712 Ldp(dst0, dst1, MemOperand(StackPointer(), offset)); in PopCPURegList() 1714 Ldr(dst0, MemOperand(StackPointer(), offset)); in PopCPURegList() 1721 Ldp(bottom_0, bottom_1, MemOperand(StackPointer(), size, PostIndex)); in PopCPURegList() 1723 Ldr(bottom_0, MemOperand(StackPointer(), size, PostIndex)); in PopCPURegList() 1770 str(src0, MemOperand(StackPointer(), -1 * size, PreIndex)); in PushHelper() 1774 stp(src1, src0, MemOperand(StackPointer(), -2 * size, PreIndex)); in PushHelper() [all …]
|
D | macro-assembler-a64.h | 855 LoadCPURegList(registers, MemOperand(StackPointer(), offset)); in PeekCPURegList() 858 StoreCPURegList(registers, MemOperand(StackPointer(), offset)); in PokeCPURegList() 2990 const Register& StackPointer() const { in StackPointer() function
|
/external/vixl/examples/ |
D | swap-int32.cc | 36 const Register old_stack_pointer = __ StackPointer(); in GenerateSwapInt32() 37 __ Mov(x2, __ StackPointer()); in GenerateSwapInt32() 54 __ Mov(old_stack_pointer, __ StackPointer()); in GenerateSwapInt32()
|
/external/v8/src/arm64/ |
D | macro-assembler-arm64.cc | 935 str(src1, MemOperand(StackPointer(), -size, PreIndex)); in Push() 937 str(src0, MemOperand(StackPointer(), src1.SizeInBytes())); in Push() 1131 str(src0, MemOperand(StackPointer(), -1 * size, PreIndex)); in PushHelper() 1135 stp(src1, src0, MemOperand(StackPointer(), -2 * size, PreIndex)); in PushHelper() 1139 stp(src2, src1, MemOperand(StackPointer(), -3 * size, PreIndex)); in PushHelper() 1140 str(src0, MemOperand(StackPointer(), 2 * size)); in PushHelper() 1146 stp(src3, src2, MemOperand(StackPointer(), -4 * size, PreIndex)); in PushHelper() 1147 stp(src1, src0, MemOperand(StackPointer(), 2 * size)); in PushHelper() 1171 ldr(dst0, MemOperand(StackPointer(), 1 * size, PostIndex)); in PopHelper() 1175 ldp(dst0, dst1, MemOperand(StackPointer(), 2 * size, PostIndex)); in PopHelper() [all …]
|
D | macro-assembler-arm64-inl.h | 1230 Sub(csp, StackPointer(), space); in BumpSystemStackPointer() 1244 Register source = StackPointer(); in BumpSystemStackPointer() 1267 mov(csp, StackPointer()); in SyncSystemStackPointer() 1473 if (csp.Is(StackPointer())) { in Claim() 1479 Sub(StackPointer(), StackPointer(), size); in Claim() 1495 if (!csp.Is(StackPointer())) { in Claim() 1499 Sub(StackPointer(), StackPointer(), size); in Claim() 1514 if (!csp.Is(StackPointer())) { in ClaimBySMI() 1518 Sub(StackPointer(), StackPointer(), size); in ClaimBySMI() 1530 Add(StackPointer(), StackPointer(), size); in Drop() [all …]
|
D | deoptimizer-arm64.cc | 158 __ Add(fp_to_sp, masm()->StackPointer(), in Generate() 209 __ Add(unwind_limit, unwind_limit, __ StackPointer()); in Generate() 222 __ Cmp(unwind_limit, __ StackPointer()); in Generate()
|
D | code-stubs-arm64.cc | 134 DCHECK(jssp.Is(masm->StackPointer())); in Generate() 1082 DCHECK(jssp.Is(__ StackPointer())); in Generate() 1111 DCHECK(csp.Is(__ StackPointer())); in Generate() 1161 DCHECK(csp.Is(__ StackPointer())); in Generate() 1208 DCHECK(jssp.Is(__ StackPointer())); in Generate() 1238 DCHECK(csp.Is(masm->StackPointer())); in Generate() 1290 DCHECK(jssp.Is(__ StackPointer())); in Generate() 1446 DCHECK(jssp.Is(__ StackPointer())); in Generate() 2242 DCHECK(jssp.Is(__ StackPointer())); in Generate() 2279 DCHECK(jssp.Is(__ StackPointer())); in Generate() [all …]
|
D | builtins-arm64.cc | 408 __ CompareRoot(masm->StackPointer(), Heap::kStackLimitRootIndex); in Generate_InOptimizationQueue() 920 DCHECK(jssp.Is(__ StackPointer())); in Generate_InterpreterEntryTrampoline() 1904 __ Sub(x10, masm->StackPointer(), x10); in Generate_Apply()
|
D | macro-assembler-arm64.h | 825 inline const Register& StackPointer() const { in StackPointer() function 836 Bic(csp, StackPointer(), sp_alignment - 1); in AlignAndSetCSPForFrame()
|
/external/v8/src/regexp/arm64/ |
D | regexp-macro-assembler-arm64.cc | 723 DCHECK(csp.Is(__ StackPointer())); in GetCode() 1016 DCHECK(csp.Is(__ StackPointer())); in GetCode() 1374 DCHECK(csp.Is(__ StackPointer())); in CallCheckStackGuardState() 1404 DCHECK(csp.Is(__ StackPointer())); in CallCheckStackGuardState() 1453 DCHECK(csp.Is(__ StackPointer())); in CheckPreemption() 1564 DCHECK(csp.Is(__ StackPointer())); in RestoreLinkRegister() 1571 DCHECK(csp.Is(__ StackPointer())); in SaveLinkRegister()
|
/external/v8/src/debug/arm64/ |
D | debug-arm64.cc | 129 __ Sub(masm->StackPointer(), fp, kPointerSize); in GenerateFrameDropperLiveEdit()
|
/external/v8/test/cctest/ |
D | test-hashing.cc | 76 Register old_stack_pointer = __ StackPointer(); in generate()
|
D | test-assembler-arm64.cc | 8310 CHECK(csp.Is(__ StackPointer())); in TEST() 8312 CHECK(x0.Is(__ StackPointer())); in TEST() 8314 CHECK(jssp.Is(__ StackPointer())); in TEST() 8316 CHECK(csp.Is(__ StackPointer())); in TEST() 8549 CHECK(__ StackPointer().Is(csp)); in TEST() 8550 __ Mov(x4, __ StackPointer()); in TEST() 8564 __ Mov(csp, __ StackPointer()); in TEST() 8646 CHECK(__ StackPointer().Is(csp)); in PushPopJsspSimpleHelper() 8647 __ Mov(jssp, __ StackPointer()); in PushPopJsspSimpleHelper() 8712 __ Mov(csp, __ StackPointer()); in PushPopJsspSimpleHelper() [all …]
|
D | test-utils-arm64.cc | 323 CHECK(__ StackPointer().Is(csp)); in Dump()
|
/external/valgrind/VEX/priv/ |
D | host_mips_isel.c | 195 HReg sp = StackPointer(mode64); in add_to_sp() 207 HReg sp = StackPointer(mode64); in sub_from_sp() 297 am_addr = MIPSAMode_IR(0, StackPointer(mode64)); in set_MIPS_rounding_mode() 311 am_addr = MIPSAMode_IR(0, StackPointer(mode64)); in set_MIPS_rounding_default() 364 am_addr0 = MIPSAMode_IR(0, StackPointer(mode64)); in mk_LoadRR32toFPR() 365 am_addr1 = MIPSAMode_IR(4, StackPointer(mode64)); in mk_LoadRR32toFPR() 2777 am_addr = MIPSAMode_IR(0, StackPointer(mode64)); in iselInt64Expr_wrk() 2930 am_addr = MIPSAMode_IR(0, StackPointer(mode64)); in iselInt64Expr_wrk() 3237 am_addr = MIPSAMode_IR(0, StackPointer(mode64)); in iselFltExpr_wrk() 3270 am_addr = MIPSAMode_IR(0, StackPointer(mode64)); in iselFltExpr_wrk() [all …]
|
D | host_mips_defs.h | 124 #define StackPointer(_mode64) hregMIPS_GPR29(_mode64) macro
|
/external/vixl/test/ |
D | test-utils-a64.cc | 355 VIXL_ASSERT(__ StackPointer().Is(sp)); in Dump()
|
D | test-assembler-a64.cc | 12746 assert(sp.Is(__ StackPointer())); in TEST() 12748 assert(x0.Is(__ StackPointer())); in TEST() 12750 assert(x28.Is(__ StackPointer())); in TEST() 12752 assert(sp.Is(__ StackPointer())); in TEST() 12985 VIXL_ASSERT(__ StackPointer().Is(sp)); in TEST() 12986 __ Mov(x4, __ StackPointer()); in TEST() 13000 __ Mov(sp, __ StackPointer()); in TEST() 13290 VIXL_ASSERT(__ StackPointer().Is(sp)); in PushPopXRegSimpleHelper() 13291 __ Mov(stack_pointer, __ StackPointer()); in PushPopXRegSimpleHelper() 13350 __ Mov(sp, __ StackPointer()); in PushPopXRegSimpleHelper() [all …]
|
/external/v8/src/compiler/arm64/ |
D | code-generator-arm64.cc | 223 return MemOperand(offset.from_stack_pointer() ? masm->StackPointer() : fp, in ToMemOperand() 607 __ mov(i.OutputRegister(), masm()->StackPointer()); in AssembleArchInstruction() 1386 if (csp.Is(masm()->StackPointer()) && frame()->needs_frame()) { in AssemblePrologue()
|
/external/v8/src/crankshaft/arm64/ |
D | lithium-codegen-arm64.cc | 617 DCHECK(__ StackPointer().Is(jssp)); in GeneratePrologue() 758 __ Add(fp, __ StackPointer(), in GenerateDeferredCode() 839 __ Add(fp, __ StackPointer(), 2 * kPointerSize); in GenerateJumpTable() 1185 return MemOperand(masm()->StackPointer(), jssp_offset); in ToMemOperand() 1192 return MemOperand(masm()->StackPointer(), in ToMemOperand() 1598 DCHECK(masm()->StackPointer().Is(jssp)); in DoArgumentsElements() 4464 Register stack_pointer = masm()->StackPointer(); in DoReturn() 4760 __ CompareRoot(masm()->StackPointer(), Heap::kStackLimitRootIndex); in DoStackCheck() 4776 __ CompareRoot(masm()->StackPointer(), Heap::kStackLimitRootIndex); in DoStackCheck()
|
/external/v8/src/full-codegen/arm64/ |
D | full-codegen-arm64.cc | 148 DCHECK(jssp.Is(__ StackPointer())); in Generate() 343 DCHECK(jssp.Is(__ StackPointer())); in Generate() 397 DCHECK(jssp.Is(__ StackPointer())); in EmitBackEdgeBookkeeping() 468 const Register& current_sp = __ StackPointer(); in EmitReturnSequence() 4401 __ Cmp(__ StackPointer(), x1); in VisitYield() 4592 __ Add(fp, __ StackPointer(), kPointerSize * 2); in EmitGeneratorResume()
|
/external/v8/src/ic/arm64/ |
D | handler-compiler-arm64.cc | 679 __ Add(args_addr, __ StackPointer(), kPointerSize); in GenerateLoadCallback()
|