/art/runtime/interpreter/mterp/arm64ng/ |
D | object.S | 76 $load x0, [x3, x0] 79 $load w0, [x3, x0] 84 $load w0, [x3, x0] 100 mov x3, #0 109 add x3, x3, x0 111 $volatile_load x0, [x3] 114 $volatile_load w0, [x3] 118 $volatile_load w0, [x3] 176 mov x3, x26 178 mov x3, #0 [all …]
|
D | other.S | 96 orr x0, x0, x3, lsl #48 // w0<- HHHHhhhhBBBBbbbb 193 ldr x3, [xSELF, #THREAD_EXCEPTION_OFFSET] 254 GET_VREG_WIDE x3, w3 257 SET_VREG_WIDE x3, w2 265 GET_VREG_WIDE x3, w3 267 SET_VREG_WIDE x3, w2 276 GET_VREG_WIDE x3, w3 279 SET_VREG_WIDE x3, w2
|
D | main.S | 476 stp x2, x3, [sp, #16] 486 ldp x2, x3, [sp, #16] 568 asr x2, x3, #12 575 ubfx x2, x3, #8, #4 582 ubfx x2, x3, #4, #4 594 and x2, x3, #0xf 628 add x3, x3, #1 631 add x3, x3, #1 635 add ip2, xREFS, x3, lsl #2 // pointer to first argument in reference array 637 add x5, xFP, x3, lsl #2 // pointer to first argument in register array [all …]
|
D | array.S | 153 mov x3, xPC
|
/art/runtime/interpreter/mterp/arm64/ |
D | other.S | 11 mov x3, xSELF 84 mov x3, xSELF 103 orr x0, x0, x3, lsl #48 // w0<- HHHHhhhhBBBBbbbb 208 ldr x3, [xSELF, #THREAD_EXCEPTION_OFFSET] 274 GET_VREG_WIDE x3, w3 277 SET_VREG_WIDE x3, w2 285 GET_VREG_WIDE x3, w3 287 SET_VREG_WIDE x3, w2 296 GET_VREG_WIDE x3, w3 299 SET_VREG_WIDE x3, w2
|
D | invoke.S | 12 mov x3, xINST 33 mov x3, xINST
|
D | object.S | 9 mov x3, xSELF // arg3: Thread* self 27 mov x3, xSELF // w3<- self 114 mov x3, xSELF // w3<- self
|
D | array.S | 230 mov x3, xSELF
|
D | main.S | 412 str x3, [x2, #SHADOWFRAME_RESULT_REGISTER_OFFSET]
|
/art/runtime/arch/arm64/ |
D | quick_entrypoints_arm64.S | 116 SAVE_TWO_REGS x2, x3, 288 174 RESTORE_TWO_REGS x2, x3, 288 339 mov x3, sp 376 mov xSELF, x3 // Move thread pointer into SELF register. 583 LOADREG x8 8 x3 .LfillRegisters 708 LOADREG x8 8 x3 .LfillRegisters2 754 SAVE_TWO_REGS_INCREASE_FRAME x3, x4, SAVE_SIZE 787 RESTORE_TWO_REGS_DECREASE_FRAME x3, x4, SAVE_SIZE 790 str x0, [x3] 853 ldp x2, x3, [x0, #16] [all …]
|
D | jni_entrypoints_arm64.S | 44 stp x2, x3, [sp, #-16]! 69 ldp x2, x3, [sp], #16 103 stp x2, x3, [sp, #16] 122 ldp x2, x3, [sp, #16] 167 stp x3, x4, [x13, #96] 243 ldp x3, x4, [x29, #96]
|
D | asm_support_arm64.S | 247 SAVE_TWO_REGS_BASE \base, x3, x4, 96 275 RESTORE_TWO_REGS_BASE \base, x3, x4, 96
|
D | memcmp16_arm64.S | 34 #define data1 x3
|
/art/test/541-regression-inlined-deopt/smali/ |
D | TestCase.smali | 25 const v1, 0x3 50 const v5, 0x3
|
/art/test/458-checker-instruct-simplification/smali/ |
D | SmaliTests.smali | 34 const v1, 0x3 56 const v1, 0x3 76 const v1, 0x3 98 const v1, 0x3 120 const v1, 0x3 142 const v1, 0x3 166 const v1, 0x3 188 const v1, 0x3
|
/art/test/452-multiple-returns2/smali/ |
D | MultipleReturns.smali | 37 const/4 v1, 0x3
|
/art/test/552-checker-primitive-typeprop/smali/ |
D | SsaBuilder.smali | 41 const v1, 0x3
|
/art/test/586-checker-null-array-get/smali/ |
D | SmaliTests.smali | 35 const/4 v6, 0x3 104 const/4 v6, 0x3
|
/art/test/546-regression-simplify-catch/smali/ |
D | TestCase.smali | 56 const v1, 0x3
|
/art/test/121-modifiers/smali/ |
D | NonInf.smali | 64 const/4 v0, 0x3
|
/art/test/449-checker-bce/src/ |
D | Main.java | 1794 int[] x3 = new int[10]; in main() local 1797 modArrayIndex5(x3, 5); in main() 1802 if (x1[i] != e1 || x2[i] != e2 || x3[i] != e3) { in main()
|
/art/test/530-checker-loops2/src/ |
D | Main.java | 1236 char[] x3 = { 3 }; in main() 1243 dynamicBCEAndConstantIndicesAllPrimTypes(x, x1, x2, x3, x4, x5, x6, x7, x8, 0, 10)); in main()
|
/art/compiler/optimizing/ |
D | code_generator_arm64.h | 60 vixl::aarch64::x3, 159 vixl::aarch64::x3,
|
/art/compiler/utils/arm64/ |
D | managed_register_arm64_test.cc | 598 EXPECT_TRUE(vixl::aarch64::x3.Is(Arm64Assembler::reg_x(X3))); in TEST()
|
/art/test/510-checker-try-catch/smali/ |
D | Builder.smali | 129 const/4 p0, -0x3
|