1 /* 2 * Array put, 64 bits. vBB[vCC] <- vAA. 3 * 4 * Arrays of long/double are 64-bit aligned, so it's okay to use STRD. 5 */ 6 /* aput-wide vAA, vBB, vCC */ 7 FETCH(a0, 1) # a0 <- CCBB 8 GET_OPA(t0) # t0 <- AA 9 and a2, a0, 255 # a2 <- BB 10 srl a3, a0, 8 # a3 <- CC 11 GET_VREG(a0, a2) # a0 <- vBB (array object) 12 GET_VREG(a1, a3) # a1 <- vCC (requested index) 13 # null array object? 14 beqz a0, common_errNullObject # yes, bail 15 LOAD_base_offMirrorArray_length(a3, a0) # a3 <- arrayObj->length 16 EAS3(a0, a0, a1) # a0 <- arrayObj + index*width 17 EAS2(rOBJ, rFP, t0) # rOBJ <- &fp[AA] 18 # compare unsigned index, length 19 bgeu a1, a3, common_errArrayIndex # index >= length, bail 20 21 FETCH_ADVANCE_INST(2) # advance rPC, load rINST 22 LOAD64(a2, a3, rOBJ) # a2/a3 <- vAA/vAA+1 23 GET_INST_OPCODE(t0) # extract opcode from rINST 24 STORE64_off(a2, a3, a0, MIRROR_WIDE_ARRAY_DATA_OFFSET) # a2/a3 <- vBB[vCC] 25 GOTO_OPCODE(t0) # jump to next instruction 26