1 /*
2 * Copyright (C) 2011 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17 #include "assembler_x86.h"
18
19 #include "base/casts.h"
20 #include "entrypoints/quick/quick_entrypoints.h"
21 #include "memory_region.h"
22 #include "thread.h"
23
24 namespace art {
25 namespace x86 {
26
operator <<(std::ostream & os,const XmmRegister & reg)27 std::ostream& operator<<(std::ostream& os, const XmmRegister& reg) {
28 return os << "XMM" << static_cast<int>(reg);
29 }
30
operator <<(std::ostream & os,const X87Register & reg)31 std::ostream& operator<<(std::ostream& os, const X87Register& reg) {
32 return os << "ST" << static_cast<int>(reg);
33 }
34
call(Register reg)35 void X86Assembler::call(Register reg) {
36 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
37 EmitUint8(0xFF);
38 EmitRegisterOperand(2, reg);
39 }
40
41
call(const Address & address)42 void X86Assembler::call(const Address& address) {
43 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
44 EmitUint8(0xFF);
45 EmitOperand(2, address);
46 }
47
48
call(Label * label)49 void X86Assembler::call(Label* label) {
50 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
51 EmitUint8(0xE8);
52 static const int kSize = 5;
53 EmitLabel(label, kSize);
54 }
55
56
call(const ExternalLabel & label)57 void X86Assembler::call(const ExternalLabel& label) {
58 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
59 intptr_t call_start = buffer_.GetPosition();
60 EmitUint8(0xE8);
61 EmitInt32(label.address());
62 static const intptr_t kCallExternalLabelSize = 5;
63 DCHECK_EQ((buffer_.GetPosition() - call_start), kCallExternalLabelSize);
64 }
65
66
pushl(Register reg)67 void X86Assembler::pushl(Register reg) {
68 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
69 EmitUint8(0x50 + reg);
70 }
71
72
pushl(const Address & address)73 void X86Assembler::pushl(const Address& address) {
74 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
75 EmitUint8(0xFF);
76 EmitOperand(6, address);
77 }
78
79
pushl(const Immediate & imm)80 void X86Assembler::pushl(const Immediate& imm) {
81 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
82 if (imm.is_int8()) {
83 EmitUint8(0x6A);
84 EmitUint8(imm.value() & 0xFF);
85 } else {
86 EmitUint8(0x68);
87 EmitImmediate(imm);
88 }
89 }
90
91
popl(Register reg)92 void X86Assembler::popl(Register reg) {
93 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
94 EmitUint8(0x58 + reg);
95 }
96
97
popl(const Address & address)98 void X86Assembler::popl(const Address& address) {
99 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
100 EmitUint8(0x8F);
101 EmitOperand(0, address);
102 }
103
104
movl(Register dst,const Immediate & imm)105 void X86Assembler::movl(Register dst, const Immediate& imm) {
106 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
107 EmitUint8(0xB8 + dst);
108 EmitImmediate(imm);
109 }
110
111
movl(Register dst,Register src)112 void X86Assembler::movl(Register dst, Register src) {
113 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
114 EmitUint8(0x89);
115 EmitRegisterOperand(src, dst);
116 }
117
118
movl(Register dst,const Address & src)119 void X86Assembler::movl(Register dst, const Address& src) {
120 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
121 EmitUint8(0x8B);
122 EmitOperand(dst, src);
123 }
124
125
movl(const Address & dst,Register src)126 void X86Assembler::movl(const Address& dst, Register src) {
127 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
128 EmitUint8(0x89);
129 EmitOperand(src, dst);
130 }
131
132
movl(const Address & dst,const Immediate & imm)133 void X86Assembler::movl(const Address& dst, const Immediate& imm) {
134 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
135 EmitUint8(0xC7);
136 EmitOperand(0, dst);
137 EmitImmediate(imm);
138 }
139
movl(const Address & dst,Label * lbl)140 void X86Assembler::movl(const Address& dst, Label* lbl) {
141 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
142 EmitUint8(0xC7);
143 EmitOperand(0, dst);
144 EmitLabel(lbl, dst.length_ + 5);
145 }
146
movzxb(Register dst,ByteRegister src)147 void X86Assembler::movzxb(Register dst, ByteRegister src) {
148 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
149 EmitUint8(0x0F);
150 EmitUint8(0xB6);
151 EmitRegisterOperand(dst, src);
152 }
153
154
movzxb(Register dst,const Address & src)155 void X86Assembler::movzxb(Register dst, const Address& src) {
156 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
157 EmitUint8(0x0F);
158 EmitUint8(0xB6);
159 EmitOperand(dst, src);
160 }
161
162
movsxb(Register dst,ByteRegister src)163 void X86Assembler::movsxb(Register dst, ByteRegister src) {
164 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
165 EmitUint8(0x0F);
166 EmitUint8(0xBE);
167 EmitRegisterOperand(dst, src);
168 }
169
170
movsxb(Register dst,const Address & src)171 void X86Assembler::movsxb(Register dst, const Address& src) {
172 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
173 EmitUint8(0x0F);
174 EmitUint8(0xBE);
175 EmitOperand(dst, src);
176 }
177
178
movb(Register,const Address &)179 void X86Assembler::movb(Register /*dst*/, const Address& /*src*/) {
180 LOG(FATAL) << "Use movzxb or movsxb instead.";
181 }
182
183
movb(const Address & dst,ByteRegister src)184 void X86Assembler::movb(const Address& dst, ByteRegister src) {
185 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
186 EmitUint8(0x88);
187 EmitOperand(src, dst);
188 }
189
190
movb(const Address & dst,const Immediate & imm)191 void X86Assembler::movb(const Address& dst, const Immediate& imm) {
192 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
193 EmitUint8(0xC6);
194 EmitOperand(EAX, dst);
195 CHECK(imm.is_int8());
196 EmitUint8(imm.value() & 0xFF);
197 }
198
199
movzxw(Register dst,Register src)200 void X86Assembler::movzxw(Register dst, Register src) {
201 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
202 EmitUint8(0x0F);
203 EmitUint8(0xB7);
204 EmitRegisterOperand(dst, src);
205 }
206
207
movzxw(Register dst,const Address & src)208 void X86Assembler::movzxw(Register dst, const Address& src) {
209 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
210 EmitUint8(0x0F);
211 EmitUint8(0xB7);
212 EmitOperand(dst, src);
213 }
214
215
movsxw(Register dst,Register src)216 void X86Assembler::movsxw(Register dst, Register src) {
217 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
218 EmitUint8(0x0F);
219 EmitUint8(0xBF);
220 EmitRegisterOperand(dst, src);
221 }
222
223
movsxw(Register dst,const Address & src)224 void X86Assembler::movsxw(Register dst, const Address& src) {
225 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
226 EmitUint8(0x0F);
227 EmitUint8(0xBF);
228 EmitOperand(dst, src);
229 }
230
231
movw(Register,const Address &)232 void X86Assembler::movw(Register /*dst*/, const Address& /*src*/) {
233 LOG(FATAL) << "Use movzxw or movsxw instead.";
234 }
235
236
movw(const Address & dst,Register src)237 void X86Assembler::movw(const Address& dst, Register src) {
238 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
239 EmitOperandSizeOverride();
240 EmitUint8(0x89);
241 EmitOperand(src, dst);
242 }
243
244
leal(Register dst,const Address & src)245 void X86Assembler::leal(Register dst, const Address& src) {
246 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
247 EmitUint8(0x8D);
248 EmitOperand(dst, src);
249 }
250
251
cmovl(Condition condition,Register dst,Register src)252 void X86Assembler::cmovl(Condition condition, Register dst, Register src) {
253 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
254 EmitUint8(0x0F);
255 EmitUint8(0x40 + condition);
256 EmitRegisterOperand(dst, src);
257 }
258
259
setb(Condition condition,Register dst)260 void X86Assembler::setb(Condition condition, Register dst) {
261 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
262 EmitUint8(0x0F);
263 EmitUint8(0x90 + condition);
264 EmitOperand(0, Operand(dst));
265 }
266
267
movss(XmmRegister dst,const Address & src)268 void X86Assembler::movss(XmmRegister dst, const Address& src) {
269 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
270 EmitUint8(0xF3);
271 EmitUint8(0x0F);
272 EmitUint8(0x10);
273 EmitOperand(dst, src);
274 }
275
276
movss(const Address & dst,XmmRegister src)277 void X86Assembler::movss(const Address& dst, XmmRegister src) {
278 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
279 EmitUint8(0xF3);
280 EmitUint8(0x0F);
281 EmitUint8(0x11);
282 EmitOperand(src, dst);
283 }
284
285
movss(XmmRegister dst,XmmRegister src)286 void X86Assembler::movss(XmmRegister dst, XmmRegister src) {
287 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
288 EmitUint8(0xF3);
289 EmitUint8(0x0F);
290 EmitUint8(0x11);
291 EmitXmmRegisterOperand(src, dst);
292 }
293
294
movd(XmmRegister dst,Register src)295 void X86Assembler::movd(XmmRegister dst, Register src) {
296 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
297 EmitUint8(0x66);
298 EmitUint8(0x0F);
299 EmitUint8(0x6E);
300 EmitOperand(dst, Operand(src));
301 }
302
303
movd(Register dst,XmmRegister src)304 void X86Assembler::movd(Register dst, XmmRegister src) {
305 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
306 EmitUint8(0x66);
307 EmitUint8(0x0F);
308 EmitUint8(0x7E);
309 EmitOperand(src, Operand(dst));
310 }
311
312
addss(XmmRegister dst,XmmRegister src)313 void X86Assembler::addss(XmmRegister dst, XmmRegister src) {
314 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
315 EmitUint8(0xF3);
316 EmitUint8(0x0F);
317 EmitUint8(0x58);
318 EmitXmmRegisterOperand(dst, src);
319 }
320
321
addss(XmmRegister dst,const Address & src)322 void X86Assembler::addss(XmmRegister dst, const Address& src) {
323 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
324 EmitUint8(0xF3);
325 EmitUint8(0x0F);
326 EmitUint8(0x58);
327 EmitOperand(dst, src);
328 }
329
330
subss(XmmRegister dst,XmmRegister src)331 void X86Assembler::subss(XmmRegister dst, XmmRegister src) {
332 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
333 EmitUint8(0xF3);
334 EmitUint8(0x0F);
335 EmitUint8(0x5C);
336 EmitXmmRegisterOperand(dst, src);
337 }
338
339
subss(XmmRegister dst,const Address & src)340 void X86Assembler::subss(XmmRegister dst, const Address& src) {
341 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
342 EmitUint8(0xF3);
343 EmitUint8(0x0F);
344 EmitUint8(0x5C);
345 EmitOperand(dst, src);
346 }
347
348
mulss(XmmRegister dst,XmmRegister src)349 void X86Assembler::mulss(XmmRegister dst, XmmRegister src) {
350 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
351 EmitUint8(0xF3);
352 EmitUint8(0x0F);
353 EmitUint8(0x59);
354 EmitXmmRegisterOperand(dst, src);
355 }
356
357
mulss(XmmRegister dst,const Address & src)358 void X86Assembler::mulss(XmmRegister dst, const Address& src) {
359 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
360 EmitUint8(0xF3);
361 EmitUint8(0x0F);
362 EmitUint8(0x59);
363 EmitOperand(dst, src);
364 }
365
366
divss(XmmRegister dst,XmmRegister src)367 void X86Assembler::divss(XmmRegister dst, XmmRegister src) {
368 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
369 EmitUint8(0xF3);
370 EmitUint8(0x0F);
371 EmitUint8(0x5E);
372 EmitXmmRegisterOperand(dst, src);
373 }
374
375
divss(XmmRegister dst,const Address & src)376 void X86Assembler::divss(XmmRegister dst, const Address& src) {
377 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
378 EmitUint8(0xF3);
379 EmitUint8(0x0F);
380 EmitUint8(0x5E);
381 EmitOperand(dst, src);
382 }
383
384
flds(const Address & src)385 void X86Assembler::flds(const Address& src) {
386 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
387 EmitUint8(0xD9);
388 EmitOperand(0, src);
389 }
390
391
fstps(const Address & dst)392 void X86Assembler::fstps(const Address& dst) {
393 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
394 EmitUint8(0xD9);
395 EmitOperand(3, dst);
396 }
397
398
movsd(XmmRegister dst,const Address & src)399 void X86Assembler::movsd(XmmRegister dst, const Address& src) {
400 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
401 EmitUint8(0xF2);
402 EmitUint8(0x0F);
403 EmitUint8(0x10);
404 EmitOperand(dst, src);
405 }
406
407
movsd(const Address & dst,XmmRegister src)408 void X86Assembler::movsd(const Address& dst, XmmRegister src) {
409 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
410 EmitUint8(0xF2);
411 EmitUint8(0x0F);
412 EmitUint8(0x11);
413 EmitOperand(src, dst);
414 }
415
416
movsd(XmmRegister dst,XmmRegister src)417 void X86Assembler::movsd(XmmRegister dst, XmmRegister src) {
418 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
419 EmitUint8(0xF2);
420 EmitUint8(0x0F);
421 EmitUint8(0x11);
422 EmitXmmRegisterOperand(src, dst);
423 }
424
425
addsd(XmmRegister dst,XmmRegister src)426 void X86Assembler::addsd(XmmRegister dst, XmmRegister src) {
427 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
428 EmitUint8(0xF2);
429 EmitUint8(0x0F);
430 EmitUint8(0x58);
431 EmitXmmRegisterOperand(dst, src);
432 }
433
434
addsd(XmmRegister dst,const Address & src)435 void X86Assembler::addsd(XmmRegister dst, const Address& src) {
436 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
437 EmitUint8(0xF2);
438 EmitUint8(0x0F);
439 EmitUint8(0x58);
440 EmitOperand(dst, src);
441 }
442
443
subsd(XmmRegister dst,XmmRegister src)444 void X86Assembler::subsd(XmmRegister dst, XmmRegister src) {
445 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
446 EmitUint8(0xF2);
447 EmitUint8(0x0F);
448 EmitUint8(0x5C);
449 EmitXmmRegisterOperand(dst, src);
450 }
451
452
subsd(XmmRegister dst,const Address & src)453 void X86Assembler::subsd(XmmRegister dst, const Address& src) {
454 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
455 EmitUint8(0xF2);
456 EmitUint8(0x0F);
457 EmitUint8(0x5C);
458 EmitOperand(dst, src);
459 }
460
461
mulsd(XmmRegister dst,XmmRegister src)462 void X86Assembler::mulsd(XmmRegister dst, XmmRegister src) {
463 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
464 EmitUint8(0xF2);
465 EmitUint8(0x0F);
466 EmitUint8(0x59);
467 EmitXmmRegisterOperand(dst, src);
468 }
469
470
mulsd(XmmRegister dst,const Address & src)471 void X86Assembler::mulsd(XmmRegister dst, const Address& src) {
472 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
473 EmitUint8(0xF2);
474 EmitUint8(0x0F);
475 EmitUint8(0x59);
476 EmitOperand(dst, src);
477 }
478
479
divsd(XmmRegister dst,XmmRegister src)480 void X86Assembler::divsd(XmmRegister dst, XmmRegister src) {
481 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
482 EmitUint8(0xF2);
483 EmitUint8(0x0F);
484 EmitUint8(0x5E);
485 EmitXmmRegisterOperand(dst, src);
486 }
487
488
divsd(XmmRegister dst,const Address & src)489 void X86Assembler::divsd(XmmRegister dst, const Address& src) {
490 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
491 EmitUint8(0xF2);
492 EmitUint8(0x0F);
493 EmitUint8(0x5E);
494 EmitOperand(dst, src);
495 }
496
497
cvtsi2ss(XmmRegister dst,Register src)498 void X86Assembler::cvtsi2ss(XmmRegister dst, Register src) {
499 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
500 EmitUint8(0xF3);
501 EmitUint8(0x0F);
502 EmitUint8(0x2A);
503 EmitOperand(dst, Operand(src));
504 }
505
506
cvtsi2sd(XmmRegister dst,Register src)507 void X86Assembler::cvtsi2sd(XmmRegister dst, Register src) {
508 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
509 EmitUint8(0xF2);
510 EmitUint8(0x0F);
511 EmitUint8(0x2A);
512 EmitOperand(dst, Operand(src));
513 }
514
515
cvtss2si(Register dst,XmmRegister src)516 void X86Assembler::cvtss2si(Register dst, XmmRegister src) {
517 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
518 EmitUint8(0xF3);
519 EmitUint8(0x0F);
520 EmitUint8(0x2D);
521 EmitXmmRegisterOperand(dst, src);
522 }
523
524
cvtss2sd(XmmRegister dst,XmmRegister src)525 void X86Assembler::cvtss2sd(XmmRegister dst, XmmRegister src) {
526 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
527 EmitUint8(0xF3);
528 EmitUint8(0x0F);
529 EmitUint8(0x5A);
530 EmitXmmRegisterOperand(dst, src);
531 }
532
533
cvtsd2si(Register dst,XmmRegister src)534 void X86Assembler::cvtsd2si(Register dst, XmmRegister src) {
535 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
536 EmitUint8(0xF2);
537 EmitUint8(0x0F);
538 EmitUint8(0x2D);
539 EmitXmmRegisterOperand(dst, src);
540 }
541
542
cvttss2si(Register dst,XmmRegister src)543 void X86Assembler::cvttss2si(Register dst, XmmRegister src) {
544 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
545 EmitUint8(0xF3);
546 EmitUint8(0x0F);
547 EmitUint8(0x2C);
548 EmitXmmRegisterOperand(dst, src);
549 }
550
551
cvttsd2si(Register dst,XmmRegister src)552 void X86Assembler::cvttsd2si(Register dst, XmmRegister src) {
553 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
554 EmitUint8(0xF2);
555 EmitUint8(0x0F);
556 EmitUint8(0x2C);
557 EmitXmmRegisterOperand(dst, src);
558 }
559
560
cvtsd2ss(XmmRegister dst,XmmRegister src)561 void X86Assembler::cvtsd2ss(XmmRegister dst, XmmRegister src) {
562 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
563 EmitUint8(0xF2);
564 EmitUint8(0x0F);
565 EmitUint8(0x5A);
566 EmitXmmRegisterOperand(dst, src);
567 }
568
569
cvtdq2pd(XmmRegister dst,XmmRegister src)570 void X86Assembler::cvtdq2pd(XmmRegister dst, XmmRegister src) {
571 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
572 EmitUint8(0xF3);
573 EmitUint8(0x0F);
574 EmitUint8(0xE6);
575 EmitXmmRegisterOperand(dst, src);
576 }
577
578
comiss(XmmRegister a,XmmRegister b)579 void X86Assembler::comiss(XmmRegister a, XmmRegister b) {
580 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
581 EmitUint8(0x0F);
582 EmitUint8(0x2F);
583 EmitXmmRegisterOperand(a, b);
584 }
585
586
comisd(XmmRegister a,XmmRegister b)587 void X86Assembler::comisd(XmmRegister a, XmmRegister b) {
588 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
589 EmitUint8(0x66);
590 EmitUint8(0x0F);
591 EmitUint8(0x2F);
592 EmitXmmRegisterOperand(a, b);
593 }
594
595
sqrtsd(XmmRegister dst,XmmRegister src)596 void X86Assembler::sqrtsd(XmmRegister dst, XmmRegister src) {
597 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
598 EmitUint8(0xF2);
599 EmitUint8(0x0F);
600 EmitUint8(0x51);
601 EmitXmmRegisterOperand(dst, src);
602 }
603
604
sqrtss(XmmRegister dst,XmmRegister src)605 void X86Assembler::sqrtss(XmmRegister dst, XmmRegister src) {
606 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
607 EmitUint8(0xF3);
608 EmitUint8(0x0F);
609 EmitUint8(0x51);
610 EmitXmmRegisterOperand(dst, src);
611 }
612
613
xorpd(XmmRegister dst,const Address & src)614 void X86Assembler::xorpd(XmmRegister dst, const Address& src) {
615 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
616 EmitUint8(0x66);
617 EmitUint8(0x0F);
618 EmitUint8(0x57);
619 EmitOperand(dst, src);
620 }
621
622
xorpd(XmmRegister dst,XmmRegister src)623 void X86Assembler::xorpd(XmmRegister dst, XmmRegister src) {
624 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
625 EmitUint8(0x66);
626 EmitUint8(0x0F);
627 EmitUint8(0x57);
628 EmitXmmRegisterOperand(dst, src);
629 }
630
631
xorps(XmmRegister dst,const Address & src)632 void X86Assembler::xorps(XmmRegister dst, const Address& src) {
633 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
634 EmitUint8(0x0F);
635 EmitUint8(0x57);
636 EmitOperand(dst, src);
637 }
638
639
xorps(XmmRegister dst,XmmRegister src)640 void X86Assembler::xorps(XmmRegister dst, XmmRegister src) {
641 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
642 EmitUint8(0x0F);
643 EmitUint8(0x57);
644 EmitXmmRegisterOperand(dst, src);
645 }
646
647
andpd(XmmRegister dst,const Address & src)648 void X86Assembler::andpd(XmmRegister dst, const Address& src) {
649 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
650 EmitUint8(0x66);
651 EmitUint8(0x0F);
652 EmitUint8(0x54);
653 EmitOperand(dst, src);
654 }
655
656
fldl(const Address & src)657 void X86Assembler::fldl(const Address& src) {
658 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
659 EmitUint8(0xDD);
660 EmitOperand(0, src);
661 }
662
663
fstpl(const Address & dst)664 void X86Assembler::fstpl(const Address& dst) {
665 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
666 EmitUint8(0xDD);
667 EmitOperand(3, dst);
668 }
669
670
fnstcw(const Address & dst)671 void X86Assembler::fnstcw(const Address& dst) {
672 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
673 EmitUint8(0xD9);
674 EmitOperand(7, dst);
675 }
676
677
fldcw(const Address & src)678 void X86Assembler::fldcw(const Address& src) {
679 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
680 EmitUint8(0xD9);
681 EmitOperand(5, src);
682 }
683
684
fistpl(const Address & dst)685 void X86Assembler::fistpl(const Address& dst) {
686 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
687 EmitUint8(0xDF);
688 EmitOperand(7, dst);
689 }
690
691
fistps(const Address & dst)692 void X86Assembler::fistps(const Address& dst) {
693 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
694 EmitUint8(0xDB);
695 EmitOperand(3, dst);
696 }
697
698
fildl(const Address & src)699 void X86Assembler::fildl(const Address& src) {
700 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
701 EmitUint8(0xDF);
702 EmitOperand(5, src);
703 }
704
705
fincstp()706 void X86Assembler::fincstp() {
707 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
708 EmitUint8(0xD9);
709 EmitUint8(0xF7);
710 }
711
712
ffree(const Immediate & index)713 void X86Assembler::ffree(const Immediate& index) {
714 CHECK_LT(index.value(), 7);
715 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
716 EmitUint8(0xDD);
717 EmitUint8(0xC0 + index.value());
718 }
719
720
fsin()721 void X86Assembler::fsin() {
722 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
723 EmitUint8(0xD9);
724 EmitUint8(0xFE);
725 }
726
727
fcos()728 void X86Assembler::fcos() {
729 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
730 EmitUint8(0xD9);
731 EmitUint8(0xFF);
732 }
733
734
fptan()735 void X86Assembler::fptan() {
736 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
737 EmitUint8(0xD9);
738 EmitUint8(0xF2);
739 }
740
741
xchgl(Register dst,Register src)742 void X86Assembler::xchgl(Register dst, Register src) {
743 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
744 EmitUint8(0x87);
745 EmitRegisterOperand(dst, src);
746 }
747
xchgl(Register reg,const Address & address)748 void X86Assembler::xchgl(Register reg, const Address& address) {
749 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
750 EmitUint8(0x87);
751 EmitOperand(reg, address);
752 }
753
754
cmpl(Register reg,const Immediate & imm)755 void X86Assembler::cmpl(Register reg, const Immediate& imm) {
756 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
757 EmitComplex(7, Operand(reg), imm);
758 }
759
760
cmpl(Register reg0,Register reg1)761 void X86Assembler::cmpl(Register reg0, Register reg1) {
762 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
763 EmitUint8(0x3B);
764 EmitOperand(reg0, Operand(reg1));
765 }
766
767
cmpl(Register reg,const Address & address)768 void X86Assembler::cmpl(Register reg, const Address& address) {
769 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
770 EmitUint8(0x3B);
771 EmitOperand(reg, address);
772 }
773
774
addl(Register dst,Register src)775 void X86Assembler::addl(Register dst, Register src) {
776 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
777 EmitUint8(0x03);
778 EmitRegisterOperand(dst, src);
779 }
780
781
addl(Register reg,const Address & address)782 void X86Assembler::addl(Register reg, const Address& address) {
783 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
784 EmitUint8(0x03);
785 EmitOperand(reg, address);
786 }
787
788
cmpl(const Address & address,Register reg)789 void X86Assembler::cmpl(const Address& address, Register reg) {
790 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
791 EmitUint8(0x39);
792 EmitOperand(reg, address);
793 }
794
795
cmpl(const Address & address,const Immediate & imm)796 void X86Assembler::cmpl(const Address& address, const Immediate& imm) {
797 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
798 EmitComplex(7, address, imm);
799 }
800
801
testl(Register reg1,Register reg2)802 void X86Assembler::testl(Register reg1, Register reg2) {
803 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
804 EmitUint8(0x85);
805 EmitRegisterOperand(reg1, reg2);
806 }
807
808
testl(Register reg,const Address & address)809 void X86Assembler::testl(Register reg, const Address& address) {
810 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
811 EmitUint8(0x85);
812 EmitOperand(reg, address);
813 }
814
815
testl(Register reg,const Immediate & immediate)816 void X86Assembler::testl(Register reg, const Immediate& immediate) {
817 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
818 // For registers that have a byte variant (EAX, EBX, ECX, and EDX)
819 // we only test the byte register to keep the encoding short.
820 if (immediate.is_uint8() && reg < 4) {
821 // Use zero-extended 8-bit immediate.
822 if (reg == EAX) {
823 EmitUint8(0xA8);
824 } else {
825 EmitUint8(0xF6);
826 EmitUint8(0xC0 + reg);
827 }
828 EmitUint8(immediate.value() & 0xFF);
829 } else if (reg == EAX) {
830 // Use short form if the destination is EAX.
831 EmitUint8(0xA9);
832 EmitImmediate(immediate);
833 } else {
834 EmitUint8(0xF7);
835 EmitOperand(0, Operand(reg));
836 EmitImmediate(immediate);
837 }
838 }
839
840
andl(Register dst,Register src)841 void X86Assembler::andl(Register dst, Register src) {
842 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
843 EmitUint8(0x23);
844 EmitOperand(dst, Operand(src));
845 }
846
847
andl(Register dst,const Immediate & imm)848 void X86Assembler::andl(Register dst, const Immediate& imm) {
849 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
850 EmitComplex(4, Operand(dst), imm);
851 }
852
853
orl(Register dst,Register src)854 void X86Assembler::orl(Register dst, Register src) {
855 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
856 EmitUint8(0x0B);
857 EmitOperand(dst, Operand(src));
858 }
859
860
orl(Register dst,const Immediate & imm)861 void X86Assembler::orl(Register dst, const Immediate& imm) {
862 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
863 EmitComplex(1, Operand(dst), imm);
864 }
865
866
xorl(Register dst,Register src)867 void X86Assembler::xorl(Register dst, Register src) {
868 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
869 EmitUint8(0x33);
870 EmitOperand(dst, Operand(src));
871 }
872
xorl(Register dst,const Immediate & imm)873 void X86Assembler::xorl(Register dst, const Immediate& imm) {
874 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
875 EmitComplex(6, Operand(dst), imm);
876 }
877
addl(Register reg,const Immediate & imm)878 void X86Assembler::addl(Register reg, const Immediate& imm) {
879 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
880 EmitComplex(0, Operand(reg), imm);
881 }
882
883
addl(const Address & address,Register reg)884 void X86Assembler::addl(const Address& address, Register reg) {
885 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
886 EmitUint8(0x01);
887 EmitOperand(reg, address);
888 }
889
890
addl(const Address & address,const Immediate & imm)891 void X86Assembler::addl(const Address& address, const Immediate& imm) {
892 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
893 EmitComplex(0, address, imm);
894 }
895
896
adcl(Register reg,const Immediate & imm)897 void X86Assembler::adcl(Register reg, const Immediate& imm) {
898 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
899 EmitComplex(2, Operand(reg), imm);
900 }
901
902
adcl(Register dst,Register src)903 void X86Assembler::adcl(Register dst, Register src) {
904 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
905 EmitUint8(0x13);
906 EmitOperand(dst, Operand(src));
907 }
908
909
adcl(Register dst,const Address & address)910 void X86Assembler::adcl(Register dst, const Address& address) {
911 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
912 EmitUint8(0x13);
913 EmitOperand(dst, address);
914 }
915
916
subl(Register dst,Register src)917 void X86Assembler::subl(Register dst, Register src) {
918 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
919 EmitUint8(0x2B);
920 EmitOperand(dst, Operand(src));
921 }
922
923
subl(Register reg,const Immediate & imm)924 void X86Assembler::subl(Register reg, const Immediate& imm) {
925 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
926 EmitComplex(5, Operand(reg), imm);
927 }
928
929
subl(Register reg,const Address & address)930 void X86Assembler::subl(Register reg, const Address& address) {
931 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
932 EmitUint8(0x2B);
933 EmitOperand(reg, address);
934 }
935
936
cdq()937 void X86Assembler::cdq() {
938 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
939 EmitUint8(0x99);
940 }
941
942
idivl(Register reg)943 void X86Assembler::idivl(Register reg) {
944 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
945 EmitUint8(0xF7);
946 EmitUint8(0xF8 | reg);
947 }
948
949
imull(Register dst,Register src)950 void X86Assembler::imull(Register dst, Register src) {
951 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
952 EmitUint8(0x0F);
953 EmitUint8(0xAF);
954 EmitOperand(dst, Operand(src));
955 }
956
957
imull(Register reg,const Immediate & imm)958 void X86Assembler::imull(Register reg, const Immediate& imm) {
959 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
960 EmitUint8(0x69);
961 EmitOperand(reg, Operand(reg));
962 EmitImmediate(imm);
963 }
964
965
imull(Register reg,const Address & address)966 void X86Assembler::imull(Register reg, const Address& address) {
967 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
968 EmitUint8(0x0F);
969 EmitUint8(0xAF);
970 EmitOperand(reg, address);
971 }
972
973
imull(Register reg)974 void X86Assembler::imull(Register reg) {
975 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
976 EmitUint8(0xF7);
977 EmitOperand(5, Operand(reg));
978 }
979
980
imull(const Address & address)981 void X86Assembler::imull(const Address& address) {
982 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
983 EmitUint8(0xF7);
984 EmitOperand(5, address);
985 }
986
987
mull(Register reg)988 void X86Assembler::mull(Register reg) {
989 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
990 EmitUint8(0xF7);
991 EmitOperand(4, Operand(reg));
992 }
993
994
mull(const Address & address)995 void X86Assembler::mull(const Address& address) {
996 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
997 EmitUint8(0xF7);
998 EmitOperand(4, address);
999 }
1000
1001
sbbl(Register dst,Register src)1002 void X86Assembler::sbbl(Register dst, Register src) {
1003 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1004 EmitUint8(0x1B);
1005 EmitOperand(dst, Operand(src));
1006 }
1007
1008
sbbl(Register reg,const Immediate & imm)1009 void X86Assembler::sbbl(Register reg, const Immediate& imm) {
1010 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1011 EmitComplex(3, Operand(reg), imm);
1012 }
1013
1014
sbbl(Register dst,const Address & address)1015 void X86Assembler::sbbl(Register dst, const Address& address) {
1016 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1017 EmitUint8(0x1B);
1018 EmitOperand(dst, address);
1019 }
1020
1021
incl(Register reg)1022 void X86Assembler::incl(Register reg) {
1023 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1024 EmitUint8(0x40 + reg);
1025 }
1026
1027
incl(const Address & address)1028 void X86Assembler::incl(const Address& address) {
1029 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1030 EmitUint8(0xFF);
1031 EmitOperand(0, address);
1032 }
1033
1034
decl(Register reg)1035 void X86Assembler::decl(Register reg) {
1036 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1037 EmitUint8(0x48 + reg);
1038 }
1039
1040
decl(const Address & address)1041 void X86Assembler::decl(const Address& address) {
1042 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1043 EmitUint8(0xFF);
1044 EmitOperand(1, address);
1045 }
1046
1047
shll(Register reg,const Immediate & imm)1048 void X86Assembler::shll(Register reg, const Immediate& imm) {
1049 EmitGenericShift(4, reg, imm);
1050 }
1051
1052
shll(Register operand,Register shifter)1053 void X86Assembler::shll(Register operand, Register shifter) {
1054 EmitGenericShift(4, operand, shifter);
1055 }
1056
1057
shrl(Register reg,const Immediate & imm)1058 void X86Assembler::shrl(Register reg, const Immediate& imm) {
1059 EmitGenericShift(5, reg, imm);
1060 }
1061
1062
shrl(Register operand,Register shifter)1063 void X86Assembler::shrl(Register operand, Register shifter) {
1064 EmitGenericShift(5, operand, shifter);
1065 }
1066
1067
sarl(Register reg,const Immediate & imm)1068 void X86Assembler::sarl(Register reg, const Immediate& imm) {
1069 EmitGenericShift(7, reg, imm);
1070 }
1071
1072
sarl(Register operand,Register shifter)1073 void X86Assembler::sarl(Register operand, Register shifter) {
1074 EmitGenericShift(7, operand, shifter);
1075 }
1076
1077
shld(Register dst,Register src)1078 void X86Assembler::shld(Register dst, Register src) {
1079 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1080 EmitUint8(0x0F);
1081 EmitUint8(0xA5);
1082 EmitRegisterOperand(src, dst);
1083 }
1084
1085
negl(Register reg)1086 void X86Assembler::negl(Register reg) {
1087 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1088 EmitUint8(0xF7);
1089 EmitOperand(3, Operand(reg));
1090 }
1091
1092
notl(Register reg)1093 void X86Assembler::notl(Register reg) {
1094 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1095 EmitUint8(0xF7);
1096 EmitUint8(0xD0 | reg);
1097 }
1098
1099
enter(const Immediate & imm)1100 void X86Assembler::enter(const Immediate& imm) {
1101 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1102 EmitUint8(0xC8);
1103 CHECK(imm.is_uint16());
1104 EmitUint8(imm.value() & 0xFF);
1105 EmitUint8((imm.value() >> 8) & 0xFF);
1106 EmitUint8(0x00);
1107 }
1108
1109
leave()1110 void X86Assembler::leave() {
1111 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1112 EmitUint8(0xC9);
1113 }
1114
1115
ret()1116 void X86Assembler::ret() {
1117 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1118 EmitUint8(0xC3);
1119 }
1120
1121
ret(const Immediate & imm)1122 void X86Assembler::ret(const Immediate& imm) {
1123 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1124 EmitUint8(0xC2);
1125 CHECK(imm.is_uint16());
1126 EmitUint8(imm.value() & 0xFF);
1127 EmitUint8((imm.value() >> 8) & 0xFF);
1128 }
1129
1130
1131
nop()1132 void X86Assembler::nop() {
1133 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1134 EmitUint8(0x90);
1135 }
1136
1137
int3()1138 void X86Assembler::int3() {
1139 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1140 EmitUint8(0xCC);
1141 }
1142
1143
hlt()1144 void X86Assembler::hlt() {
1145 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1146 EmitUint8(0xF4);
1147 }
1148
1149
j(Condition condition,Label * label)1150 void X86Assembler::j(Condition condition, Label* label) {
1151 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1152 if (label->IsBound()) {
1153 static const int kShortSize = 2;
1154 static const int kLongSize = 6;
1155 int offset = label->Position() - buffer_.Size();
1156 CHECK_LE(offset, 0);
1157 if (IsInt(8, offset - kShortSize)) {
1158 EmitUint8(0x70 + condition);
1159 EmitUint8((offset - kShortSize) & 0xFF);
1160 } else {
1161 EmitUint8(0x0F);
1162 EmitUint8(0x80 + condition);
1163 EmitInt32(offset - kLongSize);
1164 }
1165 } else {
1166 EmitUint8(0x0F);
1167 EmitUint8(0x80 + condition);
1168 EmitLabelLink(label);
1169 }
1170 }
1171
1172
jmp(Register reg)1173 void X86Assembler::jmp(Register reg) {
1174 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1175 EmitUint8(0xFF);
1176 EmitRegisterOperand(4, reg);
1177 }
1178
jmp(const Address & address)1179 void X86Assembler::jmp(const Address& address) {
1180 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1181 EmitUint8(0xFF);
1182 EmitOperand(4, address);
1183 }
1184
jmp(Label * label)1185 void X86Assembler::jmp(Label* label) {
1186 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1187 if (label->IsBound()) {
1188 static const int kShortSize = 2;
1189 static const int kLongSize = 5;
1190 int offset = label->Position() - buffer_.Size();
1191 CHECK_LE(offset, 0);
1192 if (IsInt(8, offset - kShortSize)) {
1193 EmitUint8(0xEB);
1194 EmitUint8((offset - kShortSize) & 0xFF);
1195 } else {
1196 EmitUint8(0xE9);
1197 EmitInt32(offset - kLongSize);
1198 }
1199 } else {
1200 EmitUint8(0xE9);
1201 EmitLabelLink(label);
1202 }
1203 }
1204
1205
lock()1206 X86Assembler* X86Assembler::lock() {
1207 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1208 EmitUint8(0xF0);
1209 return this;
1210 }
1211
1212
cmpxchgl(const Address & address,Register reg)1213 void X86Assembler::cmpxchgl(const Address& address, Register reg) {
1214 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1215 EmitUint8(0x0F);
1216 EmitUint8(0xB1);
1217 EmitOperand(reg, address);
1218 }
1219
mfence()1220 void X86Assembler::mfence() {
1221 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1222 EmitUint8(0x0F);
1223 EmitUint8(0xAE);
1224 EmitUint8(0xF0);
1225 }
1226
fs()1227 X86Assembler* X86Assembler::fs() {
1228 // TODO: fs is a prefix and not an instruction
1229 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1230 EmitUint8(0x64);
1231 return this;
1232 }
1233
gs()1234 X86Assembler* X86Assembler::gs() {
1235 // TODO: fs is a prefix and not an instruction
1236 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1237 EmitUint8(0x65);
1238 return this;
1239 }
1240
AddImmediate(Register reg,const Immediate & imm)1241 void X86Assembler::AddImmediate(Register reg, const Immediate& imm) {
1242 int value = imm.value();
1243 if (value > 0) {
1244 if (value == 1) {
1245 incl(reg);
1246 } else if (value != 0) {
1247 addl(reg, imm);
1248 }
1249 } else if (value < 0) {
1250 value = -value;
1251 if (value == 1) {
1252 decl(reg);
1253 } else if (value != 0) {
1254 subl(reg, Immediate(value));
1255 }
1256 }
1257 }
1258
1259
LoadDoubleConstant(XmmRegister dst,double value)1260 void X86Assembler::LoadDoubleConstant(XmmRegister dst, double value) {
1261 // TODO: Need to have a code constants table.
1262 int64_t constant = bit_cast<int64_t, double>(value);
1263 pushl(Immediate(High32Bits(constant)));
1264 pushl(Immediate(Low32Bits(constant)));
1265 movsd(dst, Address(ESP, 0));
1266 addl(ESP, Immediate(2 * kWordSize));
1267 }
1268
1269
FloatNegate(XmmRegister f)1270 void X86Assembler::FloatNegate(XmmRegister f) {
1271 static const struct {
1272 uint32_t a;
1273 uint32_t b;
1274 uint32_t c;
1275 uint32_t d;
1276 } float_negate_constant __attribute__((aligned(16))) =
1277 { 0x80000000, 0x00000000, 0x80000000, 0x00000000 };
1278 xorps(f, Address::Absolute(reinterpret_cast<uword>(&float_negate_constant)));
1279 }
1280
1281
DoubleNegate(XmmRegister d)1282 void X86Assembler::DoubleNegate(XmmRegister d) {
1283 static const struct {
1284 uint64_t a;
1285 uint64_t b;
1286 } double_negate_constant __attribute__((aligned(16))) =
1287 {0x8000000000000000LL, 0x8000000000000000LL};
1288 xorpd(d, Address::Absolute(reinterpret_cast<uword>(&double_negate_constant)));
1289 }
1290
1291
DoubleAbs(XmmRegister reg)1292 void X86Assembler::DoubleAbs(XmmRegister reg) {
1293 static const struct {
1294 uint64_t a;
1295 uint64_t b;
1296 } double_abs_constant __attribute__((aligned(16))) =
1297 {0x7FFFFFFFFFFFFFFFLL, 0x7FFFFFFFFFFFFFFFLL};
1298 andpd(reg, Address::Absolute(reinterpret_cast<uword>(&double_abs_constant)));
1299 }
1300
1301
Align(int alignment,int offset)1302 void X86Assembler::Align(int alignment, int offset) {
1303 CHECK(IsPowerOfTwo(alignment));
1304 // Emit nop instruction until the real position is aligned.
1305 while (((offset + buffer_.GetPosition()) & (alignment-1)) != 0) {
1306 nop();
1307 }
1308 }
1309
1310
Bind(Label * label)1311 void X86Assembler::Bind(Label* label) {
1312 int bound = buffer_.Size();
1313 CHECK(!label->IsBound()); // Labels can only be bound once.
1314 while (label->IsLinked()) {
1315 int position = label->LinkPosition();
1316 int next = buffer_.Load<int32_t>(position);
1317 buffer_.Store<int32_t>(position, bound - (position + 4));
1318 label->position_ = next;
1319 }
1320 label->BindTo(bound);
1321 }
1322
1323
EmitOperand(int reg_or_opcode,const Operand & operand)1324 void X86Assembler::EmitOperand(int reg_or_opcode, const Operand& operand) {
1325 CHECK_GE(reg_or_opcode, 0);
1326 CHECK_LT(reg_or_opcode, 8);
1327 const int length = operand.length_;
1328 CHECK_GT(length, 0);
1329 // Emit the ModRM byte updated with the given reg value.
1330 CHECK_EQ(operand.encoding_[0] & 0x38, 0);
1331 EmitUint8(operand.encoding_[0] + (reg_or_opcode << 3));
1332 // Emit the rest of the encoded operand.
1333 for (int i = 1; i < length; i++) {
1334 EmitUint8(operand.encoding_[i]);
1335 }
1336 }
1337
1338
EmitImmediate(const Immediate & imm)1339 void X86Assembler::EmitImmediate(const Immediate& imm) {
1340 EmitInt32(imm.value());
1341 }
1342
1343
EmitComplex(int reg_or_opcode,const Operand & operand,const Immediate & immediate)1344 void X86Assembler::EmitComplex(int reg_or_opcode,
1345 const Operand& operand,
1346 const Immediate& immediate) {
1347 CHECK_GE(reg_or_opcode, 0);
1348 CHECK_LT(reg_or_opcode, 8);
1349 if (immediate.is_int8()) {
1350 // Use sign-extended 8-bit immediate.
1351 EmitUint8(0x83);
1352 EmitOperand(reg_or_opcode, operand);
1353 EmitUint8(immediate.value() & 0xFF);
1354 } else if (operand.IsRegister(EAX)) {
1355 // Use short form if the destination is eax.
1356 EmitUint8(0x05 + (reg_or_opcode << 3));
1357 EmitImmediate(immediate);
1358 } else {
1359 EmitUint8(0x81);
1360 EmitOperand(reg_or_opcode, operand);
1361 EmitImmediate(immediate);
1362 }
1363 }
1364
1365
EmitLabel(Label * label,int instruction_size)1366 void X86Assembler::EmitLabel(Label* label, int instruction_size) {
1367 if (label->IsBound()) {
1368 int offset = label->Position() - buffer_.Size();
1369 CHECK_LE(offset, 0);
1370 EmitInt32(offset - instruction_size);
1371 } else {
1372 EmitLabelLink(label);
1373 }
1374 }
1375
1376
EmitLabelLink(Label * label)1377 void X86Assembler::EmitLabelLink(Label* label) {
1378 CHECK(!label->IsBound());
1379 int position = buffer_.Size();
1380 EmitInt32(label->position_);
1381 label->LinkTo(position);
1382 }
1383
1384
EmitGenericShift(int reg_or_opcode,Register reg,const Immediate & imm)1385 void X86Assembler::EmitGenericShift(int reg_or_opcode,
1386 Register reg,
1387 const Immediate& imm) {
1388 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1389 CHECK(imm.is_int8());
1390 if (imm.value() == 1) {
1391 EmitUint8(0xD1);
1392 EmitOperand(reg_or_opcode, Operand(reg));
1393 } else {
1394 EmitUint8(0xC1);
1395 EmitOperand(reg_or_opcode, Operand(reg));
1396 EmitUint8(imm.value() & 0xFF);
1397 }
1398 }
1399
1400
EmitGenericShift(int reg_or_opcode,Register operand,Register shifter)1401 void X86Assembler::EmitGenericShift(int reg_or_opcode,
1402 Register operand,
1403 Register shifter) {
1404 AssemblerBuffer::EnsureCapacity ensured(&buffer_);
1405 CHECK_EQ(shifter, ECX);
1406 EmitUint8(0xD3);
1407 EmitOperand(reg_or_opcode, Operand(operand));
1408 }
1409
1410 constexpr size_t kFramePointerSize = 4;
1411
BuildFrame(size_t frame_size,ManagedRegister method_reg,const std::vector<ManagedRegister> & spill_regs,const ManagedRegisterEntrySpills & entry_spills)1412 void X86Assembler::BuildFrame(size_t frame_size, ManagedRegister method_reg,
1413 const std::vector<ManagedRegister>& spill_regs,
1414 const ManagedRegisterEntrySpills& entry_spills) {
1415 CHECK_ALIGNED(frame_size, kStackAlignment);
1416 for (int i = spill_regs.size() - 1; i >= 0; --i) {
1417 pushl(spill_regs.at(i).AsX86().AsCpuRegister());
1418 }
1419 // return address then method on stack
1420 addl(ESP, Immediate(-frame_size + (spill_regs.size() * kFramePointerSize) +
1421 sizeof(StackReference<mirror::ArtMethod>) /*method*/ +
1422 kFramePointerSize /*return address*/));
1423 pushl(method_reg.AsX86().AsCpuRegister());
1424 for (size_t i = 0; i < entry_spills.size(); ++i) {
1425 movl(Address(ESP, frame_size + sizeof(StackReference<mirror::ArtMethod>) +
1426 (i * kFramePointerSize)),
1427 entry_spills.at(i).AsX86().AsCpuRegister());
1428 }
1429 }
1430
RemoveFrame(size_t frame_size,const std::vector<ManagedRegister> & spill_regs)1431 void X86Assembler::RemoveFrame(size_t frame_size,
1432 const std::vector<ManagedRegister>& spill_regs) {
1433 CHECK_ALIGNED(frame_size, kStackAlignment);
1434 addl(ESP, Immediate(frame_size - (spill_regs.size() * kFramePointerSize) -
1435 sizeof(StackReference<mirror::ArtMethod>)));
1436 for (size_t i = 0; i < spill_regs.size(); ++i) {
1437 popl(spill_regs.at(i).AsX86().AsCpuRegister());
1438 }
1439 ret();
1440 }
1441
IncreaseFrameSize(size_t adjust)1442 void X86Assembler::IncreaseFrameSize(size_t adjust) {
1443 CHECK_ALIGNED(adjust, kStackAlignment);
1444 addl(ESP, Immediate(-adjust));
1445 }
1446
DecreaseFrameSize(size_t adjust)1447 void X86Assembler::DecreaseFrameSize(size_t adjust) {
1448 CHECK_ALIGNED(adjust, kStackAlignment);
1449 addl(ESP, Immediate(adjust));
1450 }
1451
Store(FrameOffset offs,ManagedRegister msrc,size_t size)1452 void X86Assembler::Store(FrameOffset offs, ManagedRegister msrc, size_t size) {
1453 X86ManagedRegister src = msrc.AsX86();
1454 if (src.IsNoRegister()) {
1455 CHECK_EQ(0u, size);
1456 } else if (src.IsCpuRegister()) {
1457 CHECK_EQ(4u, size);
1458 movl(Address(ESP, offs), src.AsCpuRegister());
1459 } else if (src.IsRegisterPair()) {
1460 CHECK_EQ(8u, size);
1461 movl(Address(ESP, offs), src.AsRegisterPairLow());
1462 movl(Address(ESP, FrameOffset(offs.Int32Value()+4)),
1463 src.AsRegisterPairHigh());
1464 } else if (src.IsX87Register()) {
1465 if (size == 4) {
1466 fstps(Address(ESP, offs));
1467 } else {
1468 fstpl(Address(ESP, offs));
1469 }
1470 } else {
1471 CHECK(src.IsXmmRegister());
1472 if (size == 4) {
1473 movss(Address(ESP, offs), src.AsXmmRegister());
1474 } else {
1475 movsd(Address(ESP, offs), src.AsXmmRegister());
1476 }
1477 }
1478 }
1479
StoreRef(FrameOffset dest,ManagedRegister msrc)1480 void X86Assembler::StoreRef(FrameOffset dest, ManagedRegister msrc) {
1481 X86ManagedRegister src = msrc.AsX86();
1482 CHECK(src.IsCpuRegister());
1483 movl(Address(ESP, dest), src.AsCpuRegister());
1484 }
1485
StoreRawPtr(FrameOffset dest,ManagedRegister msrc)1486 void X86Assembler::StoreRawPtr(FrameOffset dest, ManagedRegister msrc) {
1487 X86ManagedRegister src = msrc.AsX86();
1488 CHECK(src.IsCpuRegister());
1489 movl(Address(ESP, dest), src.AsCpuRegister());
1490 }
1491
StoreImmediateToFrame(FrameOffset dest,uint32_t imm,ManagedRegister)1492 void X86Assembler::StoreImmediateToFrame(FrameOffset dest, uint32_t imm,
1493 ManagedRegister) {
1494 movl(Address(ESP, dest), Immediate(imm));
1495 }
1496
StoreImmediateToThread32(ThreadOffset<4> dest,uint32_t imm,ManagedRegister)1497 void X86Assembler::StoreImmediateToThread32(ThreadOffset<4> dest, uint32_t imm,
1498 ManagedRegister) {
1499 fs()->movl(Address::Absolute(dest), Immediate(imm));
1500 }
1501
StoreStackOffsetToThread32(ThreadOffset<4> thr_offs,FrameOffset fr_offs,ManagedRegister mscratch)1502 void X86Assembler::StoreStackOffsetToThread32(ThreadOffset<4> thr_offs,
1503 FrameOffset fr_offs,
1504 ManagedRegister mscratch) {
1505 X86ManagedRegister scratch = mscratch.AsX86();
1506 CHECK(scratch.IsCpuRegister());
1507 leal(scratch.AsCpuRegister(), Address(ESP, fr_offs));
1508 fs()->movl(Address::Absolute(thr_offs), scratch.AsCpuRegister());
1509 }
1510
StoreStackPointerToThread32(ThreadOffset<4> thr_offs)1511 void X86Assembler::StoreStackPointerToThread32(ThreadOffset<4> thr_offs) {
1512 fs()->movl(Address::Absolute(thr_offs), ESP);
1513 }
1514
StoreSpanning(FrameOffset,ManagedRegister,FrameOffset,ManagedRegister)1515 void X86Assembler::StoreSpanning(FrameOffset /*dst*/, ManagedRegister /*src*/,
1516 FrameOffset /*in_off*/, ManagedRegister /*scratch*/) {
1517 UNIMPLEMENTED(FATAL); // this case only currently exists for ARM
1518 }
1519
Load(ManagedRegister mdest,FrameOffset src,size_t size)1520 void X86Assembler::Load(ManagedRegister mdest, FrameOffset src, size_t size) {
1521 X86ManagedRegister dest = mdest.AsX86();
1522 if (dest.IsNoRegister()) {
1523 CHECK_EQ(0u, size);
1524 } else if (dest.IsCpuRegister()) {
1525 CHECK_EQ(4u, size);
1526 movl(dest.AsCpuRegister(), Address(ESP, src));
1527 } else if (dest.IsRegisterPair()) {
1528 CHECK_EQ(8u, size);
1529 movl(dest.AsRegisterPairLow(), Address(ESP, src));
1530 movl(dest.AsRegisterPairHigh(), Address(ESP, FrameOffset(src.Int32Value()+4)));
1531 } else if (dest.IsX87Register()) {
1532 if (size == 4) {
1533 flds(Address(ESP, src));
1534 } else {
1535 fldl(Address(ESP, src));
1536 }
1537 } else {
1538 CHECK(dest.IsXmmRegister());
1539 if (size == 4) {
1540 movss(dest.AsXmmRegister(), Address(ESP, src));
1541 } else {
1542 movsd(dest.AsXmmRegister(), Address(ESP, src));
1543 }
1544 }
1545 }
1546
LoadFromThread32(ManagedRegister mdest,ThreadOffset<4> src,size_t size)1547 void X86Assembler::LoadFromThread32(ManagedRegister mdest, ThreadOffset<4> src, size_t size) {
1548 X86ManagedRegister dest = mdest.AsX86();
1549 if (dest.IsNoRegister()) {
1550 CHECK_EQ(0u, size);
1551 } else if (dest.IsCpuRegister()) {
1552 CHECK_EQ(4u, size);
1553 fs()->movl(dest.AsCpuRegister(), Address::Absolute(src));
1554 } else if (dest.IsRegisterPair()) {
1555 CHECK_EQ(8u, size);
1556 fs()->movl(dest.AsRegisterPairLow(), Address::Absolute(src));
1557 fs()->movl(dest.AsRegisterPairHigh(), Address::Absolute(ThreadOffset<4>(src.Int32Value()+4)));
1558 } else if (dest.IsX87Register()) {
1559 if (size == 4) {
1560 fs()->flds(Address::Absolute(src));
1561 } else {
1562 fs()->fldl(Address::Absolute(src));
1563 }
1564 } else {
1565 CHECK(dest.IsXmmRegister());
1566 if (size == 4) {
1567 fs()->movss(dest.AsXmmRegister(), Address::Absolute(src));
1568 } else {
1569 fs()->movsd(dest.AsXmmRegister(), Address::Absolute(src));
1570 }
1571 }
1572 }
1573
LoadRef(ManagedRegister mdest,FrameOffset src)1574 void X86Assembler::LoadRef(ManagedRegister mdest, FrameOffset src) {
1575 X86ManagedRegister dest = mdest.AsX86();
1576 CHECK(dest.IsCpuRegister());
1577 movl(dest.AsCpuRegister(), Address(ESP, src));
1578 }
1579
LoadRef(ManagedRegister mdest,ManagedRegister base,MemberOffset offs)1580 void X86Assembler::LoadRef(ManagedRegister mdest, ManagedRegister base,
1581 MemberOffset offs) {
1582 X86ManagedRegister dest = mdest.AsX86();
1583 CHECK(dest.IsCpuRegister() && dest.IsCpuRegister());
1584 movl(dest.AsCpuRegister(), Address(base.AsX86().AsCpuRegister(), offs));
1585 if (kPoisonHeapReferences) {
1586 negl(dest.AsCpuRegister());
1587 }
1588 }
1589
LoadRawPtr(ManagedRegister mdest,ManagedRegister base,Offset offs)1590 void X86Assembler::LoadRawPtr(ManagedRegister mdest, ManagedRegister base,
1591 Offset offs) {
1592 X86ManagedRegister dest = mdest.AsX86();
1593 CHECK(dest.IsCpuRegister() && dest.IsCpuRegister());
1594 movl(dest.AsCpuRegister(), Address(base.AsX86().AsCpuRegister(), offs));
1595 }
1596
LoadRawPtrFromThread32(ManagedRegister mdest,ThreadOffset<4> offs)1597 void X86Assembler::LoadRawPtrFromThread32(ManagedRegister mdest,
1598 ThreadOffset<4> offs) {
1599 X86ManagedRegister dest = mdest.AsX86();
1600 CHECK(dest.IsCpuRegister());
1601 fs()->movl(dest.AsCpuRegister(), Address::Absolute(offs));
1602 }
1603
SignExtend(ManagedRegister mreg,size_t size)1604 void X86Assembler::SignExtend(ManagedRegister mreg, size_t size) {
1605 X86ManagedRegister reg = mreg.AsX86();
1606 CHECK(size == 1 || size == 2) << size;
1607 CHECK(reg.IsCpuRegister()) << reg;
1608 if (size == 1) {
1609 movsxb(reg.AsCpuRegister(), reg.AsByteRegister());
1610 } else {
1611 movsxw(reg.AsCpuRegister(), reg.AsCpuRegister());
1612 }
1613 }
1614
ZeroExtend(ManagedRegister mreg,size_t size)1615 void X86Assembler::ZeroExtend(ManagedRegister mreg, size_t size) {
1616 X86ManagedRegister reg = mreg.AsX86();
1617 CHECK(size == 1 || size == 2) << size;
1618 CHECK(reg.IsCpuRegister()) << reg;
1619 if (size == 1) {
1620 movzxb(reg.AsCpuRegister(), reg.AsByteRegister());
1621 } else {
1622 movzxw(reg.AsCpuRegister(), reg.AsCpuRegister());
1623 }
1624 }
1625
Move(ManagedRegister mdest,ManagedRegister msrc,size_t size)1626 void X86Assembler::Move(ManagedRegister mdest, ManagedRegister msrc, size_t size) {
1627 X86ManagedRegister dest = mdest.AsX86();
1628 X86ManagedRegister src = msrc.AsX86();
1629 if (!dest.Equals(src)) {
1630 if (dest.IsCpuRegister() && src.IsCpuRegister()) {
1631 movl(dest.AsCpuRegister(), src.AsCpuRegister());
1632 } else if (src.IsX87Register() && dest.IsXmmRegister()) {
1633 // Pass via stack and pop X87 register
1634 subl(ESP, Immediate(16));
1635 if (size == 4) {
1636 CHECK_EQ(src.AsX87Register(), ST0);
1637 fstps(Address(ESP, 0));
1638 movss(dest.AsXmmRegister(), Address(ESP, 0));
1639 } else {
1640 CHECK_EQ(src.AsX87Register(), ST0);
1641 fstpl(Address(ESP, 0));
1642 movsd(dest.AsXmmRegister(), Address(ESP, 0));
1643 }
1644 addl(ESP, Immediate(16));
1645 } else {
1646 // TODO: x87, SSE
1647 UNIMPLEMENTED(FATAL) << ": Move " << dest << ", " << src;
1648 }
1649 }
1650 }
1651
CopyRef(FrameOffset dest,FrameOffset src,ManagedRegister mscratch)1652 void X86Assembler::CopyRef(FrameOffset dest, FrameOffset src,
1653 ManagedRegister mscratch) {
1654 X86ManagedRegister scratch = mscratch.AsX86();
1655 CHECK(scratch.IsCpuRegister());
1656 movl(scratch.AsCpuRegister(), Address(ESP, src));
1657 movl(Address(ESP, dest), scratch.AsCpuRegister());
1658 }
1659
CopyRawPtrFromThread32(FrameOffset fr_offs,ThreadOffset<4> thr_offs,ManagedRegister mscratch)1660 void X86Assembler::CopyRawPtrFromThread32(FrameOffset fr_offs,
1661 ThreadOffset<4> thr_offs,
1662 ManagedRegister mscratch) {
1663 X86ManagedRegister scratch = mscratch.AsX86();
1664 CHECK(scratch.IsCpuRegister());
1665 fs()->movl(scratch.AsCpuRegister(), Address::Absolute(thr_offs));
1666 Store(fr_offs, scratch, 4);
1667 }
1668
CopyRawPtrToThread32(ThreadOffset<4> thr_offs,FrameOffset fr_offs,ManagedRegister mscratch)1669 void X86Assembler::CopyRawPtrToThread32(ThreadOffset<4> thr_offs,
1670 FrameOffset fr_offs,
1671 ManagedRegister mscratch) {
1672 X86ManagedRegister scratch = mscratch.AsX86();
1673 CHECK(scratch.IsCpuRegister());
1674 Load(scratch, fr_offs, 4);
1675 fs()->movl(Address::Absolute(thr_offs), scratch.AsCpuRegister());
1676 }
1677
Copy(FrameOffset dest,FrameOffset src,ManagedRegister mscratch,size_t size)1678 void X86Assembler::Copy(FrameOffset dest, FrameOffset src,
1679 ManagedRegister mscratch,
1680 size_t size) {
1681 X86ManagedRegister scratch = mscratch.AsX86();
1682 if (scratch.IsCpuRegister() && size == 8) {
1683 Load(scratch, src, 4);
1684 Store(dest, scratch, 4);
1685 Load(scratch, FrameOffset(src.Int32Value() + 4), 4);
1686 Store(FrameOffset(dest.Int32Value() + 4), scratch, 4);
1687 } else {
1688 Load(scratch, src, size);
1689 Store(dest, scratch, size);
1690 }
1691 }
1692
Copy(FrameOffset,ManagedRegister,Offset,ManagedRegister,size_t)1693 void X86Assembler::Copy(FrameOffset /*dst*/, ManagedRegister /*src_base*/, Offset /*src_offset*/,
1694 ManagedRegister /*scratch*/, size_t /*size*/) {
1695 UNIMPLEMENTED(FATAL);
1696 }
1697
Copy(ManagedRegister dest_base,Offset dest_offset,FrameOffset src,ManagedRegister scratch,size_t size)1698 void X86Assembler::Copy(ManagedRegister dest_base, Offset dest_offset, FrameOffset src,
1699 ManagedRegister scratch, size_t size) {
1700 CHECK(scratch.IsNoRegister());
1701 CHECK_EQ(size, 4u);
1702 pushl(Address(ESP, src));
1703 popl(Address(dest_base.AsX86().AsCpuRegister(), dest_offset));
1704 }
1705
Copy(FrameOffset dest,FrameOffset src_base,Offset src_offset,ManagedRegister mscratch,size_t size)1706 void X86Assembler::Copy(FrameOffset dest, FrameOffset src_base, Offset src_offset,
1707 ManagedRegister mscratch, size_t size) {
1708 Register scratch = mscratch.AsX86().AsCpuRegister();
1709 CHECK_EQ(size, 4u);
1710 movl(scratch, Address(ESP, src_base));
1711 movl(scratch, Address(scratch, src_offset));
1712 movl(Address(ESP, dest), scratch);
1713 }
1714
Copy(ManagedRegister dest,Offset dest_offset,ManagedRegister src,Offset src_offset,ManagedRegister scratch,size_t size)1715 void X86Assembler::Copy(ManagedRegister dest, Offset dest_offset,
1716 ManagedRegister src, Offset src_offset,
1717 ManagedRegister scratch, size_t size) {
1718 CHECK_EQ(size, 4u);
1719 CHECK(scratch.IsNoRegister());
1720 pushl(Address(src.AsX86().AsCpuRegister(), src_offset));
1721 popl(Address(dest.AsX86().AsCpuRegister(), dest_offset));
1722 }
1723
Copy(FrameOffset dest,Offset dest_offset,FrameOffset src,Offset src_offset,ManagedRegister mscratch,size_t size)1724 void X86Assembler::Copy(FrameOffset dest, Offset dest_offset, FrameOffset src, Offset src_offset,
1725 ManagedRegister mscratch, size_t size) {
1726 Register scratch = mscratch.AsX86().AsCpuRegister();
1727 CHECK_EQ(size, 4u);
1728 CHECK_EQ(dest.Int32Value(), src.Int32Value());
1729 movl(scratch, Address(ESP, src));
1730 pushl(Address(scratch, src_offset));
1731 popl(Address(scratch, dest_offset));
1732 }
1733
MemoryBarrier(ManagedRegister)1734 void X86Assembler::MemoryBarrier(ManagedRegister) {
1735 #if ANDROID_SMP != 0
1736 mfence();
1737 #endif
1738 }
1739
CreateHandleScopeEntry(ManagedRegister mout_reg,FrameOffset handle_scope_offset,ManagedRegister min_reg,bool null_allowed)1740 void X86Assembler::CreateHandleScopeEntry(ManagedRegister mout_reg,
1741 FrameOffset handle_scope_offset,
1742 ManagedRegister min_reg, bool null_allowed) {
1743 X86ManagedRegister out_reg = mout_reg.AsX86();
1744 X86ManagedRegister in_reg = min_reg.AsX86();
1745 CHECK(in_reg.IsCpuRegister());
1746 CHECK(out_reg.IsCpuRegister());
1747 VerifyObject(in_reg, null_allowed);
1748 if (null_allowed) {
1749 Label null_arg;
1750 if (!out_reg.Equals(in_reg)) {
1751 xorl(out_reg.AsCpuRegister(), out_reg.AsCpuRegister());
1752 }
1753 testl(in_reg.AsCpuRegister(), in_reg.AsCpuRegister());
1754 j(kZero, &null_arg);
1755 leal(out_reg.AsCpuRegister(), Address(ESP, handle_scope_offset));
1756 Bind(&null_arg);
1757 } else {
1758 leal(out_reg.AsCpuRegister(), Address(ESP, handle_scope_offset));
1759 }
1760 }
1761
CreateHandleScopeEntry(FrameOffset out_off,FrameOffset handle_scope_offset,ManagedRegister mscratch,bool null_allowed)1762 void X86Assembler::CreateHandleScopeEntry(FrameOffset out_off,
1763 FrameOffset handle_scope_offset,
1764 ManagedRegister mscratch,
1765 bool null_allowed) {
1766 X86ManagedRegister scratch = mscratch.AsX86();
1767 CHECK(scratch.IsCpuRegister());
1768 if (null_allowed) {
1769 Label null_arg;
1770 movl(scratch.AsCpuRegister(), Address(ESP, handle_scope_offset));
1771 testl(scratch.AsCpuRegister(), scratch.AsCpuRegister());
1772 j(kZero, &null_arg);
1773 leal(scratch.AsCpuRegister(), Address(ESP, handle_scope_offset));
1774 Bind(&null_arg);
1775 } else {
1776 leal(scratch.AsCpuRegister(), Address(ESP, handle_scope_offset));
1777 }
1778 Store(out_off, scratch, 4);
1779 }
1780
1781 // Given a handle scope entry, load the associated reference.
LoadReferenceFromHandleScope(ManagedRegister mout_reg,ManagedRegister min_reg)1782 void X86Assembler::LoadReferenceFromHandleScope(ManagedRegister mout_reg,
1783 ManagedRegister min_reg) {
1784 X86ManagedRegister out_reg = mout_reg.AsX86();
1785 X86ManagedRegister in_reg = min_reg.AsX86();
1786 CHECK(out_reg.IsCpuRegister());
1787 CHECK(in_reg.IsCpuRegister());
1788 Label null_arg;
1789 if (!out_reg.Equals(in_reg)) {
1790 xorl(out_reg.AsCpuRegister(), out_reg.AsCpuRegister());
1791 }
1792 testl(in_reg.AsCpuRegister(), in_reg.AsCpuRegister());
1793 j(kZero, &null_arg);
1794 movl(out_reg.AsCpuRegister(), Address(in_reg.AsCpuRegister(), 0));
1795 Bind(&null_arg);
1796 }
1797
VerifyObject(ManagedRegister,bool)1798 void X86Assembler::VerifyObject(ManagedRegister /*src*/, bool /*could_be_null*/) {
1799 // TODO: not validating references
1800 }
1801
VerifyObject(FrameOffset,bool)1802 void X86Assembler::VerifyObject(FrameOffset /*src*/, bool /*could_be_null*/) {
1803 // TODO: not validating references
1804 }
1805
Call(ManagedRegister mbase,Offset offset,ManagedRegister)1806 void X86Assembler::Call(ManagedRegister mbase, Offset offset, ManagedRegister) {
1807 X86ManagedRegister base = mbase.AsX86();
1808 CHECK(base.IsCpuRegister());
1809 call(Address(base.AsCpuRegister(), offset.Int32Value()));
1810 // TODO: place reference map on call
1811 }
1812
Call(FrameOffset base,Offset offset,ManagedRegister mscratch)1813 void X86Assembler::Call(FrameOffset base, Offset offset, ManagedRegister mscratch) {
1814 Register scratch = mscratch.AsX86().AsCpuRegister();
1815 movl(scratch, Address(ESP, base));
1816 call(Address(scratch, offset));
1817 }
1818
CallFromThread32(ThreadOffset<4> offset,ManagedRegister)1819 void X86Assembler::CallFromThread32(ThreadOffset<4> offset, ManagedRegister /*mscratch*/) {
1820 fs()->call(Address::Absolute(offset));
1821 }
1822
GetCurrentThread(ManagedRegister tr)1823 void X86Assembler::GetCurrentThread(ManagedRegister tr) {
1824 fs()->movl(tr.AsX86().AsCpuRegister(),
1825 Address::Absolute(Thread::SelfOffset<4>()));
1826 }
1827
GetCurrentThread(FrameOffset offset,ManagedRegister mscratch)1828 void X86Assembler::GetCurrentThread(FrameOffset offset,
1829 ManagedRegister mscratch) {
1830 X86ManagedRegister scratch = mscratch.AsX86();
1831 fs()->movl(scratch.AsCpuRegister(), Address::Absolute(Thread::SelfOffset<4>()));
1832 movl(Address(ESP, offset), scratch.AsCpuRegister());
1833 }
1834
ExceptionPoll(ManagedRegister,size_t stack_adjust)1835 void X86Assembler::ExceptionPoll(ManagedRegister /*scratch*/, size_t stack_adjust) {
1836 X86ExceptionSlowPath* slow = new X86ExceptionSlowPath(stack_adjust);
1837 buffer_.EnqueueSlowPath(slow);
1838 fs()->cmpl(Address::Absolute(Thread::ExceptionOffset<4>()), Immediate(0));
1839 j(kNotEqual, slow->Entry());
1840 }
1841
Emit(Assembler * sasm)1842 void X86ExceptionSlowPath::Emit(Assembler *sasm) {
1843 X86Assembler* sp_asm = down_cast<X86Assembler*>(sasm);
1844 #define __ sp_asm->
1845 __ Bind(&entry_);
1846 // Note: the return value is dead
1847 if (stack_adjust_ != 0) { // Fix up the frame.
1848 __ DecreaseFrameSize(stack_adjust_);
1849 }
1850 // Pass exception as argument in EAX
1851 __ fs()->movl(EAX, Address::Absolute(Thread::ExceptionOffset<4>()));
1852 __ fs()->call(Address::Absolute(QUICK_ENTRYPOINT_OFFSET(4, pDeliverException)));
1853 // this call should never return
1854 __ int3();
1855 #undef __
1856 }
1857
1858 } // namespace x86
1859 } // namespace art
1860