mirror of
https://github.com/dart-lang/sdk
synced 2024-11-05 18:22:09 +00:00
8378b8fdbf
Inline instance object hash code into object header on 64 bit. 64 bit objects have 32 bits of free space in the header word. This is used for the hash code in string objects. We take it for the default hash code on all objects that don't override the hashCode getter. This is both faster and a memory reduction. Eg it makes the MegaHashCode part of the Megamorphic benchmark 6 times faster. This is a reland of https://codereview.chromium.org/2912863006/ It fixes issues with the 32 bit compare-swap instruction on ARM64 and fixes a fragile tree shaking test that is sensitive to which private methods are in the core libraries. R=kustermann@google.com, vegorov@google.com BUG= Review-Url: https://codereview.chromium.org/2954453002 .
3031 lines
76 KiB
C++
3031 lines
76 KiB
C++
// Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file
|
|
// for details. All rights reserved. Use of this source code is governed by a
|
|
// BSD-style license that can be found in the LICENSE file.
|
|
|
|
#include "vm/globals.h" // NOLINT
|
|
#if defined(TARGET_ARCH_IA32)
|
|
|
|
#include "vm/assembler.h"
|
|
#include "vm/cpu.h"
|
|
#include "vm/heap.h"
|
|
#include "vm/instructions.h"
|
|
#include "vm/memory_region.h"
|
|
#include "vm/runtime_entry.h"
|
|
#include "vm/stack_frame.h"
|
|
#include "vm/stub_code.h"
|
|
|
|
namespace dart {
|
|
|
|
DECLARE_FLAG(bool, inline_alloc);
|
|
|
|
|
|
class DirectCallRelocation : public AssemblerFixup {
|
|
public:
|
|
void Process(const MemoryRegion& region, intptr_t position) {
|
|
// Direct calls are relative to the following instruction on x86.
|
|
int32_t pointer = region.Load<int32_t>(position);
|
|
int32_t delta = region.start() + position + sizeof(int32_t);
|
|
region.Store<int32_t>(position, pointer - delta);
|
|
}
|
|
|
|
virtual bool IsPointerOffset() const { return false; }
|
|
};
|
|
|
|
|
|
int32_t Assembler::jit_cookie() {
|
|
if (jit_cookie_ == 0) {
|
|
jit_cookie_ =
|
|
static_cast<int32_t>(Isolate::Current()->random()->NextUInt32());
|
|
}
|
|
return jit_cookie_;
|
|
}
|
|
|
|
|
|
void Assembler::InitializeMemoryWithBreakpoints(uword data, intptr_t length) {
|
|
memset(reinterpret_cast<void*>(data), Instr::kBreakPointInstruction, length);
|
|
}
|
|
|
|
|
|
void Assembler::call(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xFF);
|
|
EmitRegisterOperand(2, reg);
|
|
}
|
|
|
|
|
|
void Assembler::call(const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xFF);
|
|
EmitOperand(2, address);
|
|
}
|
|
|
|
|
|
void Assembler::call(Label* label) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xE8);
|
|
static const int kSize = 5;
|
|
EmitLabel(label, kSize);
|
|
}
|
|
|
|
|
|
void Assembler::call(const ExternalLabel* label) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
intptr_t call_start = buffer_.GetPosition();
|
|
EmitUint8(0xE8);
|
|
EmitFixup(new DirectCallRelocation());
|
|
EmitInt32(label->address());
|
|
ASSERT((buffer_.GetPosition() - call_start) == kCallExternalLabelSize);
|
|
}
|
|
|
|
|
|
void Assembler::pushl(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x50 + reg);
|
|
}
|
|
|
|
|
|
void Assembler::pushl(const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xFF);
|
|
EmitOperand(6, address);
|
|
}
|
|
|
|
|
|
void Assembler::pushl(const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
if (imm.is_int8()) {
|
|
EmitUint8(0x6A);
|
|
EmitUint8(imm.value() & 0xFF);
|
|
} else {
|
|
EmitUint8(0x68);
|
|
EmitImmediate(imm);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::popl(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x58 + reg);
|
|
}
|
|
|
|
|
|
void Assembler::popl(const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x8F);
|
|
EmitOperand(0, address);
|
|
}
|
|
|
|
|
|
void Assembler::pushal() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x60);
|
|
}
|
|
|
|
|
|
void Assembler::popal() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x61);
|
|
}
|
|
|
|
|
|
void Assembler::setcc(Condition condition, ByteRegister dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x90 + condition);
|
|
EmitUint8(0xC0 + dst);
|
|
}
|
|
|
|
|
|
void Assembler::movl(Register dst, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xB8 + dst);
|
|
EmitImmediate(imm);
|
|
}
|
|
|
|
|
|
void Assembler::movl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x89);
|
|
EmitRegisterOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movl(Register dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x8B);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movl(const Address& dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x89);
|
|
EmitOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movl(const Address& dst, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xC7);
|
|
EmitOperand(0, dst);
|
|
EmitImmediate(imm);
|
|
}
|
|
|
|
|
|
void Assembler::movzxb(Register dst, ByteRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xB6);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movzxb(Register dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xB6);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movsxb(Register dst, ByteRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xBE);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movsxb(Register dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xBE);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movb(Register dst, const Address& src) {
|
|
FATAL("Use movzxb or movsxb instead.");
|
|
}
|
|
|
|
|
|
void Assembler::movb(const Address& dst, ByteRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x88);
|
|
EmitOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movb(const Address& dst, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xC6);
|
|
EmitOperand(EAX, dst);
|
|
ASSERT(imm.is_int8());
|
|
EmitUint8(imm.value() & 0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::movzxw(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xB7);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movzxw(Register dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xB7);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movsxw(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xBF);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movsxw(Register dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xBF);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movw(Register dst, const Address& src) {
|
|
FATAL("Use movzxw or movsxw instead.");
|
|
}
|
|
|
|
|
|
void Assembler::movw(const Address& dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitOperandSizeOverride();
|
|
EmitUint8(0x89);
|
|
EmitOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movw(const Address& dst, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitOperandSizeOverride();
|
|
EmitUint8(0xC7);
|
|
EmitOperand(0, dst);
|
|
EmitUint8(imm.value() & 0xFF);
|
|
EmitUint8((imm.value() >> 8) & 0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::leal(Register dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x8D);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
// Move if not overflow.
|
|
void Assembler::cmovno(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x41);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cmove(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x44);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cmovne(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x45);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cmovs(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x48);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cmovns(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x49);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cmovgel(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x4D);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cmovlessl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x4C);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::rep_movsb() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0xA4);
|
|
}
|
|
|
|
|
|
void Assembler::movss(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x10);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movss(const Address& dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x11);
|
|
EmitOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movss(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x11);
|
|
EmitXmmRegisterOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movd(XmmRegister dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x6E);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::movd(Register dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x7E);
|
|
EmitOperand(src, Operand(dst));
|
|
}
|
|
|
|
|
|
void Assembler::movq(const Address& dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xD6);
|
|
EmitOperand(src, Operand(dst));
|
|
}
|
|
|
|
|
|
void Assembler::movq(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x7E);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::addss(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x58);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::addss(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x58);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::subss(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5C);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::subss(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5C);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::mulss(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x59);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::mulss(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x59);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::divss(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5E);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::divss(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5E);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::flds(const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitOperand(0, src);
|
|
}
|
|
|
|
|
|
void Assembler::fstps(const Address& dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitOperand(3, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movsd(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x10);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movsd(const Address& dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x11);
|
|
EmitOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movsd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x11);
|
|
EmitXmmRegisterOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movaps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x28);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movups(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x10);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movups(const Address& dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x11);
|
|
EmitOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::addsd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x58);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::addsd(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x58);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::addpl(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xFE);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::subpl(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xFA);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::addps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x58);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::subps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5C);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::divps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5E);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::mulps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x59);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::minps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5D);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::maxps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5F);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::andps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x54);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::andps(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x54);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::orps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x56);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::notps(XmmRegister dst) {
|
|
static const struct ALIGN16 {
|
|
uint32_t a;
|
|
uint32_t b;
|
|
uint32_t c;
|
|
uint32_t d;
|
|
} float_not_constant = {0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF};
|
|
xorps(dst, Address::Absolute(reinterpret_cast<uword>(&float_not_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::negateps(XmmRegister dst) {
|
|
static const struct ALIGN16 {
|
|
uint32_t a;
|
|
uint32_t b;
|
|
uint32_t c;
|
|
uint32_t d;
|
|
} float_negate_constant = {0x80000000, 0x80000000, 0x80000000, 0x80000000};
|
|
xorps(dst,
|
|
Address::Absolute(reinterpret_cast<uword>(&float_negate_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::absps(XmmRegister dst) {
|
|
static const struct ALIGN16 {
|
|
uint32_t a;
|
|
uint32_t b;
|
|
uint32_t c;
|
|
uint32_t d;
|
|
} float_absolute_constant = {0x7FFFFFFF, 0x7FFFFFFF, 0x7FFFFFFF, 0x7FFFFFFF};
|
|
andps(dst,
|
|
Address::Absolute(reinterpret_cast<uword>(&float_absolute_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::zerowps(XmmRegister dst) {
|
|
static const struct ALIGN16 {
|
|
uint32_t a;
|
|
uint32_t b;
|
|
uint32_t c;
|
|
uint32_t d;
|
|
} float_zerow_constant = {0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF, 0x00000000};
|
|
andps(dst, Address::Absolute(reinterpret_cast<uword>(&float_zerow_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::cmppseq(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xC2);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
EmitUint8(0x0);
|
|
}
|
|
|
|
|
|
void Assembler::cmppsneq(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xC2);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
EmitUint8(0x4);
|
|
}
|
|
|
|
|
|
void Assembler::cmppslt(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xC2);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
EmitUint8(0x1);
|
|
}
|
|
|
|
|
|
void Assembler::cmppsle(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xC2);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
EmitUint8(0x2);
|
|
}
|
|
|
|
|
|
void Assembler::cmppsnlt(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xC2);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
EmitUint8(0x5);
|
|
}
|
|
|
|
|
|
void Assembler::cmppsnle(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xC2);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
EmitUint8(0x6);
|
|
}
|
|
|
|
|
|
void Assembler::sqrtps(XmmRegister dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x51);
|
|
EmitXmmRegisterOperand(dst, dst);
|
|
}
|
|
|
|
|
|
void Assembler::rsqrtps(XmmRegister dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x52);
|
|
EmitXmmRegisterOperand(dst, dst);
|
|
}
|
|
|
|
|
|
void Assembler::reciprocalps(XmmRegister dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x53);
|
|
EmitXmmRegisterOperand(dst, dst);
|
|
}
|
|
|
|
|
|
void Assembler::movhlps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x12);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movlhps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x16);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::unpcklps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x14);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::unpckhps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x15);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::unpcklpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x14);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::unpckhpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x15);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::set1ps(XmmRegister dst, Register tmp1, const Immediate& imm) {
|
|
// Load 32-bit immediate value into tmp1.
|
|
movl(tmp1, imm);
|
|
// Move value from tmp1 into dst.
|
|
movd(dst, tmp1);
|
|
// Broadcast low lane into other three lanes.
|
|
shufps(dst, dst, Immediate(0x0));
|
|
}
|
|
|
|
|
|
void Assembler::shufps(XmmRegister dst, XmmRegister src, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xC6);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
ASSERT(imm.is_uint8());
|
|
EmitUint8(imm.value());
|
|
}
|
|
|
|
|
|
void Assembler::addpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x58);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::negatepd(XmmRegister dst) {
|
|
static const struct ALIGN16 {
|
|
uint64_t a;
|
|
uint64_t b;
|
|
} double_negate_constant = {0x8000000000000000LL, 0x8000000000000000LL};
|
|
xorpd(dst,
|
|
Address::Absolute(reinterpret_cast<uword>(&double_negate_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::subpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5C);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::mulpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x59);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::divpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5E);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::abspd(XmmRegister dst) {
|
|
static const struct ALIGN16 {
|
|
uint64_t a;
|
|
uint64_t b;
|
|
} double_absolute_constant = {0x7FFFFFFFFFFFFFFFLL, 0x7FFFFFFFFFFFFFFFLL};
|
|
andpd(dst,
|
|
Address::Absolute(reinterpret_cast<uword>(&double_absolute_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::minpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5D);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::maxpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5F);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::sqrtpd(XmmRegister dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x51);
|
|
EmitXmmRegisterOperand(dst, dst);
|
|
}
|
|
|
|
|
|
void Assembler::cvtps2pd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5A);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cvtpd2ps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5A);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::shufpd(XmmRegister dst, XmmRegister src, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xC6);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
ASSERT(imm.is_uint8());
|
|
EmitUint8(imm.value());
|
|
}
|
|
|
|
|
|
void Assembler::subsd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5C);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::subsd(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5C);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::mulsd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x59);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::mulsd(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x59);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::divsd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5E);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::divsd(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5E);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cvtsi2ss(XmmRegister dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x2A);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::cvtsi2sd(XmmRegister dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x2A);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::cvtss2si(Register dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x2D);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cvtss2sd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5A);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cvtsd2si(Register dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x2D);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cvttss2si(Register dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x2C);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cvttsd2si(Register dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x2C);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cvtsd2ss(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x5A);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cvtdq2pd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xE6);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::comiss(XmmRegister a, XmmRegister b) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x2F);
|
|
EmitXmmRegisterOperand(a, b);
|
|
}
|
|
|
|
|
|
void Assembler::comisd(XmmRegister a, XmmRegister b) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x2F);
|
|
EmitXmmRegisterOperand(a, b);
|
|
}
|
|
|
|
|
|
void Assembler::movmskpd(Register dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x50);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::movmskps(Register dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x50);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::sqrtsd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF2);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x51);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::sqrtss(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF3);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x51);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::xorpd(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x57);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::xorpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x57);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::orpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x56);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::xorps(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x57);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::xorps(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x57);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::andpd(XmmRegister dst, const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x54);
|
|
EmitOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::andpd(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x54);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::pextrd(Register dst, XmmRegister src, const Immediate& imm) {
|
|
ASSERT(TargetCPUFeatures::sse4_1_supported());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x3A);
|
|
EmitUint8(0x16);
|
|
EmitOperand(src, Operand(dst));
|
|
ASSERT(imm.is_uint8());
|
|
EmitUint8(imm.value());
|
|
}
|
|
|
|
|
|
void Assembler::pmovsxdq(XmmRegister dst, XmmRegister src) {
|
|
ASSERT(TargetCPUFeatures::sse4_1_supported());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x38);
|
|
EmitUint8(0x25);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::pcmpeqq(XmmRegister dst, XmmRegister src) {
|
|
ASSERT(TargetCPUFeatures::sse4_1_supported());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x38);
|
|
EmitUint8(0x29);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::pxor(XmmRegister dst, XmmRegister src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xEF);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::roundsd(XmmRegister dst, XmmRegister src, RoundingMode mode) {
|
|
ASSERT(TargetCPUFeatures::sse4_1_supported());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x3A);
|
|
EmitUint8(0x0B);
|
|
EmitXmmRegisterOperand(dst, src);
|
|
// Mask precision exeption.
|
|
EmitUint8(static_cast<uint8_t>(mode) | 0x8);
|
|
}
|
|
|
|
|
|
void Assembler::fldl(const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xDD);
|
|
EmitOperand(0, src);
|
|
}
|
|
|
|
|
|
void Assembler::fstpl(const Address& dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xDD);
|
|
EmitOperand(3, dst);
|
|
}
|
|
|
|
|
|
void Assembler::fnstcw(const Address& dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitOperand(7, dst);
|
|
}
|
|
|
|
|
|
void Assembler::fldcw(const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitOperand(5, src);
|
|
}
|
|
|
|
|
|
void Assembler::fistpl(const Address& dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xDF);
|
|
EmitOperand(7, dst);
|
|
}
|
|
|
|
|
|
void Assembler::fistps(const Address& dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xDB);
|
|
EmitOperand(3, dst);
|
|
}
|
|
|
|
|
|
void Assembler::fildl(const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xDF);
|
|
EmitOperand(5, src);
|
|
}
|
|
|
|
|
|
void Assembler::filds(const Address& src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xDB);
|
|
EmitOperand(0, src);
|
|
}
|
|
|
|
|
|
void Assembler::fincstp() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitUint8(0xF7);
|
|
}
|
|
|
|
|
|
void Assembler::ffree(intptr_t value) {
|
|
ASSERT(value < 7);
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xDD);
|
|
EmitUint8(0xC0 + value);
|
|
}
|
|
|
|
|
|
void Assembler::fsin() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitUint8(0xFE);
|
|
}
|
|
|
|
|
|
void Assembler::fcos() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitUint8(0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::fsincos() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitUint8(0xFB);
|
|
}
|
|
|
|
|
|
void Assembler::fptan() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xD9);
|
|
EmitUint8(0xF2);
|
|
}
|
|
|
|
|
|
void Assembler::xchgl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x87);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::cmpl(Register reg, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(7, Operand(reg), imm);
|
|
}
|
|
|
|
|
|
void Assembler::cmpl(Register reg0, Register reg1) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x3B);
|
|
EmitOperand(reg0, Operand(reg1));
|
|
}
|
|
|
|
|
|
void Assembler::cmpl(Register reg, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x3B);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::addl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x03);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::addl(Register reg, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x03);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::cmpl(const Address& address, Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x39);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::cmpl(const Address& address, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(7, address, imm);
|
|
}
|
|
|
|
|
|
void Assembler::cmpw(Register reg, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitOperandSizeOverride();
|
|
EmitUint8(0x3B);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::cmpw(const Address& address, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitOperandSizeOverride();
|
|
EmitUint8(0x81);
|
|
EmitOperand(7, address);
|
|
EmitUint8(imm.value() & 0xFF);
|
|
EmitUint8((imm.value() >> 8) & 0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::cmpb(const Address& address, const Immediate& imm) {
|
|
ASSERT(imm.is_int8());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x80);
|
|
EmitOperand(7, address);
|
|
EmitUint8(imm.value() & 0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::testl(Register reg1, Register reg2) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x85);
|
|
EmitRegisterOperand(reg1, reg2);
|
|
}
|
|
|
|
|
|
void Assembler::testl(Register reg, const Immediate& immediate) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
// For registers that have a byte variant (EAX, EBX, ECX, and EDX)
|
|
// we only test the byte register to keep the encoding short.
|
|
if (immediate.is_uint8() && reg < 4) {
|
|
// Use zero-extended 8-bit immediate.
|
|
if (reg == EAX) {
|
|
EmitUint8(0xA8);
|
|
} else {
|
|
EmitUint8(0xF6);
|
|
EmitUint8(0xC0 + reg);
|
|
}
|
|
EmitUint8(immediate.value() & 0xFF);
|
|
} else if (reg == EAX) {
|
|
// Use short form if the destination is EAX.
|
|
EmitUint8(0xA9);
|
|
EmitImmediate(immediate);
|
|
} else {
|
|
EmitUint8(0xF7);
|
|
EmitOperand(0, Operand(reg));
|
|
EmitImmediate(immediate);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::testb(const Address& address, const Immediate& imm) {
|
|
ASSERT(imm.is_int8());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF6);
|
|
EmitOperand(0, address);
|
|
EmitUint8(imm.value() & 0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::andl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x23);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::andl(Register dst, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(4, Operand(dst), imm);
|
|
}
|
|
|
|
|
|
void Assembler::andl(Register dst, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x23);
|
|
EmitOperand(dst, address);
|
|
}
|
|
|
|
|
|
void Assembler::orl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0B);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::orl(Register dst, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(1, Operand(dst), imm);
|
|
}
|
|
|
|
|
|
void Assembler::orl(Register dst, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0B);
|
|
EmitOperand(dst, address);
|
|
}
|
|
|
|
|
|
void Assembler::orl(const Address& address, Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x09);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::xorl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x33);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::xorl(Register dst, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(6, Operand(dst), imm);
|
|
}
|
|
|
|
|
|
void Assembler::xorl(Register dst, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x33);
|
|
EmitOperand(dst, address);
|
|
}
|
|
|
|
|
|
void Assembler::addl(Register reg, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(0, Operand(reg), imm);
|
|
}
|
|
|
|
|
|
void Assembler::addl(const Address& address, Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x01);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::addl(const Address& address, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(0, address, imm);
|
|
}
|
|
|
|
|
|
void Assembler::adcl(Register reg, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(2, Operand(reg), imm);
|
|
}
|
|
|
|
|
|
void Assembler::adcl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x13);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::adcl(Register dst, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x13);
|
|
EmitOperand(dst, address);
|
|
}
|
|
|
|
|
|
void Assembler::adcl(const Address& address, Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x11);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::subl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x2B);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::subl(Register reg, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(5, Operand(reg), imm);
|
|
}
|
|
|
|
|
|
void Assembler::subl(Register reg, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x2B);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::subl(const Address& address, Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x29);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::cdq() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x99);
|
|
}
|
|
|
|
|
|
void Assembler::idivl(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF7);
|
|
EmitOperand(7, Operand(reg));
|
|
}
|
|
|
|
|
|
void Assembler::divl(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF7);
|
|
EmitOperand(6, Operand(reg));
|
|
}
|
|
|
|
|
|
void Assembler::imull(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xAF);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::imull(Register reg, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x69);
|
|
EmitOperand(reg, Operand(reg));
|
|
EmitImmediate(imm);
|
|
}
|
|
|
|
|
|
void Assembler::imull(Register reg, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xAF);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::imull(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF7);
|
|
EmitOperand(5, Operand(reg));
|
|
}
|
|
|
|
|
|
void Assembler::imull(const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF7);
|
|
EmitOperand(5, address);
|
|
}
|
|
|
|
|
|
void Assembler::mull(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF7);
|
|
EmitOperand(4, Operand(reg));
|
|
}
|
|
|
|
|
|
void Assembler::mull(const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF7);
|
|
EmitOperand(4, address);
|
|
}
|
|
|
|
|
|
void Assembler::sbbl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x1B);
|
|
EmitOperand(dst, Operand(src));
|
|
}
|
|
|
|
|
|
void Assembler::sbbl(Register reg, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitComplex(3, Operand(reg), imm);
|
|
}
|
|
|
|
|
|
void Assembler::sbbl(Register dst, const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x1B);
|
|
EmitOperand(dst, address);
|
|
}
|
|
|
|
|
|
void Assembler::sbbl(const Address& address, Register dst) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x19);
|
|
EmitOperand(dst, address);
|
|
}
|
|
|
|
|
|
void Assembler::incl(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x40 + reg);
|
|
}
|
|
|
|
|
|
void Assembler::incl(const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xFF);
|
|
EmitOperand(0, address);
|
|
}
|
|
|
|
|
|
void Assembler::decl(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x48 + reg);
|
|
}
|
|
|
|
|
|
void Assembler::decl(const Address& address) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xFF);
|
|
EmitOperand(1, address);
|
|
}
|
|
|
|
|
|
void Assembler::shll(Register reg, const Immediate& imm) {
|
|
EmitGenericShift(4, reg, imm);
|
|
}
|
|
|
|
|
|
void Assembler::shll(Register operand, Register shifter) {
|
|
EmitGenericShift(4, Operand(operand), shifter);
|
|
}
|
|
|
|
|
|
void Assembler::shll(const Address& operand, Register shifter) {
|
|
EmitGenericShift(4, Operand(operand), shifter);
|
|
}
|
|
|
|
|
|
void Assembler::shrl(Register reg, const Immediate& imm) {
|
|
EmitGenericShift(5, reg, imm);
|
|
}
|
|
|
|
|
|
void Assembler::shrl(Register operand, Register shifter) {
|
|
EmitGenericShift(5, Operand(operand), shifter);
|
|
}
|
|
|
|
|
|
void Assembler::sarl(Register reg, const Immediate& imm) {
|
|
EmitGenericShift(7, reg, imm);
|
|
}
|
|
|
|
|
|
void Assembler::sarl(Register operand, Register shifter) {
|
|
EmitGenericShift(7, Operand(operand), shifter);
|
|
}
|
|
|
|
|
|
void Assembler::sarl(const Address& address, Register shifter) {
|
|
EmitGenericShift(7, Operand(address), shifter);
|
|
}
|
|
|
|
|
|
void Assembler::shldl(Register dst, Register src, Register shifter) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
ASSERT(shifter == ECX);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xA5);
|
|
EmitRegisterOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::shldl(Register dst, Register src, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
ASSERT(imm.is_int8());
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xA4);
|
|
EmitRegisterOperand(src, dst);
|
|
EmitUint8(imm.value() & 0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::shldl(const Address& operand, Register src, Register shifter) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
ASSERT(shifter == ECX);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xA5);
|
|
EmitOperand(src, Operand(operand));
|
|
}
|
|
|
|
|
|
void Assembler::shrdl(Register dst, Register src, Register shifter) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
ASSERT(shifter == ECX);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xAD);
|
|
EmitRegisterOperand(src, dst);
|
|
}
|
|
|
|
|
|
void Assembler::shrdl(Register dst, Register src, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
ASSERT(imm.is_int8());
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xAC);
|
|
EmitRegisterOperand(src, dst);
|
|
EmitUint8(imm.value() & 0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::shrdl(const Address& dst, Register src, Register shifter) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
ASSERT(shifter == ECX);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xAD);
|
|
EmitOperand(src, Operand(dst));
|
|
}
|
|
|
|
|
|
void Assembler::negl(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF7);
|
|
EmitOperand(3, Operand(reg));
|
|
}
|
|
|
|
|
|
void Assembler::notl(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF7);
|
|
EmitUint8(0xD0 | reg);
|
|
}
|
|
|
|
|
|
void Assembler::bsrl(Register dst, Register src) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xBD);
|
|
EmitRegisterOperand(dst, src);
|
|
}
|
|
|
|
|
|
void Assembler::bt(Register base, Register offset) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xA3);
|
|
EmitRegisterOperand(offset, base);
|
|
}
|
|
|
|
|
|
void Assembler::enter(const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xC8);
|
|
ASSERT(imm.is_uint16());
|
|
EmitUint8(imm.value() & 0xFF);
|
|
EmitUint8((imm.value() >> 8) & 0xFF);
|
|
EmitUint8(0x00);
|
|
}
|
|
|
|
|
|
void Assembler::leave() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xC9);
|
|
}
|
|
|
|
|
|
void Assembler::ret() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xC3);
|
|
}
|
|
|
|
|
|
void Assembler::ret(const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xC2);
|
|
ASSERT(imm.is_uint16());
|
|
EmitUint8(imm.value() & 0xFF);
|
|
EmitUint8((imm.value() >> 8) & 0xFF);
|
|
}
|
|
|
|
|
|
void Assembler::nop(int size) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
// There are nops up to size 15, but for now just provide up to size 8.
|
|
ASSERT(0 < size && size <= MAX_NOP_SIZE);
|
|
switch (size) {
|
|
case 1:
|
|
EmitUint8(0x90);
|
|
break;
|
|
case 2:
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x90);
|
|
break;
|
|
case 3:
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x1F);
|
|
EmitUint8(0x00);
|
|
break;
|
|
case 4:
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x1F);
|
|
EmitUint8(0x40);
|
|
EmitUint8(0x00);
|
|
break;
|
|
case 5:
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x1F);
|
|
EmitUint8(0x44);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
break;
|
|
case 6:
|
|
EmitUint8(0x66);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x1F);
|
|
EmitUint8(0x44);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
break;
|
|
case 7:
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x1F);
|
|
EmitUint8(0x80);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
break;
|
|
case 8:
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x1F);
|
|
EmitUint8(0x84);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
EmitUint8(0x00);
|
|
break;
|
|
default:
|
|
UNIMPLEMENTED();
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::int3() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xCC);
|
|
}
|
|
|
|
|
|
void Assembler::hlt() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF4);
|
|
}
|
|
|
|
|
|
void Assembler::j(Condition condition, Label* label, bool near) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
if (label->IsBound()) {
|
|
static const int kShortSize = 2;
|
|
static const int kLongSize = 6;
|
|
intptr_t offset = label->Position() - buffer_.Size();
|
|
ASSERT(offset <= 0);
|
|
if (Utils::IsInt(8, offset - kShortSize)) {
|
|
EmitUint8(0x70 + condition);
|
|
EmitUint8((offset - kShortSize) & 0xFF);
|
|
} else {
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x80 + condition);
|
|
EmitInt32(offset - kLongSize);
|
|
}
|
|
} else if (near) {
|
|
EmitUint8(0x70 + condition);
|
|
EmitNearLabelLink(label);
|
|
} else {
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x80 + condition);
|
|
EmitLabelLink(label);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::j(Condition condition, const ExternalLabel* label) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0x80 + condition);
|
|
EmitFixup(new DirectCallRelocation());
|
|
EmitInt32(label->address());
|
|
}
|
|
|
|
|
|
void Assembler::jmp(Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xFF);
|
|
EmitRegisterOperand(4, reg);
|
|
}
|
|
|
|
|
|
void Assembler::jmp(Label* label, bool near) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
if (label->IsBound()) {
|
|
static const int kShortSize = 2;
|
|
static const int kLongSize = 5;
|
|
intptr_t offset = label->Position() - buffer_.Size();
|
|
ASSERT(offset <= 0);
|
|
if (Utils::IsInt(8, offset - kShortSize)) {
|
|
EmitUint8(0xEB);
|
|
EmitUint8((offset - kShortSize) & 0xFF);
|
|
} else {
|
|
EmitUint8(0xE9);
|
|
EmitInt32(offset - kLongSize);
|
|
}
|
|
} else if (near) {
|
|
EmitUint8(0xEB);
|
|
EmitNearLabelLink(label);
|
|
} else {
|
|
EmitUint8(0xE9);
|
|
EmitLabelLink(label);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::jmp(const ExternalLabel* label) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xE9);
|
|
EmitFixup(new DirectCallRelocation());
|
|
EmitInt32(label->address());
|
|
}
|
|
|
|
|
|
void Assembler::lock() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xF0);
|
|
}
|
|
|
|
|
|
void Assembler::cmpxchgl(const Address& address, Register reg) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xB1);
|
|
EmitOperand(reg, address);
|
|
}
|
|
|
|
|
|
void Assembler::cpuid() {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x0F);
|
|
EmitUint8(0xA2);
|
|
}
|
|
|
|
|
|
void Assembler::CompareRegisters(Register a, Register b) {
|
|
cmpl(a, b);
|
|
}
|
|
|
|
|
|
void Assembler::MoveRegister(Register to, Register from) {
|
|
if (to != from) {
|
|
movl(to, from);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::PopRegister(Register r) {
|
|
popl(r);
|
|
}
|
|
|
|
|
|
void Assembler::AddImmediate(Register reg, const Immediate& imm) {
|
|
const intptr_t value = imm.value();
|
|
if (value == 0) {
|
|
return;
|
|
}
|
|
if ((value > 0) || (value == kMinInt32)) {
|
|
if (value == 1) {
|
|
incl(reg);
|
|
} else {
|
|
addl(reg, imm);
|
|
}
|
|
} else {
|
|
SubImmediate(reg, Immediate(-value));
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::SubImmediate(Register reg, const Immediate& imm) {
|
|
const intptr_t value = imm.value();
|
|
if (value == 0) {
|
|
return;
|
|
}
|
|
if ((value > 0) || (value == kMinInt32)) {
|
|
if (value == 1) {
|
|
decl(reg);
|
|
} else {
|
|
subl(reg, imm);
|
|
}
|
|
} else {
|
|
AddImmediate(reg, Immediate(-value));
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::Drop(intptr_t stack_elements) {
|
|
ASSERT(stack_elements >= 0);
|
|
if (stack_elements > 0) {
|
|
addl(ESP, Immediate(stack_elements * kWordSize));
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::LoadIsolate(Register dst) {
|
|
movl(dst, Address(THR, Thread::isolate_offset()));
|
|
}
|
|
|
|
|
|
void Assembler::LoadObject(Register dst, const Object& object) {
|
|
ASSERT(!object.IsICData() || ICData::Cast(object).IsOriginal());
|
|
ASSERT(!object.IsField() || Field::Cast(object).IsOriginal());
|
|
if (object.IsSmi() || object.InVMHeap()) {
|
|
movl(dst, Immediate(reinterpret_cast<int32_t>(object.raw())));
|
|
} else {
|
|
ASSERT(object.IsNotTemporaryScopedHandle());
|
|
ASSERT(object.IsOld());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xB8 + dst);
|
|
buffer_.EmitObject(object);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::LoadObjectSafely(Register dst, const Object& object) {
|
|
ASSERT(!object.IsICData() || ICData::Cast(object).IsOriginal());
|
|
ASSERT(!object.IsField() || Field::Cast(object).IsOriginal());
|
|
if (Assembler::IsSafe(object)) {
|
|
LoadObject(dst, object);
|
|
} else {
|
|
int32_t cookie = jit_cookie();
|
|
movl(dst, Immediate(reinterpret_cast<int32_t>(object.raw()) ^ cookie));
|
|
xorl(dst, Immediate(cookie));
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::PushObject(const Object& object) {
|
|
ASSERT(!object.IsICData() || ICData::Cast(object).IsOriginal());
|
|
ASSERT(!object.IsField() || Field::Cast(object).IsOriginal());
|
|
if (object.IsSmi() || object.InVMHeap()) {
|
|
pushl(Immediate(reinterpret_cast<int32_t>(object.raw())));
|
|
} else {
|
|
ASSERT(object.IsNotTemporaryScopedHandle());
|
|
ASSERT(object.IsOld());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x68);
|
|
buffer_.EmitObject(object);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::CompareObject(Register reg, const Object& object) {
|
|
ASSERT(!object.IsICData() || ICData::Cast(object).IsOriginal());
|
|
ASSERT(!object.IsField() || Field::Cast(object).IsOriginal());
|
|
if (object.IsSmi() || object.InVMHeap()) {
|
|
cmpl(reg, Immediate(reinterpret_cast<int32_t>(object.raw())));
|
|
} else {
|
|
ASSERT(object.IsNotTemporaryScopedHandle());
|
|
ASSERT(object.IsOld());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
if (reg == EAX) {
|
|
EmitUint8(0x05 + (7 << 3));
|
|
buffer_.EmitObject(object);
|
|
} else {
|
|
EmitUint8(0x81);
|
|
EmitOperand(7, Operand(reg));
|
|
buffer_.EmitObject(object);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// Destroys the value register.
|
|
void Assembler::StoreIntoObjectFilterNoSmi(Register object,
|
|
Register value,
|
|
Label* no_update) {
|
|
COMPILE_ASSERT((kNewObjectAlignmentOffset == kWordSize) &&
|
|
(kOldObjectAlignmentOffset == 0));
|
|
|
|
// Write-barrier triggers if the value is in the new space (has bit set) and
|
|
// the object is in the old space (has bit cleared).
|
|
// To check that we could compute value & ~object and skip the write barrier
|
|
// if the bit is not set. However we can't destroy the object.
|
|
// However to preserve the object we compute negated expression
|
|
// ~value | object instead and skip the write barrier if the bit is set.
|
|
notl(value);
|
|
orl(value, object);
|
|
testl(value, Immediate(kNewObjectAlignmentOffset));
|
|
j(NOT_ZERO, no_update, Assembler::kNearJump);
|
|
}
|
|
|
|
|
|
// Destroys the value register.
|
|
void Assembler::StoreIntoObjectFilter(Register object,
|
|
Register value,
|
|
Label* no_update) {
|
|
// For the value we are only interested in the new/old bit and the tag bit.
|
|
andl(value, Immediate(kNewObjectAlignmentOffset | kHeapObjectTag));
|
|
// Shift the tag bit into the carry.
|
|
shrl(value, Immediate(1));
|
|
// Add the tag bits together, if the value is not a Smi the addition will
|
|
// overflow into the next bit, leaving us with a zero low bit.
|
|
adcl(value, object);
|
|
// Mask out higher, uninteresting bits which were polluted by dest.
|
|
andl(value, Immediate(kObjectAlignment - 1));
|
|
// Compare with the expected bit pattern.
|
|
cmpl(value, Immediate((kNewObjectAlignmentOffset >> 1) + kHeapObjectTag +
|
|
kOldObjectAlignmentOffset + kHeapObjectTag));
|
|
j(NOT_ZERO, no_update, Assembler::kNearJump);
|
|
}
|
|
|
|
|
|
// Destroys the value register.
|
|
void Assembler::StoreIntoObject(Register object,
|
|
const Address& dest,
|
|
Register value,
|
|
bool can_value_be_smi) {
|
|
ASSERT(object != value);
|
|
movl(dest, value);
|
|
Label done;
|
|
if (can_value_be_smi) {
|
|
StoreIntoObjectFilter(object, value, &done);
|
|
} else {
|
|
StoreIntoObjectFilterNoSmi(object, value, &done);
|
|
}
|
|
// A store buffer update is required.
|
|
if (value != EDX) {
|
|
pushl(EDX); // Preserve EDX.
|
|
}
|
|
if (object != EDX) {
|
|
movl(EDX, object);
|
|
}
|
|
call(Address(THR, Thread::update_store_buffer_entry_point_offset()));
|
|
if (value != EDX) {
|
|
popl(EDX); // Restore EDX.
|
|
}
|
|
Bind(&done);
|
|
}
|
|
|
|
|
|
void Assembler::StoreIntoObjectNoBarrier(Register object,
|
|
const Address& dest,
|
|
Register value) {
|
|
movl(dest, value);
|
|
#if defined(DEBUG)
|
|
Label done;
|
|
pushl(value);
|
|
StoreIntoObjectFilter(object, value, &done);
|
|
Stop("Store buffer update is required");
|
|
Bind(&done);
|
|
popl(value);
|
|
#endif // defined(DEBUG)
|
|
// No store buffer update.
|
|
}
|
|
|
|
|
|
void Assembler::UnverifiedStoreOldObject(const Address& dest,
|
|
const Object& value) {
|
|
ASSERT(!value.IsICData() || ICData::Cast(value).IsOriginal());
|
|
ASSERT(!value.IsField() || Field::Cast(value).IsOriginal());
|
|
ASSERT(value.IsOld());
|
|
ASSERT(!value.InVMHeap());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0xC7);
|
|
EmitOperand(0, dest);
|
|
buffer_.EmitObject(value);
|
|
}
|
|
|
|
|
|
void Assembler::StoreIntoObjectNoBarrier(Register object,
|
|
const Address& dest,
|
|
const Object& value) {
|
|
ASSERT(!value.IsICData() || ICData::Cast(value).IsOriginal());
|
|
ASSERT(!value.IsField() || Field::Cast(value).IsOriginal());
|
|
if (value.IsSmi() || value.InVMHeap()) {
|
|
Immediate imm_value(reinterpret_cast<int32_t>(value.raw()));
|
|
movl(dest, imm_value);
|
|
} else {
|
|
UnverifiedStoreOldObject(dest, value);
|
|
}
|
|
// No store buffer update.
|
|
}
|
|
|
|
|
|
void Assembler::StoreIntoSmiField(const Address& dest, Register value) {
|
|
#if defined(DEBUG)
|
|
Label done;
|
|
testl(value, Immediate(kHeapObjectTag));
|
|
j(ZERO, &done);
|
|
Stop("New value must be Smi.");
|
|
Bind(&done);
|
|
#endif // defined(DEBUG)
|
|
movl(dest, value);
|
|
}
|
|
|
|
|
|
void Assembler::ZeroInitSmiField(const Address& dest) {
|
|
Immediate zero(Smi::RawValue(0));
|
|
movl(dest, zero);
|
|
}
|
|
|
|
|
|
void Assembler::IncrementSmiField(const Address& dest, int32_t increment) {
|
|
// Note: FlowGraphCompiler::EdgeCounterIncrementSizeInBytes depends on
|
|
// the length of this instruction sequence.
|
|
Immediate inc_imm(Smi::RawValue(increment));
|
|
addl(dest, inc_imm);
|
|
}
|
|
|
|
|
|
void Assembler::LoadDoubleConstant(XmmRegister dst, double value) {
|
|
// TODO(5410843): Need to have a code constants table.
|
|
int64_t constant = bit_cast<int64_t, double>(value);
|
|
pushl(Immediate(Utils::High32Bits(constant)));
|
|
pushl(Immediate(Utils::Low32Bits(constant)));
|
|
movsd(dst, Address(ESP, 0));
|
|
addl(ESP, Immediate(2 * kWordSize));
|
|
}
|
|
|
|
|
|
void Assembler::FloatNegate(XmmRegister f) {
|
|
static const struct ALIGN16 {
|
|
uint32_t a;
|
|
uint32_t b;
|
|
uint32_t c;
|
|
uint32_t d;
|
|
} float_negate_constant = {0x80000000, 0x00000000, 0x80000000, 0x00000000};
|
|
xorps(f, Address::Absolute(reinterpret_cast<uword>(&float_negate_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::DoubleNegate(XmmRegister d) {
|
|
static const struct ALIGN16 {
|
|
uint64_t a;
|
|
uint64_t b;
|
|
} double_negate_constant = {0x8000000000000000LL, 0x8000000000000000LL};
|
|
xorpd(d, Address::Absolute(reinterpret_cast<uword>(&double_negate_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::DoubleAbs(XmmRegister reg) {
|
|
static const struct ALIGN16 {
|
|
uint64_t a;
|
|
uint64_t b;
|
|
} double_abs_constant = {0x7FFFFFFFFFFFFFFFLL, 0x7FFFFFFFFFFFFFFFLL};
|
|
andpd(reg, Address::Absolute(reinterpret_cast<uword>(&double_abs_constant)));
|
|
}
|
|
|
|
|
|
void Assembler::EnterFrame(intptr_t frame_size) {
|
|
if (prologue_offset_ == -1) {
|
|
Comment("PrologueOffset = %" Pd "", CodeSize());
|
|
prologue_offset_ = CodeSize();
|
|
}
|
|
#ifdef DEBUG
|
|
intptr_t check_offset = CodeSize();
|
|
#endif
|
|
pushl(EBP);
|
|
movl(EBP, ESP);
|
|
#ifdef DEBUG
|
|
ProloguePattern pp(CodeAddress(check_offset));
|
|
ASSERT(pp.IsValid());
|
|
#endif
|
|
if (frame_size != 0) {
|
|
Immediate frame_space(frame_size);
|
|
subl(ESP, frame_space);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::LeaveFrame() {
|
|
movl(ESP, EBP);
|
|
popl(EBP);
|
|
}
|
|
|
|
|
|
void Assembler::ReserveAlignedFrameSpace(intptr_t frame_space) {
|
|
// Reserve space for arguments and align frame before entering
|
|
// the C++ world.
|
|
AddImmediate(ESP, Immediate(-frame_space));
|
|
if (OS::ActivationFrameAlignment() > 1) {
|
|
andl(ESP, Immediate(~(OS::ActivationFrameAlignment() - 1)));
|
|
}
|
|
}
|
|
|
|
|
|
static const intptr_t kNumberOfVolatileCpuRegisters = 3;
|
|
static const Register volatile_cpu_registers[kNumberOfVolatileCpuRegisters] = {
|
|
EAX, ECX, EDX};
|
|
|
|
|
|
// XMM0 is used only as a scratch register in the optimized code. No need to
|
|
// save it.
|
|
static const intptr_t kNumberOfVolatileXmmRegisters = kNumberOfXmmRegisters - 1;
|
|
|
|
|
|
void Assembler::EnterCallRuntimeFrame(intptr_t frame_space) {
|
|
Comment("EnterCallRuntimeFrame");
|
|
EnterFrame(0);
|
|
|
|
// Preserve volatile CPU registers.
|
|
for (intptr_t i = 0; i < kNumberOfVolatileCpuRegisters; i++) {
|
|
pushl(volatile_cpu_registers[i]);
|
|
}
|
|
|
|
// Preserve all XMM registers except XMM0
|
|
subl(ESP, Immediate((kNumberOfXmmRegisters - 1) * kFpuRegisterSize));
|
|
// Store XMM registers with the lowest register number at the lowest
|
|
// address.
|
|
intptr_t offset = 0;
|
|
for (intptr_t reg_idx = 1; reg_idx < kNumberOfXmmRegisters; ++reg_idx) {
|
|
XmmRegister xmm_reg = static_cast<XmmRegister>(reg_idx);
|
|
movups(Address(ESP, offset), xmm_reg);
|
|
offset += kFpuRegisterSize;
|
|
}
|
|
|
|
ReserveAlignedFrameSpace(frame_space);
|
|
}
|
|
|
|
|
|
void Assembler::LeaveCallRuntimeFrame() {
|
|
// ESP might have been modified to reserve space for arguments
|
|
// and ensure proper alignment of the stack frame.
|
|
// We need to restore it before restoring registers.
|
|
const intptr_t kPushedRegistersSize =
|
|
kNumberOfVolatileCpuRegisters * kWordSize +
|
|
kNumberOfVolatileXmmRegisters * kFpuRegisterSize;
|
|
leal(ESP, Address(EBP, -kPushedRegistersSize));
|
|
|
|
// Restore all XMM registers except XMM0
|
|
// XMM registers have the lowest register number at the lowest address.
|
|
intptr_t offset = 0;
|
|
for (intptr_t reg_idx = 1; reg_idx < kNumberOfXmmRegisters; ++reg_idx) {
|
|
XmmRegister xmm_reg = static_cast<XmmRegister>(reg_idx);
|
|
movups(xmm_reg, Address(ESP, offset));
|
|
offset += kFpuRegisterSize;
|
|
}
|
|
addl(ESP, Immediate(offset));
|
|
|
|
// Restore volatile CPU registers.
|
|
for (intptr_t i = kNumberOfVolatileCpuRegisters - 1; i >= 0; i--) {
|
|
popl(volatile_cpu_registers[i]);
|
|
}
|
|
|
|
leave();
|
|
}
|
|
|
|
|
|
void Assembler::CallRuntime(const RuntimeEntry& entry,
|
|
intptr_t argument_count) {
|
|
entry.Call(this, argument_count);
|
|
}
|
|
|
|
|
|
void Assembler::Call(const StubEntry& stub_entry) {
|
|
const Code& target = Code::ZoneHandle(stub_entry.code());
|
|
LoadObject(CODE_REG, target);
|
|
call(FieldAddress(CODE_REG, Code::entry_point_offset()));
|
|
}
|
|
|
|
|
|
void Assembler::CallToRuntime() {
|
|
movl(CODE_REG, Address(THR, Thread::call_to_runtime_stub_offset()));
|
|
call(Address(THR, Thread::call_to_runtime_entry_point_offset()));
|
|
}
|
|
|
|
|
|
void Assembler::Jmp(const StubEntry& stub_entry) {
|
|
const ExternalLabel label(stub_entry.EntryPoint());
|
|
jmp(&label);
|
|
}
|
|
|
|
|
|
void Assembler::J(Condition condition, const StubEntry& stub_entry) {
|
|
const ExternalLabel label(stub_entry.EntryPoint());
|
|
j(condition, &label);
|
|
}
|
|
|
|
|
|
void Assembler::Align(intptr_t alignment, intptr_t offset) {
|
|
ASSERT(Utils::IsPowerOfTwo(alignment));
|
|
intptr_t pos = offset + buffer_.GetPosition();
|
|
intptr_t mod = pos & (alignment - 1);
|
|
if (mod == 0) {
|
|
return;
|
|
}
|
|
intptr_t bytes_needed = alignment - mod;
|
|
while (bytes_needed > MAX_NOP_SIZE) {
|
|
nop(MAX_NOP_SIZE);
|
|
bytes_needed -= MAX_NOP_SIZE;
|
|
}
|
|
if (bytes_needed) {
|
|
nop(bytes_needed);
|
|
}
|
|
ASSERT(((offset + buffer_.GetPosition()) & (alignment - 1)) == 0);
|
|
}
|
|
|
|
|
|
void Assembler::Bind(Label* label) {
|
|
intptr_t bound = buffer_.Size();
|
|
ASSERT(!label->IsBound()); // Labels can only be bound once.
|
|
while (label->IsLinked()) {
|
|
intptr_t position = label->LinkPosition();
|
|
intptr_t next = buffer_.Load<int32_t>(position);
|
|
buffer_.Store<int32_t>(position, bound - (position + 4));
|
|
label->position_ = next;
|
|
}
|
|
while (label->HasNear()) {
|
|
intptr_t position = label->NearPosition();
|
|
intptr_t offset = bound - (position + 1);
|
|
ASSERT(Utils::IsInt(8, offset));
|
|
buffer_.Store<int8_t>(position, offset);
|
|
}
|
|
label->BindTo(bound);
|
|
}
|
|
|
|
|
|
#ifndef PRODUCT
|
|
void Assembler::MaybeTraceAllocation(intptr_t cid,
|
|
Register temp_reg,
|
|
Label* trace,
|
|
bool near_jump) {
|
|
ASSERT(cid > 0);
|
|
Address state_address(kNoRegister, 0);
|
|
intptr_t state_offset = ClassTable::StateOffsetFor(cid);
|
|
ASSERT(temp_reg != kNoRegister);
|
|
LoadIsolate(temp_reg);
|
|
intptr_t table_offset =
|
|
Isolate::class_table_offset() + ClassTable::TableOffsetFor(cid);
|
|
movl(temp_reg, Address(temp_reg, table_offset));
|
|
state_address = Address(temp_reg, state_offset);
|
|
testb(state_address, Immediate(ClassHeapStats::TraceAllocationMask()));
|
|
// We are tracing for this class, jump to the trace label which will use
|
|
// the allocation stub.
|
|
j(NOT_ZERO, trace, near_jump);
|
|
}
|
|
|
|
|
|
void Assembler::UpdateAllocationStats(intptr_t cid,
|
|
Register temp_reg,
|
|
Heap::Space space) {
|
|
ASSERT(cid > 0);
|
|
intptr_t counter_offset =
|
|
ClassTable::CounterOffsetFor(cid, space == Heap::kNew);
|
|
ASSERT(temp_reg != kNoRegister);
|
|
LoadIsolate(temp_reg);
|
|
intptr_t table_offset =
|
|
Isolate::class_table_offset() + ClassTable::TableOffsetFor(cid);
|
|
movl(temp_reg, Address(temp_reg, table_offset));
|
|
incl(Address(temp_reg, counter_offset));
|
|
}
|
|
|
|
|
|
void Assembler::UpdateAllocationStatsWithSize(intptr_t cid,
|
|
Register size_reg,
|
|
Register temp_reg,
|
|
Heap::Space space) {
|
|
ASSERT(cid > 0);
|
|
ASSERT(cid < kNumPredefinedCids);
|
|
UpdateAllocationStats(cid, temp_reg, space);
|
|
intptr_t size_offset = ClassTable::SizeOffsetFor(cid, space == Heap::kNew);
|
|
addl(Address(temp_reg, size_offset), size_reg);
|
|
}
|
|
|
|
|
|
void Assembler::UpdateAllocationStatsWithSize(intptr_t cid,
|
|
intptr_t size_in_bytes,
|
|
Register temp_reg,
|
|
Heap::Space space) {
|
|
ASSERT(cid > 0);
|
|
ASSERT(cid < kNumPredefinedCids);
|
|
UpdateAllocationStats(cid, temp_reg, space);
|
|
intptr_t size_offset = ClassTable::SizeOffsetFor(cid, space == Heap::kNew);
|
|
addl(Address(temp_reg, size_offset), Immediate(size_in_bytes));
|
|
}
|
|
#endif // !PRODUCT
|
|
|
|
|
|
void Assembler::TryAllocate(const Class& cls,
|
|
Label* failure,
|
|
bool near_jump,
|
|
Register instance_reg,
|
|
Register temp_reg) {
|
|
ASSERT(failure != NULL);
|
|
ASSERT(temp_reg != kNoRegister);
|
|
if (FLAG_inline_alloc) {
|
|
// If this allocation is traced, program will jump to failure path
|
|
// (i.e. the allocation stub) which will allocate the object and trace the
|
|
// allocation call site.
|
|
NOT_IN_PRODUCT(
|
|
MaybeTraceAllocation(cls.id(), temp_reg, failure, near_jump));
|
|
const intptr_t instance_size = cls.instance_size();
|
|
Heap::Space space = Heap::kNew;
|
|
movl(temp_reg, Address(THR, Thread::heap_offset()));
|
|
movl(instance_reg, Address(temp_reg, Heap::TopOffset(space)));
|
|
addl(instance_reg, Immediate(instance_size));
|
|
// instance_reg: potential next object start.
|
|
cmpl(instance_reg, Address(temp_reg, Heap::EndOffset(space)));
|
|
j(ABOVE_EQUAL, failure, near_jump);
|
|
// Successfully allocated the object, now update top to point to
|
|
// next object start and store the class in the class field of object.
|
|
movl(Address(temp_reg, Heap::TopOffset(space)), instance_reg);
|
|
NOT_IN_PRODUCT(UpdateAllocationStats(cls.id(), temp_reg, space));
|
|
ASSERT(instance_size >= kHeapObjectTag);
|
|
subl(instance_reg, Immediate(instance_size - kHeapObjectTag));
|
|
uint32_t tags = 0;
|
|
tags = RawObject::SizeTag::update(instance_size, tags);
|
|
ASSERT(cls.id() != kIllegalCid);
|
|
tags = RawObject::ClassIdTag::update(cls.id(), tags);
|
|
movl(FieldAddress(instance_reg, Object::tags_offset()), Immediate(tags));
|
|
} else {
|
|
jmp(failure);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::TryAllocateArray(intptr_t cid,
|
|
intptr_t instance_size,
|
|
Label* failure,
|
|
bool near_jump,
|
|
Register instance,
|
|
Register end_address,
|
|
Register temp_reg) {
|
|
ASSERT(failure != NULL);
|
|
ASSERT(temp_reg != kNoRegister);
|
|
if (FLAG_inline_alloc) {
|
|
// If this allocation is traced, program will jump to failure path
|
|
// (i.e. the allocation stub) which will allocate the object and trace the
|
|
// allocation call site.
|
|
NOT_IN_PRODUCT(MaybeTraceAllocation(cid, temp_reg, failure, near_jump));
|
|
Heap::Space space = Heap::kNew;
|
|
movl(temp_reg, Address(THR, Thread::heap_offset()));
|
|
movl(instance, Address(temp_reg, Heap::TopOffset(space)));
|
|
movl(end_address, instance);
|
|
|
|
addl(end_address, Immediate(instance_size));
|
|
j(CARRY, failure);
|
|
|
|
// Check if the allocation fits into the remaining space.
|
|
// EAX: potential new object start.
|
|
// EBX: potential next object start.
|
|
cmpl(end_address, Address(temp_reg, Heap::EndOffset(space)));
|
|
j(ABOVE_EQUAL, failure);
|
|
|
|
// Successfully allocated the object(s), now update top to point to
|
|
// next object start and initialize the object.
|
|
movl(Address(temp_reg, Heap::TopOffset(space)), end_address);
|
|
addl(instance, Immediate(kHeapObjectTag));
|
|
NOT_IN_PRODUCT(
|
|
UpdateAllocationStatsWithSize(cid, instance_size, temp_reg, space));
|
|
|
|
// Initialize the tags.
|
|
uint32_t tags = 0;
|
|
tags = RawObject::ClassIdTag::update(cid, tags);
|
|
tags = RawObject::SizeTag::update(instance_size, tags);
|
|
movl(FieldAddress(instance, Object::tags_offset()), Immediate(tags));
|
|
} else {
|
|
jmp(failure);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::PushCodeObject() {
|
|
ASSERT(code_.IsNotTemporaryScopedHandle());
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
EmitUint8(0x68);
|
|
buffer_.EmitObject(code_);
|
|
}
|
|
|
|
|
|
void Assembler::EnterDartFrame(intptr_t frame_size) {
|
|
EnterFrame(0);
|
|
|
|
PushCodeObject();
|
|
|
|
if (frame_size != 0) {
|
|
subl(ESP, Immediate(frame_size));
|
|
}
|
|
}
|
|
|
|
|
|
// On entry to a function compiled for OSR, the caller's frame pointer, the
|
|
// stack locals, and any copied parameters are already in place. The frame
|
|
// pointer is already set up. There may be extra space for spill slots to
|
|
// allocate.
|
|
void Assembler::EnterOsrFrame(intptr_t extra_size) {
|
|
Comment("EnterOsrFrame");
|
|
if (prologue_offset_ == -1) {
|
|
Comment("PrologueOffset = %" Pd "", CodeSize());
|
|
prologue_offset_ = CodeSize();
|
|
}
|
|
|
|
if (extra_size != 0) {
|
|
subl(ESP, Immediate(extra_size));
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::EnterStubFrame() {
|
|
EnterDartFrame(0);
|
|
}
|
|
|
|
|
|
void Assembler::Stop(const char* message) {
|
|
if (FLAG_print_stop_message) {
|
|
pushl(EAX); // Preserve EAX.
|
|
movl(EAX, Immediate(reinterpret_cast<int32_t>(message)));
|
|
Call(*StubCode::PrintStopMessage_entry()); // Passing message in EAX.
|
|
popl(EAX); // Restore EAX.
|
|
} else {
|
|
// Emit the message address as immediate operand in the test instruction.
|
|
testl(EAX, Immediate(reinterpret_cast<int32_t>(message)));
|
|
}
|
|
// Emit the int3 instruction.
|
|
int3(); // Execution can be resumed with the 'cont' command in gdb.
|
|
}
|
|
|
|
|
|
void Assembler::EmitOperand(int rm, const Operand& operand) {
|
|
ASSERT(rm >= 0 && rm < 8);
|
|
const intptr_t length = operand.length_;
|
|
ASSERT(length > 0);
|
|
// Emit the ModRM byte updated with the given RM value.
|
|
ASSERT((operand.encoding_[0] & 0x38) == 0);
|
|
EmitUint8(operand.encoding_[0] + (rm << 3));
|
|
// Emit the rest of the encoded operand.
|
|
for (intptr_t i = 1; i < length; i++) {
|
|
EmitUint8(operand.encoding_[i]);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::EmitImmediate(const Immediate& imm) {
|
|
EmitInt32(imm.value());
|
|
}
|
|
|
|
|
|
void Assembler::EmitComplex(int rm,
|
|
const Operand& operand,
|
|
const Immediate& immediate) {
|
|
ASSERT(rm >= 0 && rm < 8);
|
|
if (immediate.is_int8()) {
|
|
// Use sign-extended 8-bit immediate.
|
|
EmitUint8(0x83);
|
|
EmitOperand(rm, operand);
|
|
EmitUint8(immediate.value() & 0xFF);
|
|
} else if (operand.IsRegister(EAX)) {
|
|
// Use short form if the destination is eax.
|
|
EmitUint8(0x05 + (rm << 3));
|
|
EmitImmediate(immediate);
|
|
} else {
|
|
EmitUint8(0x81);
|
|
EmitOperand(rm, operand);
|
|
EmitImmediate(immediate);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::EmitLabel(Label* label, intptr_t instruction_size) {
|
|
if (label->IsBound()) {
|
|
intptr_t offset = label->Position() - buffer_.Size();
|
|
ASSERT(offset <= 0);
|
|
EmitInt32(offset - instruction_size);
|
|
} else {
|
|
EmitLabelLink(label);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::EmitLabelLink(Label* label) {
|
|
ASSERT(!label->IsBound());
|
|
intptr_t position = buffer_.Size();
|
|
EmitInt32(label->position_);
|
|
label->LinkTo(position);
|
|
}
|
|
|
|
|
|
void Assembler::EmitNearLabelLink(Label* label) {
|
|
ASSERT(!label->IsBound());
|
|
intptr_t position = buffer_.Size();
|
|
EmitUint8(0);
|
|
label->NearLinkTo(position);
|
|
}
|
|
|
|
|
|
void Assembler::EmitGenericShift(int rm, Register reg, const Immediate& imm) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
ASSERT(imm.is_int8());
|
|
if (imm.value() == 1) {
|
|
EmitUint8(0xD1);
|
|
EmitOperand(rm, Operand(reg));
|
|
} else {
|
|
EmitUint8(0xC1);
|
|
EmitOperand(rm, Operand(reg));
|
|
EmitUint8(imm.value() & 0xFF);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::EmitGenericShift(int rm,
|
|
const Operand& operand,
|
|
Register shifter) {
|
|
AssemblerBuffer::EnsureCapacity ensured(&buffer_);
|
|
ASSERT(shifter == ECX);
|
|
EmitUint8(0xD3);
|
|
EmitOperand(rm, Operand(operand));
|
|
}
|
|
|
|
|
|
void Assembler::LoadClassId(Register result, Register object) {
|
|
ASSERT(RawObject::kClassIdTagPos == 16);
|
|
ASSERT(RawObject::kClassIdTagSize == 16);
|
|
const intptr_t class_id_offset =
|
|
Object::tags_offset() + RawObject::kClassIdTagPos / kBitsPerByte;
|
|
movzxw(result, FieldAddress(object, class_id_offset));
|
|
}
|
|
|
|
|
|
void Assembler::LoadClassById(Register result, Register class_id) {
|
|
ASSERT(result != class_id);
|
|
LoadIsolate(result);
|
|
const intptr_t offset =
|
|
Isolate::class_table_offset() + ClassTable::table_offset();
|
|
movl(result, Address(result, offset));
|
|
movl(result, Address(result, class_id, TIMES_4, 0));
|
|
}
|
|
|
|
|
|
void Assembler::LoadClass(Register result, Register object, Register scratch) {
|
|
ASSERT(scratch != result);
|
|
LoadClassId(scratch, object);
|
|
LoadClassById(result, scratch);
|
|
}
|
|
|
|
|
|
void Assembler::CompareClassId(Register object,
|
|
intptr_t class_id,
|
|
Register scratch) {
|
|
LoadClassId(scratch, object);
|
|
cmpl(scratch, Immediate(class_id));
|
|
}
|
|
|
|
|
|
void Assembler::SmiUntagOrCheckClass(Register object,
|
|
intptr_t class_id,
|
|
Register scratch,
|
|
Label* is_smi) {
|
|
ASSERT(kSmiTagShift == 1);
|
|
ASSERT(RawObject::kClassIdTagPos == 16);
|
|
ASSERT(RawObject::kClassIdTagSize == 16);
|
|
const intptr_t class_id_offset =
|
|
Object::tags_offset() + RawObject::kClassIdTagPos / kBitsPerByte;
|
|
|
|
// Untag optimistically. Tag bit is shifted into the CARRY.
|
|
SmiUntag(object);
|
|
j(NOT_CARRY, is_smi, kNearJump);
|
|
// Load cid: can't use LoadClassId, object is untagged. Use TIMES_2 scale
|
|
// factor in the addressing mode to compensate for this.
|
|
movzxw(scratch, Address(object, TIMES_2, class_id_offset));
|
|
cmpl(scratch, Immediate(class_id));
|
|
}
|
|
|
|
|
|
void Assembler::LoadClassIdMayBeSmi(Register result, Register object) {
|
|
if (result == object) {
|
|
Label smi, join;
|
|
|
|
testl(object, Immediate(kSmiTagMask));
|
|
j(EQUAL, &smi, Assembler::kNearJump);
|
|
LoadClassId(result, object);
|
|
jmp(&join, Assembler::kNearJump);
|
|
|
|
Bind(&smi);
|
|
movl(result, Immediate(kSmiCid));
|
|
|
|
Bind(&join);
|
|
} else {
|
|
ASSERT(result != object);
|
|
static const intptr_t kSmiCidSource = kSmiCid << RawObject::kClassIdTagPos;
|
|
|
|
// Make a dummy "Object" whose cid is kSmiCid.
|
|
movl(result, Immediate(reinterpret_cast<int32_t>(&kSmiCidSource) + 1));
|
|
|
|
// Check if object (in tmp) is a Smi.
|
|
testl(object, Immediate(kSmiTagMask));
|
|
|
|
// If the object is not a Smi, use the original object to load the cid.
|
|
// Otherwise, the dummy object is used, and the result is kSmiCid.
|
|
cmovne(result, object);
|
|
LoadClassId(result, result);
|
|
}
|
|
}
|
|
|
|
|
|
void Assembler::LoadTaggedClassIdMayBeSmi(Register result, Register object) {
|
|
if (result == object) {
|
|
Label smi, join;
|
|
|
|
testl(object, Immediate(kSmiTagMask));
|
|
j(EQUAL, &smi, Assembler::kNearJump);
|
|
LoadClassId(result, object);
|
|
SmiTag(result);
|
|
jmp(&join, Assembler::kNearJump);
|
|
|
|
Bind(&smi);
|
|
movl(result, Immediate(Smi::RawValue(kSmiCid)));
|
|
|
|
Bind(&join);
|
|
} else {
|
|
LoadClassIdMayBeSmi(result, object);
|
|
SmiTag(result);
|
|
}
|
|
}
|
|
|
|
|
|
Address Assembler::ElementAddressForIntIndex(bool is_external,
|
|
intptr_t cid,
|
|
intptr_t index_scale,
|
|
Register array,
|
|
intptr_t index) {
|
|
if (is_external) {
|
|
return Address(array, index * index_scale);
|
|
} else {
|
|
const int64_t disp = static_cast<int64_t>(index) * index_scale +
|
|
Instance::DataOffsetFor(cid);
|
|
ASSERT(Utils::IsInt(32, disp));
|
|
return FieldAddress(array, static_cast<int32_t>(disp));
|
|
}
|
|
}
|
|
|
|
|
|
static ScaleFactor ToScaleFactor(intptr_t index_scale) {
|
|
// Note that index is expected smi-tagged, (i.e, times 2) for all arrays with
|
|
// index scale factor > 1. E.g., for Uint8Array and OneByteString the index is
|
|
// expected to be untagged before accessing.
|
|
ASSERT(kSmiTagShift == 1);
|
|
switch (index_scale) {
|
|
case 1:
|
|
return TIMES_1;
|
|
case 2:
|
|
return TIMES_1;
|
|
case 4:
|
|
return TIMES_2;
|
|
case 8:
|
|
return TIMES_4;
|
|
case 16:
|
|
return TIMES_8;
|
|
default:
|
|
UNREACHABLE();
|
|
return TIMES_1;
|
|
}
|
|
}
|
|
|
|
|
|
Address Assembler::ElementAddressForRegIndex(bool is_external,
|
|
intptr_t cid,
|
|
intptr_t index_scale,
|
|
Register array,
|
|
Register index) {
|
|
if (is_external) {
|
|
return Address(array, index, ToScaleFactor(index_scale), 0);
|
|
} else {
|
|
return FieldAddress(array, index, ToScaleFactor(index_scale),
|
|
Instance::DataOffsetFor(cid));
|
|
}
|
|
}
|
|
|
|
|
|
static const char* cpu_reg_names[kNumberOfCpuRegisters] = {
|
|
"eax", "ecx", "edx", "ebx", "esp", "ebp", "esi", "edi"};
|
|
|
|
|
|
const char* Assembler::RegisterName(Register reg) {
|
|
ASSERT((0 <= reg) && (reg < kNumberOfCpuRegisters));
|
|
return cpu_reg_names[reg];
|
|
}
|
|
|
|
|
|
static const char* xmm_reg_names[kNumberOfXmmRegisters] = {
|
|
"xmm0", "xmm1", "xmm2", "xmm3", "xmm4", "xmm5", "xmm6", "xmm7"};
|
|
|
|
|
|
const char* Assembler::FpuRegisterName(FpuRegister reg) {
|
|
ASSERT((0 <= reg) && (reg < kNumberOfXmmRegisters));
|
|
return xmm_reg_names[reg];
|
|
}
|
|
|
|
|
|
} // namespace dart
|
|
|
|
#endif // defined TARGET_ARCH_IA32
|