mirror of
https://github.com/fluencelabs/wasmer
synced 2025-06-16 02:11:20 +00:00
Remove commented-out code that I added so as to not lose its history in git. Apply trivial cleanups and reformat.
Remove expected test failure entries that are fixed so far.
This commit is contained in:
@ -2823,116 +2823,11 @@ impl FunctionCodeGenerator<CodegenError> for X64FunctionCode {
|
||||
&mut self.machine,
|
||||
&mut self.value_stack,
|
||||
|a, src1, src2, dst| {
|
||||
/*
|
||||
// TODO: we don't have access to 'machine' here.
|
||||
//let tmp1 = self.machine.acquire_temp_gpr().unwrap();
|
||||
//let tmp2 = self.machine.acquire_temp_gpr().unwrap();
|
||||
let tmp1 = GPR::RAX;
|
||||
let tmp2 = GPR::RDX;
|
||||
a.emit_mov(Size::S32, Location::XMM(src1), Location::GPR(tmp1));
|
||||
match src2 {
|
||||
XMMOrMemory::XMM(x) => {
|
||||
a.emit_mov(Size::S32, Location::XMM(x), Location::GPR(tmp2))
|
||||
}
|
||||
XMMOrMemory::Memory(base, disp) => {
|
||||
a.emit_mov(Size::S32, Location::Memory(base, disp), Location::GPR(tmp2))
|
||||
}
|
||||
};
|
||||
let (tmp_xmm, _release_tmp_xmm) = if dst != src1 && XMMOrMemory::XMM(dst) != src2
|
||||
{
|
||||
(dst, false)
|
||||
} else {
|
||||
// TODO: we don't have access to 'machine' here.
|
||||
//(self.machine.acquire_temp_xmm().unwrap(), true)
|
||||
if src1 == XMM::XMM0 {
|
||||
if src2 == XMMOrMemory::XMM(XMM::XMM1) {
|
||||
(XMM::XMM2, false)
|
||||
} else {
|
||||
(XMM::XMM1, false)
|
||||
}
|
||||
} else {
|
||||
(XMM::XMM0, false)
|
||||
}
|
||||
};
|
||||
match src2 {
|
||||
XMMOrMemory::XMM(x) => {
|
||||
a.emit_mov(Size::S64, Location::XMM(x), Location::XMM(tmp_xmm))
|
||||
}
|
||||
XMMOrMemory::Memory(base, disp) => a.emit_mov(
|
||||
Size::S64,
|
||||
Location::Memory(base, disp),
|
||||
Location::XMM(tmp_xmm),
|
||||
),
|
||||
};
|
||||
a.emit_ucomiss(XMMOrMemory::XMM(src1), tmp_xmm);
|
||||
let do_vminss = a.get_label();
|
||||
a.emit_jmp(Condition::NotEqual, do_vminss);
|
||||
a.emit_jmp(Condition::ParityEven, do_vminss);
|
||||
a.emit_cmp(Size::S32, Location::GPR(tmp1), Location::GPR(tmp2));
|
||||
a.emit_jmp(Condition::Equal, do_vminss);
|
||||
static NEG_ZERO: u128 = 0x80000000;
|
||||
match src2 {
|
||||
XMMOrMemory::XMM(x) => {
|
||||
a.emit_mov(
|
||||
Size::S64,
|
||||
Location::Imm64((&NEG_ZERO as *const u128) as u64),
|
||||
Location::GPR(tmp2),
|
||||
);
|
||||
a.emit_mov(Size::S64, Location::Memory(tmp2, 0), Location::XMM(x));
|
||||
}
|
||||
XMMOrMemory::Memory(base, disp) => {
|
||||
let neg_zero_base = if base == tmp2 { tmp1 } else { tmp2 };
|
||||
a.emit_mov(
|
||||
Size::S64,
|
||||
Location::Imm64((&NEG_ZERO as *const u128) as u64),
|
||||
Location::GPR(neg_zero_base),
|
||||
);
|
||||
a.emit_mov(
|
||||
Size::S64,
|
||||
Location::Memory(neg_zero_base, 0),
|
||||
Location::Memory(base, disp),
|
||||
);
|
||||
}
|
||||
};
|
||||
a.emit_label(do_vminss);
|
||||
a.emit_vminss(src1, src2, dst);
|
||||
// TODO: we don't have access to 'machine' here.
|
||||
//if release_tmp_xmm { self.machine.release_temp_xmm(tmp_xmm) }
|
||||
*/
|
||||
let tmp_xmm1 = /*if dst != src1 && XMMOrMemory::XMM(dst) != src2
|
||||
{
|
||||
dst
|
||||
} else {
|
||||
// TODO: we don't have access to 'machine' here.
|
||||
//(self.machine.acquire_temp_xmm().unwrap(), true)
|
||||
if src1 == XMM::XMM0 {
|
||||
if src2 == XMMOrMemory::XMM(XMM::XMM1) {
|
||||
XMM::XMM2
|
||||
} else {
|
||||
XMM::XMM1
|
||||
}
|
||||
} else {
|
||||
XMM::XMM0
|
||||
}
|
||||
};*/ XMM::XMM6;
|
||||
let tmp_xmm2 = XMM::XMM7; // TODO: pick value safely.
|
||||
let tmp_xmm3 = XMM::XMM5; // TODO: pick value safely.
|
||||
/*
|
||||
let src2 = match src2 {
|
||||
XMMOrMemory::XMM(x) => x,
|
||||
XMMOrMemory::Memory(_, _) => panic!(),
|
||||
};
|
||||
a.emit_vminss(src1, XMMOrMemory::XMM(src2), src2);
|
||||
a.emit_vcmpunordss(src1, XMMOrMemory::XMM(src1), tmp_xmm);
|
||||
a.emit_vblendvps(src1, src2, XMMOrMemory::XMM(src1), tmp_xmm);
|
||||
a.emit_vcmpordss(src1, XMMOrMemory::XMM(src1), src2);
|
||||
let canonical_nan_base = GPR::RDX;
|
||||
static CANONICAL_NAN: u128 = 0x7FC0_0000;
|
||||
a.emit_mov(Size::S64, Location::Imm64((&CANONICAL_NAN as *const u128) as u64), Location::GPR(canonical_nan_base));
|
||||
a.emit_mov(Size::S64, Location::Memory(canonical_nan_base, 0), Location::XMM(tmp_xmm));
|
||||
a.emit_vblendvps(src1, tmp_xmm, XMMOrMemory::XMM(src1), src2);
|
||||
a.emit_mov(Size::S64, Location::XMM(src2), Location::XMM(dst));
|
||||
*/
|
||||
// TODO: pick values safely.
|
||||
let tmp_xmm1 = XMM::XMM7;
|
||||
let tmp_xmm2 = XMM::XMM6;
|
||||
let tmp_xmm3 = XMM::XMM5;
|
||||
|
||||
static NEG_ZERO: u128 = 0x8000_0000;
|
||||
static CANONICAL_NAN: u128 = 0x7FC0_0000;
|
||||
let loc2 = match src2 {
|
||||
@ -2940,7 +2835,6 @@ impl FunctionCodeGenerator<CodegenError> for X64FunctionCode {
|
||||
XMMOrMemory::Memory(base, disp) => Location::Memory(base, disp),
|
||||
};
|
||||
let spare_base = GPR::RDX;
|
||||
//a.emit_ud2();
|
||||
a.emit_mov(Size::S32, Location::XMM(src1), Location::GPR(GPR::RAX));
|
||||
a.emit_mov(Size::S32, loc2, Location::GPR(GPR::RDX));
|
||||
a.emit_cmp(Size::S32, Location::GPR(GPR::RDX), Location::GPR(GPR::RAX));
|
||||
@ -2956,15 +2850,31 @@ impl FunctionCodeGenerator<CodegenError> for X64FunctionCode {
|
||||
a.emit_jmp(Condition::None, label2);
|
||||
a.emit_label(label1);
|
||||
// load float -0.0
|
||||
a.emit_mov(Size::S64, Location::Imm64((&NEG_ZERO as *const u128) as u64), Location::GPR(spare_base));
|
||||
a.emit_mov(Size::S64, Location::Memory(spare_base, 0), Location::XMM(tmp_xmm2));
|
||||
a.emit_mov(
|
||||
Size::S64,
|
||||
Location::Imm64((&NEG_ZERO as *const u128) as u64),
|
||||
Location::GPR(spare_base),
|
||||
);
|
||||
a.emit_mov(
|
||||
Size::S64,
|
||||
Location::Memory(spare_base, 0),
|
||||
Location::XMM(tmp_xmm2),
|
||||
);
|
||||
a.emit_label(label2);
|
||||
a.emit_vcmpeqss(src1, XMMOrMemory::XMM(src2), tmp_xmm3);
|
||||
a.emit_vblendvps(tmp_xmm3, XMMOrMemory::XMM(tmp_xmm2), tmp_xmm1, tmp_xmm1);
|
||||
a.emit_vcmpunordss(src1, XMMOrMemory::XMM(src2), src1);
|
||||
// load float canonical nan
|
||||
a.emit_mov(Size::S64, Location::Imm64((&CANONICAL_NAN as *const u128) as u64), Location::GPR(spare_base));
|
||||
a.emit_mov(Size::S64, Location::Memory(spare_base, 0), Location::XMM(src2));
|
||||
a.emit_mov(
|
||||
Size::S64,
|
||||
Location::Imm64((&CANONICAL_NAN as *const u128) as u64),
|
||||
Location::GPR(spare_base),
|
||||
);
|
||||
a.emit_mov(
|
||||
Size::S64,
|
||||
Location::Memory(spare_base, 0),
|
||||
Location::XMM(src2),
|
||||
);
|
||||
a.emit_vblendvps(src1, XMMOrMemory::XMM(src2), tmp_xmm1, src1);
|
||||
a.emit_vmovaps(XMMOrMemory::XMM(src1), XMMOrMemory::XMM(dst));
|
||||
},
|
||||
|
Reference in New Issue
Block a user