/* This file is part of the dynarmic project. * Copyright (c) 2016 MerryMage * This software may be used and distributed according to the terms of the GNU * General Public License version 2 or any later version. */ #include #include #include #include #include #include "backend/x64/abi.h" #include "backend/x64/block_of_code.h" #include "backend/x64/emit_x64.h" #include "common/assert.h" #include "common/bit_util.h" #include "common/common_types.h" #include "common/math_util.h" #include "common/mp/function_info.h" #include "frontend/ir/basic_block.h" #include "frontend/ir/microinstruction.h" #include "frontend/ir/opcodes.h" namespace Dynarmic::BackendX64 { using namespace Xbyak::util; namespace mp = Common::mp; template static void EmitVectorOperation(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, Function fn) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); (code.*fn)(xmm_a, xmm_b); ctx.reg_alloc.DefineValue(inst, xmm_a); } template static void EmitAVXVectorOperation(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, Function fn) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); (code.*fn)(xmm_a, xmm_a, xmm_b); ctx.reg_alloc.DefineValue(inst, xmm_a); } template static void EmitOneArgumentFallback(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, Lambda lambda) { const auto fn = static_cast*>(lambda); constexpr u32 stack_space = 2 * 16; auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm arg1 = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); ctx.reg_alloc.EndOfAllocScope(); ctx.reg_alloc.HostCall(nullptr); code.sub(rsp, stack_space + ABI_SHADOW_SPACE); code.lea(code.ABI_PARAM1, ptr[rsp + ABI_SHADOW_SPACE + 0 * 16]); code.lea(code.ABI_PARAM2, ptr[rsp + ABI_SHADOW_SPACE + 1 * 16]); code.movaps(xword[code.ABI_PARAM2], arg1); code.CallFunction(fn); code.movaps(result, xword[rsp + ABI_SHADOW_SPACE + 0 * 16]); code.add(rsp, stack_space + ABI_SHADOW_SPACE); ctx.reg_alloc.DefineValue(inst, result); } template static void EmitOneArgumentFallbackWithSaturation(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, Lambda lambda) { const auto fn = static_cast*>(lambda); constexpr u32 stack_space = 2 * 16; auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm arg1 = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); ctx.reg_alloc.EndOfAllocScope(); ctx.reg_alloc.HostCall(nullptr); code.sub(rsp, stack_space + ABI_SHADOW_SPACE); code.lea(code.ABI_PARAM1, ptr[rsp + ABI_SHADOW_SPACE + 0 * 16]); code.lea(code.ABI_PARAM2, ptr[rsp + ABI_SHADOW_SPACE + 1 * 16]); code.movaps(xword[code.ABI_PARAM2], arg1); code.CallFunction(fn); code.movaps(result, xword[rsp + ABI_SHADOW_SPACE + 0 * 16]); code.add(rsp, stack_space + ABI_SHADOW_SPACE); code.or_(code.byte[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], code.ABI_RETURN.cvt8()); ctx.reg_alloc.DefineValue(inst, result); } template static void EmitTwoArgumentFallbackWithSaturation(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, Lambda lambda) { const auto fn = static_cast*>(lambda); constexpr u32 stack_space = 3 * 16; auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm arg1 = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm arg2 = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); ctx.reg_alloc.EndOfAllocScope(); ctx.reg_alloc.HostCall(nullptr); code.sub(rsp, stack_space + ABI_SHADOW_SPACE); code.lea(code.ABI_PARAM1, ptr[rsp + ABI_SHADOW_SPACE + 0 * 16]); code.lea(code.ABI_PARAM2, ptr[rsp + ABI_SHADOW_SPACE + 1 * 16]); code.lea(code.ABI_PARAM3, ptr[rsp + ABI_SHADOW_SPACE + 2 * 16]); code.movaps(xword[code.ABI_PARAM2], arg1); code.movaps(xword[code.ABI_PARAM3], arg2); code.CallFunction(fn); code.movaps(result, xword[rsp + ABI_SHADOW_SPACE + 0 * 16]); code.add(rsp, stack_space + ABI_SHADOW_SPACE); code.or_(code.byte[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], code.ABI_RETURN.cvt8()); ctx.reg_alloc.DefineValue(inst, result); } template static void EmitTwoArgumentFallback(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, Lambda lambda) { const auto fn = static_cast*>(lambda); constexpr u32 stack_space = 3 * 16; auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm arg1 = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm arg2 = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); ctx.reg_alloc.EndOfAllocScope(); ctx.reg_alloc.HostCall(nullptr); code.sub(rsp, stack_space + ABI_SHADOW_SPACE); code.lea(code.ABI_PARAM1, ptr[rsp + ABI_SHADOW_SPACE + 0 * 16]); code.lea(code.ABI_PARAM2, ptr[rsp + ABI_SHADOW_SPACE + 1 * 16]); code.lea(code.ABI_PARAM3, ptr[rsp + ABI_SHADOW_SPACE + 2 * 16]); code.movaps(xword[code.ABI_PARAM2], arg1); code.movaps(xword[code.ABI_PARAM3], arg2); code.CallFunction(fn); code.movaps(result, xword[rsp + ABI_SHADOW_SPACE + 0 * 16]); code.add(rsp, stack_space + ABI_SHADOW_SPACE); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorGetElement8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); ASSERT(args[1].IsImmediate()); const u8 index = args[1].GetImmediateU8(); if (index == 0) { ctx.reg_alloc.DefineValue(inst, args[0]); return; } const Xbyak::Xmm source = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Reg32 dest = ctx.reg_alloc.ScratchGpr().cvt32(); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pextrb(dest, source, index); } else { code.pextrw(dest, source, index / 2); if (index % 2 == 1) { code.shr(dest, 8); } } ctx.reg_alloc.DefineValue(inst, dest); } void EmitX64::EmitVectorGetElement16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); ASSERT(args[1].IsImmediate()); const u8 index = args[1].GetImmediateU8(); if (index == 0) { ctx.reg_alloc.DefineValue(inst, args[0]); return; } Xbyak::Xmm source = ctx.reg_alloc.UseXmm(args[0]); Xbyak::Reg32 dest = ctx.reg_alloc.ScratchGpr().cvt32(); code.pextrw(dest, source, index); ctx.reg_alloc.DefineValue(inst, dest); } void EmitX64::EmitVectorGetElement32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); ASSERT(args[1].IsImmediate()); const u8 index = args[1].GetImmediateU8(); if (index == 0) { ctx.reg_alloc.DefineValue(inst, args[0]); return; } Xbyak::Reg32 dest = ctx.reg_alloc.ScratchGpr().cvt32(); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { Xbyak::Xmm source = ctx.reg_alloc.UseXmm(args[0]); code.pextrd(dest, source, index); } else { Xbyak::Xmm source = ctx.reg_alloc.UseScratchXmm(args[0]); code.pshufd(source, source, index); code.movd(dest, source); } ctx.reg_alloc.DefineValue(inst, dest); } void EmitX64::EmitVectorGetElement64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); ASSERT(args[1].IsImmediate()); u8 index = args[1].GetImmediateU8(); if (index == 0) { ctx.reg_alloc.DefineValue(inst, args[0]); return; } Xbyak::Reg64 dest = ctx.reg_alloc.ScratchGpr().cvt64(); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { Xbyak::Xmm source = ctx.reg_alloc.UseXmm(args[0]); code.pextrq(dest, source, 1); } else { Xbyak::Xmm source = ctx.reg_alloc.UseScratchXmm(args[0]); code.punpckhqdq(source, source); code.movq(dest, source); } ctx.reg_alloc.DefineValue(inst, dest); } void EmitX64::EmitVectorSetElement8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); ASSERT(args[1].IsImmediate()); const u8 index = args[1].GetImmediateU8(); const Xbyak::Xmm source_vector = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { const Xbyak::Reg8 source_elem = ctx.reg_alloc.UseGpr(args[2]).cvt8(); code.pinsrb(source_vector, source_elem.cvt32(), index); ctx.reg_alloc.DefineValue(inst, source_vector); } else { const Xbyak::Reg32 source_elem = ctx.reg_alloc.UseScratchGpr(args[2]).cvt32(); const Xbyak::Reg32 tmp = ctx.reg_alloc.ScratchGpr().cvt32(); code.pextrw(tmp, source_vector, index / 2); if (index % 2 == 0) { code.and_(tmp, 0xFF00); code.and_(source_elem, 0x00FF); code.or_(tmp, source_elem); } else { code.and_(tmp, 0x00FF); code.shl(source_elem, 8); code.or_(tmp, source_elem); } code.pinsrw(source_vector, tmp, index / 2); ctx.reg_alloc.DefineValue(inst, source_vector); } } void EmitX64::EmitVectorSetElement16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); ASSERT(args[1].IsImmediate()); u8 index = args[1].GetImmediateU8(); Xbyak::Xmm source_vector = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Reg16 source_elem = ctx.reg_alloc.UseGpr(args[2]).cvt16(); code.pinsrw(source_vector, source_elem.cvt32(), index); ctx.reg_alloc.DefineValue(inst, source_vector); } void EmitX64::EmitVectorSetElement32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); ASSERT(args[1].IsImmediate()); const u8 index = args[1].GetImmediateU8(); const Xbyak::Xmm source_vector = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { const Xbyak::Reg32 source_elem = ctx.reg_alloc.UseGpr(args[2]).cvt32(); code.pinsrd(source_vector, source_elem, index); ctx.reg_alloc.DefineValue(inst, source_vector); } else { const Xbyak::Reg32 source_elem = ctx.reg_alloc.UseScratchGpr(args[2]).cvt32(); code.pinsrw(source_vector, source_elem, index * 2); code.shr(source_elem, 16); code.pinsrw(source_vector, source_elem, index * 2 + 1); ctx.reg_alloc.DefineValue(inst, source_vector); } } void EmitX64::EmitVectorSetElement64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); ASSERT(args[1].IsImmediate()); const u8 index = args[1].GetImmediateU8(); const Xbyak::Xmm source_vector = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { const Xbyak::Reg64 source_elem = ctx.reg_alloc.UseGpr(args[2]); code.pinsrq(source_vector, source_elem, index); ctx.reg_alloc.DefineValue(inst, source_vector); } else { const Xbyak::Reg64 source_elem = ctx.reg_alloc.UseGpr(args[2]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movq(tmp, source_elem); if (index == 0) { code.movsd(source_vector, tmp); } else { code.punpcklqdq(source_vector, tmp); } ctx.reg_alloc.DefineValue(inst, source_vector); } } static void VectorAbs8(BlockOfCode& code, EmitContext& ctx, const Xbyak::Xmm& data) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { code.pabsb(data, data); } else { const Xbyak::Xmm temp = ctx.reg_alloc.ScratchXmm(); code.pxor(temp, temp); code.psubb(temp, data); code.pminub(data, temp); } } static void VectorAbs16(BlockOfCode& code, EmitContext& ctx, const Xbyak::Xmm& data) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { code.pabsw(data, data); } else { const Xbyak::Xmm temp = ctx.reg_alloc.ScratchXmm(); code.pxor(temp, temp); code.psubw(temp, data); code.pmaxsw(data, temp); } } static void VectorAbs32(BlockOfCode& code, EmitContext& ctx, const Xbyak::Xmm& data) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { code.pabsd(data, data); } else { const Xbyak::Xmm temp = ctx.reg_alloc.ScratchXmm(); code.movdqa(temp, data); code.psrad(temp, 31); code.pxor(data, temp); code.psubd(data, temp); } } static void VectorAbs64(BlockOfCode& code, EmitContext& ctx, const Xbyak::Xmm& data) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { code.vpabsq(data, data); } else { const Xbyak::Xmm temp = ctx.reg_alloc.ScratchXmm(); code.pshufd(temp, data, 0b11110101); code.psrad(temp, 31); code.pxor(data, temp); code.psubq(data, temp); } } static void EmitVectorAbs(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm data = ctx.reg_alloc.UseScratchXmm(args[0]); switch (esize) { case 8: VectorAbs8(code, ctx, data); break; case 16: VectorAbs16(code, ctx, data); break; case 32: VectorAbs32(code, ctx, data); break; case 64: VectorAbs64(code, ctx, data); break; } ctx.reg_alloc.DefineValue(inst, data); } void EmitX64::EmitVectorAbs8(EmitContext& ctx, IR::Inst* inst) { EmitVectorAbs(8, ctx, inst, code); } void EmitX64::EmitVectorAbs16(EmitContext& ctx, IR::Inst* inst) { EmitVectorAbs(16, ctx, inst, code); } void EmitX64::EmitVectorAbs32(EmitContext& ctx, IR::Inst* inst) { EmitVectorAbs(32, ctx, inst, code); } void EmitX64::EmitVectorAbs64(EmitContext& ctx, IR::Inst* inst) { EmitVectorAbs(64, ctx, inst, code); } void EmitX64::EmitVectorAdd8(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::paddb); } void EmitX64::EmitVectorAdd16(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::paddw); } void EmitX64::EmitVectorAdd32(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::paddd); } void EmitX64::EmitVectorAdd64(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::paddq); } void EmitX64::EmitVectorAnd(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pand); } static void ArithmeticShiftRightByte(EmitContext& ctx, BlockOfCode& code, const Xbyak::Xmm& result, u8 shift_amount) { const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.punpckhbw(tmp, result); code.punpcklbw(result, result); code.psraw(tmp, 8 + shift_amount); code.psraw(result, 8 + shift_amount); code.packsswb(result, tmp); } void EmitX64::EmitVectorArithmeticShiftRight8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); ArithmeticShiftRightByte(ctx, code, result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorArithmeticShiftRight16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); code.psraw(result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorArithmeticShiftRight32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); code.psrad(result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorArithmeticShiftRight64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = std::min(args[1].GetImmediateU8(), u8(63)); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { code.vpsraq(result, result, shift_amount); } else { const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); const u64 sign_bit = 0x80000000'00000000u >> shift_amount; code.pxor(tmp2, tmp2); code.psrlq(result, shift_amount); code.movdqa(tmp1, code.MConst(xword, sign_bit, sign_bit)); code.pand(tmp1, result); code.psubq(tmp2, tmp1); code.por(result, tmp2); } ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorBroadcastLower8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX2)) { code.vpbroadcastb(a, a); code.vmovq(a, a); } else if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.pxor(tmp, tmp); code.pshufb(a, tmp); code.movq(a, a); } else { code.punpcklbw(a, a); code.pshuflw(a, a, 0); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorBroadcastLower16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); code.pshuflw(a, a, 0); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorBroadcastLower32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); code.pshuflw(a, a, 0b01000100); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorBroadcast8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX2)) { code.vpbroadcastb(a, a); } else if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.pxor(tmp, tmp); code.pshufb(a, tmp); } else { code.punpcklbw(a, a); code.pshuflw(a, a, 0); code.punpcklqdq(a, a); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorBroadcast16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX2)) { code.vpbroadcastw(a, a); } else { code.pshuflw(a, a, 0); code.punpcklqdq(a, a); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorBroadcast32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX2)) { code.vpbroadcastd(a, a); } else { code.pshufd(a, a, 0); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorBroadcast64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX2)) { code.vpbroadcastq(a, a); } else { code.punpcklqdq(a, a); } ctx.reg_alloc.DefineValue(inst, a); } template static void EmitVectorCountLeadingZeros(VectorArray& result, const VectorArray& data) { for (size_t i = 0; i < result.size(); i++) { T element = data[i]; size_t count = Common::BitSize(); while (element != 0) { element >>= 1; --count; } result[i] = static_cast(count); } } void EmitX64::EmitVectorCountLeadingZeros8(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallback(code, ctx, inst, EmitVectorCountLeadingZeros); } void EmitX64::EmitVectorCountLeadingZeros16(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallback(code, ctx, inst, EmitVectorCountLeadingZeros); } void EmitX64::EmitVectorCountLeadingZeros32(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512CD) && code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm data = ctx.reg_alloc.UseScratchXmm(args[0]); code.vplzcntd(data, data); ctx.reg_alloc.DefineValue(inst, data); return; } EmitOneArgumentFallback(code, ctx, inst, EmitVectorCountLeadingZeros); } void EmitX64::EmitVectorDeinterleaveEven8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm lhs = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm rhs = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, code.MConst(xword, 0x00FF00FF00FF00FF, 0x00FF00FF00FF00FF)); code.pand(lhs, tmp); code.pand(rhs, tmp); code.packuswb(lhs, rhs); ctx.reg_alloc.DefineValue(inst, lhs); } void EmitX64::EmitVectorDeinterleaveEven16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm lhs = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm rhs = ctx.reg_alloc.UseScratchXmm(args[1]); code.pslld(lhs, 16); code.psrad(lhs, 16); code.pslld(rhs, 16); code.psrad(rhs, 16); code.packssdw(lhs, rhs); ctx.reg_alloc.DefineValue(inst, lhs); } void EmitX64::EmitVectorDeinterleaveEven32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm lhs = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm rhs = ctx.reg_alloc.UseScratchXmm(args[1]); code.pshufd(lhs, lhs, 0b10001000); code.pshufd(rhs, rhs, 0b10001000); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pblendw(lhs, rhs, 0b11110000); } else { code.punpcklqdq(lhs, rhs); } ctx.reg_alloc.DefineValue(inst, lhs); } void EmitX64::EmitVectorDeinterleaveEven64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm lhs = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm rhs = ctx.reg_alloc.UseScratchXmm(args[1]); code.movq(lhs, lhs); code.pslldq(rhs, 8); code.por(lhs, rhs); ctx.reg_alloc.DefineValue(inst, lhs); } void EmitX64::EmitVectorDeinterleaveOdd8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm lhs = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm rhs = ctx.reg_alloc.UseScratchXmm(args[1]); code.psraw(lhs, 8); code.psraw(rhs, 8); code.packsswb(lhs, rhs); ctx.reg_alloc.DefineValue(inst, lhs); } void EmitX64::EmitVectorDeinterleaveOdd16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm lhs = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm rhs = ctx.reg_alloc.UseScratchXmm(args[1]); code.psrad(lhs, 16); code.psrad(rhs, 16); code.packssdw(lhs, rhs); ctx.reg_alloc.DefineValue(inst, lhs); } void EmitX64::EmitVectorDeinterleaveOdd32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm lhs = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm rhs = ctx.reg_alloc.UseScratchXmm(args[1]); code.pshufd(lhs, lhs, 0b11011101); code.pshufd(rhs, rhs, 0b11011101); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pblendw(lhs, rhs, 0b11110000); } else { code.punpcklqdq(lhs, rhs); } ctx.reg_alloc.DefineValue(inst, lhs); } void EmitX64::EmitVectorDeinterleaveOdd64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm lhs = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm rhs = ctx.reg_alloc.UseScratchXmm(args[1]); code.punpckhqdq(lhs, rhs); ctx.reg_alloc.DefineValue(inst, lhs); } void EmitX64::EmitVectorEor(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pxor); } void EmitX64::EmitVectorEqual8(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pcmpeqb); } void EmitX64::EmitVectorEqual16(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pcmpeqw); } void EmitX64::EmitVectorEqual32(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pcmpeqd); } void EmitX64::EmitVectorEqual64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pcmpeqq); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.pcmpeqd(xmm_a, xmm_b); code.pshufd(tmp, xmm_a, 0b10110001); code.pand(xmm_a, tmp); ctx.reg_alloc.DefineValue(inst, xmm_a); } void EmitX64::EmitVectorEqual128(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.pcmpeqq(xmm_a, xmm_b); code.pshufd(tmp, xmm_a, 0b01001110); code.pand(xmm_a, tmp); ctx.reg_alloc.DefineValue(inst, xmm_a); } else { Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.pcmpeqd(xmm_a, xmm_b); code.pshufd(tmp, xmm_a, 0b10110001); code.pand(xmm_a, tmp); code.pshufd(tmp, xmm_a, 0b01001110); code.pand(xmm_a, tmp); ctx.reg_alloc.DefineValue(inst, xmm_a); } } void EmitX64::EmitVectorExtract(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 position = args[2].GetImmediateU8(); ASSERT(position % 8 == 0); if (position != 0) { const Xbyak::Xmm xmm_b = ctx.reg_alloc.UseScratchXmm(args[1]); code.psrldq(xmm_a, position / 8); code.pslldq(xmm_b, (128 - position) / 8); code.por(xmm_a, xmm_b); } ctx.reg_alloc.DefineValue(inst, xmm_a); } void EmitX64::EmitVectorExtractLower(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 position = args[2].GetImmediateU8(); ASSERT(position % 8 == 0); if (position != 0) { const Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); code.punpcklqdq(xmm_a, xmm_b); code.psrldq(xmm_a, position / 8); } code.movq(xmm_a, xmm_a); ctx.reg_alloc.DefineValue(inst, xmm_a); } void EmitX64::EmitVectorGreaterS8(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pcmpgtb); } void EmitX64::EmitVectorGreaterS16(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pcmpgtw); } void EmitX64::EmitVectorGreaterS32(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pcmpgtd); } void EmitX64::EmitVectorGreaterS64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE42)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pcmpgtq); return; } EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { for (size_t i = 0; i < result.size(); ++i) { result[i] = (a[i] > b[i]) ? ~u64(0) : 0; } }); } static void EmitVectorHalvingAddSigned(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, b); code.pand(tmp, a); code.pxor(a, b); switch (esize) { case 8: ArithmeticShiftRightByte(ctx, code, a, 1); code.paddb(a, tmp); break; case 16: code.psraw(a, 1); code.paddw(a, tmp); break; case 32: code.psrad(a, 1); code.paddd(a, tmp); break; } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorHalvingAddS8(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingAddSigned(8, ctx, inst, code); } void EmitX64::EmitVectorHalvingAddS16(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingAddSigned(16, ctx, inst, code); } void EmitX64::EmitVectorHalvingAddS32(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingAddSigned(32, ctx, inst, code); } static void EmitVectorHalvingAddUnsigned(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, b); switch (esize) { case 8: code.pavgb(tmp, a); code.pxor(a, b); code.pand(a, code.MConst(xword, 0x0101010101010101, 0x0101010101010101)); code.psubb(tmp, a); break; case 16: code.pavgw(tmp, a); code.pxor(a, b); code.pand(a, code.MConst(xword, 0x0001000100010001, 0x0001000100010001)); code.psubw(tmp, a); break; case 32: code.pand(tmp, a); code.pxor(a, b); code.psrld(a, 1); code.paddd(tmp, a); break; } ctx.reg_alloc.DefineValue(inst, tmp); } void EmitX64::EmitVectorHalvingAddU8(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingAddUnsigned(8, ctx, inst, code); } void EmitX64::EmitVectorHalvingAddU16(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingAddUnsigned(16, ctx, inst, code); } void EmitX64::EmitVectorHalvingAddU32(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingAddUnsigned(32, ctx, inst, code); } static void EmitVectorHalvingSubSigned(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); switch (esize) { case 8: { const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, code.MConst(xword, 0x8080808080808080, 0x8080808080808080)); code.pxor(a, tmp); code.pxor(b, tmp); code.pavgb(b, a); code.psubb(a, b); break; } case 16: { const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, code.MConst(xword, 0x8000800080008000, 0x8000800080008000)); code.pxor(a, tmp); code.pxor(b, tmp); code.pavgw(b, a); code.psubw(a, b); break; } case 32: code.pxor(a, b); code.pand(b, a); code.psrad(a, 1); code.psubd(a, b); break; } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorHalvingSubS8(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingSubSigned(8, ctx, inst, code); } void EmitX64::EmitVectorHalvingSubS16(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingSubSigned(16, ctx, inst, code); } void EmitX64::EmitVectorHalvingSubS32(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingSubSigned(32, ctx, inst, code); } static void EmitVectorHalvingSubUnsigned(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); switch (esize) { case 8: code.pavgb(b, a); code.psubb(a, b); break; case 16: code.pavgw(b, a); code.psubw(a, b); break; case 32: code.pxor(a, b); code.pand(b, a); code.psrld(a, 1); code.psubd(a, b); break; } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorHalvingSubU8(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingSubUnsigned(8, ctx, inst, code); } void EmitX64::EmitVectorHalvingSubU16(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingSubUnsigned(16, ctx, inst, code); } void EmitX64::EmitVectorHalvingSubU32(EmitContext& ctx, IR::Inst* inst) { EmitVectorHalvingSubUnsigned(32, ctx, inst, code); } static void EmitVectorInterleaveLower(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, int size) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); switch (size) { case 8: code.punpcklbw(a, b); break; case 16: code.punpcklwd(a, b); break; case 32: code.punpckldq(a, b); break; case 64: code.punpcklqdq(a, b); break; } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorInterleaveLower8(EmitContext& ctx, IR::Inst* inst) { EmitVectorInterleaveLower(code, ctx, inst, 8); } void EmitX64::EmitVectorInterleaveLower16(EmitContext& ctx, IR::Inst* inst) { EmitVectorInterleaveLower(code, ctx, inst, 16); } void EmitX64::EmitVectorInterleaveLower32(EmitContext& ctx, IR::Inst* inst) { EmitVectorInterleaveLower(code, ctx, inst, 32); } void EmitX64::EmitVectorInterleaveLower64(EmitContext& ctx, IR::Inst* inst) { EmitVectorInterleaveLower(code, ctx, inst, 64); } static void EmitVectorInterleaveUpper(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, int size) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); switch (size) { case 8: code.punpckhbw(a, b); break; case 16: code.punpckhwd(a, b); break; case 32: code.punpckhdq(a, b); break; case 64: code.punpckhqdq(a, b); break; } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorInterleaveUpper8(EmitContext& ctx, IR::Inst* inst) { EmitVectorInterleaveUpper(code, ctx, inst, 8); } void EmitX64::EmitVectorInterleaveUpper16(EmitContext& ctx, IR::Inst* inst) { EmitVectorInterleaveUpper(code, ctx, inst, 16); } void EmitX64::EmitVectorInterleaveUpper32(EmitContext& ctx, IR::Inst* inst) { EmitVectorInterleaveUpper(code, ctx, inst, 32); } void EmitX64::EmitVectorInterleaveUpper64(EmitContext& ctx, IR::Inst* inst) { EmitVectorInterleaveUpper(code, ctx, inst, 64); } void EmitX64::EmitVectorLogicalShiftLeft8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); if (shift_amount == 1) { code.paddb(result, result); } else if (shift_amount > 0) { const u64 replicand = (0xFFULL << shift_amount) & 0xFF; const u64 mask = Common::Replicate(replicand, Common::BitSize()); code.psllw(result, shift_amount); code.pand(result, code.MConst(xword, mask, mask)); } ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorLogicalShiftLeft16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); code.psllw(result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorLogicalShiftLeft32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); code.pslld(result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorLogicalShiftLeft64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); code.psllq(result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorLogicalShiftRight8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); if (shift_amount > 0) { const u64 replicand = 0xFEULL >> shift_amount; const u64 mask = Common::Replicate(replicand, Common::BitSize()); code.psrlw(result, shift_amount); code.pand(result, code.MConst(xword, mask, mask)); } ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorLogicalShiftRight16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); code.psrlw(result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorLogicalShiftRight32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); code.psrld(result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorLogicalShiftRight64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const u8 shift_amount = args[1].GetImmediateU8(); code.psrlq(result, shift_amount); ctx.reg_alloc.DefineValue(inst, result); } template static constexpr T LogicalVShift(T x, T y) { const s8 shift_amount = static_cast(static_cast(y)); const s64 bit_size = static_cast(Common::BitSize()); if constexpr (std::is_signed_v) { if (shift_amount >= bit_size) { return 0; } if (shift_amount <= -bit_size) { // Parentheses necessary, as MSVC doesn't appear to consider cast parentheses // as a grouping in terms of precedence, causing warning C4554 to fire. See: // https://developercommunity.visualstudio.com/content/problem/144783/msvc-2017-does-not-understand-that-static-cast-cou.html return x >> (T(bit_size - 1)); } } else if (shift_amount <= -bit_size || shift_amount >= bit_size) { return 0; } if (shift_amount < 0) { return x >> T(-shift_amount); } using unsigned_type = std::make_unsigned_t; return static_cast(static_cast(x) << static_cast(shift_amount)); } void EmitX64::EmitVectorLogicalVShiftS8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), LogicalVShift); }); } void EmitX64::EmitVectorLogicalVShiftS16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), LogicalVShift); }); } void EmitX64::EmitVectorLogicalVShiftS32(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), LogicalVShift); }); } void EmitX64::EmitVectorLogicalVShiftS64(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), LogicalVShift); }); } void EmitX64::EmitVectorLogicalVShiftU8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), LogicalVShift); }); } void EmitX64::EmitVectorLogicalVShiftU16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), LogicalVShift); }); } void EmitX64::EmitVectorLogicalVShiftU32(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), LogicalVShift); }); } void EmitX64::EmitVectorLogicalVShiftU64(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), LogicalVShift); }); } void EmitX64::EmitVectorMaxS8(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pmaxsb); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp_b = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp_b, b); code.pcmpgtb(tmp_b, a); code.pand(b, tmp_b); code.pandn(tmp_b, a); code.por(tmp_b, b); ctx.reg_alloc.DefineValue(inst, tmp_b); } void EmitX64::EmitVectorMaxS16(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pmaxsw); } void EmitX64::EmitVectorMaxS32(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pmaxsd); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp_b = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp_b, b); code.pcmpgtd(tmp_b, a); code.pand(b, tmp_b); code.pandn(tmp_b, a); code.por(tmp_b, b); ctx.reg_alloc.DefineValue(inst, tmp_b); } void EmitX64::EmitVectorMaxS64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { EmitAVXVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::vpmaxsq); return; } if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); code.vpcmpgtq(xmm0, y, x); code.pblendvb(x, y); ctx.reg_alloc.DefineValue(inst, x); return; } EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), [](auto x, auto y) { return std::max(x, y); }); }); } void EmitX64::EmitVectorMaxU8(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pmaxub); } void EmitX64::EmitVectorMaxU16(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pmaxuw); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); code.psubusw(a, b); code.paddw(a, b); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorMaxU32(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pmaxud); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, code.MConst(xword, 0x8000000080000000, 0x8000000080000000)); const Xbyak::Xmm tmp_b = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp_b, b); code.pxor(tmp_b, tmp); code.pxor(tmp, a); code.pcmpgtd(tmp, tmp_b); code.pand(a, tmp); code.pandn(tmp, b); code.por(a, tmp); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorMaxU64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { EmitAVXVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::vpmaxuq); return; } if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.vmovdqa(xmm0, code.MConst(xword, 0x8000000000000000, 0x8000000000000000)); code.vpsubq(tmp, y, xmm0); code.vpsubq(xmm0, x, xmm0); code.vpcmpgtq(xmm0, tmp, xmm0); code.pblendvb(x, y); ctx.reg_alloc.DefineValue(inst, x); return; } EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), [](auto x, auto y) { return std::max(x, y); }); }); } void EmitX64::EmitVectorMinS8(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pminsb); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm tmp_b = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp_b, b); code.pcmpgtb(tmp_b, a); code.pand(a, tmp_b); code.pandn(tmp_b, b); code.por(a, tmp_b); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorMinS16(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pminsw); } void EmitX64::EmitVectorMinS32(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pminsd); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm tmp_b = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp_b, b); code.pcmpgtd(tmp_b, a); code.pand(a, tmp_b); code.pandn(tmp_b, b); code.por(a, tmp_b); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorMinS64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { EmitAVXVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::vpminsq); return; } if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseScratchXmm(args[1]); code.vpcmpgtq(xmm0, y, x); code.pblendvb(y, x); ctx.reg_alloc.DefineValue(inst, y); return; } EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b){ std::transform(a.begin(), a.end(), b.begin(), result.begin(), [](auto x, auto y) { return std::min(x, y); }); }); } void EmitX64::EmitVectorMinU8(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pminub); } void EmitX64::EmitVectorMinU16(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pminuw); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp_b = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp_b, b); code.psubusw(tmp_b, a); code.psubw(b, tmp_b); ctx.reg_alloc.DefineValue(inst, b); } void EmitX64::EmitVectorMinU32(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pminud); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm sint_max_plus_one = ctx.reg_alloc.ScratchXmm(); code.movdqa(sint_max_plus_one, code.MConst(xword, 0x8000000080000000, 0x8000000080000000)); const Xbyak::Xmm tmp_a = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp_a, a); code.psubd(tmp_a, sint_max_plus_one); const Xbyak::Xmm tmp_b = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp_b, b); code.psubd(tmp_b, sint_max_plus_one); code.pcmpgtd(tmp_b, tmp_a); code.pand(a, tmp_b); code.pandn(tmp_b, b); code.por(a, tmp_b); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorMinU64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { EmitAVXVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::vpminuq); return; } if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.vmovdqa(xmm0, code.MConst(xword, 0x8000000000000000, 0x8000000000000000)); code.vpsubq(tmp, y, xmm0); code.vpsubq(xmm0, x, xmm0); code.vpcmpgtq(xmm0, tmp, xmm0); code.pblendvb(y, x); ctx.reg_alloc.DefineValue(inst, y); return; } EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b){ std::transform(a.begin(), a.end(), b.begin(), result.begin(), [](auto x, auto y) { return std::min(x, y); }); }); } void EmitX64::EmitVectorMultiply8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); Xbyak::Xmm tmp_a = ctx.reg_alloc.ScratchXmm(); Xbyak::Xmm tmp_b = ctx.reg_alloc.ScratchXmm(); // TODO: Optimize code.movdqa(tmp_a, a); code.movdqa(tmp_b, b); code.pmullw(a, b); code.psrlw(tmp_a, 8); code.psrlw(tmp_b, 8); code.pmullw(tmp_a, tmp_b); code.pand(a, code.MConst(xword, 0x00FF00FF00FF00FF, 0x00FF00FF00FF00FF)); code.psllw(tmp_a, 8); code.por(a, tmp_a); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorMultiply16(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pmullw); } void EmitX64::EmitVectorMultiply32(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pmulld); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, a); code.psrlq(a, 32); code.pmuludq(tmp, b); code.psrlq(b, 32); code.pmuludq(a, b); code.pshufd(tmp, tmp, 0b00001000); code.pshufd(b, a, 0b00001000); code.punpckldq(tmp, b); ctx.reg_alloc.DefineValue(inst, tmp); } void EmitX64::EmitVectorMultiply64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512DQ) && code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { EmitAVXVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::vpmullq); return; } auto args = ctx.reg_alloc.GetArgumentInfo(inst); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); Xbyak::Reg64 tmp1 = ctx.reg_alloc.ScratchGpr(); Xbyak::Reg64 tmp2 = ctx.reg_alloc.ScratchGpr(); code.movq(tmp1, a); code.movq(tmp2, b); code.imul(tmp2, tmp1); code.pextrq(tmp1, a, 1); code.movq(a, tmp2); code.pextrq(tmp2, b, 1); code.imul(tmp1, tmp2); code.pinsrq(a, tmp1, 1); ctx.reg_alloc.DefineValue(inst, a); return; } const Xbyak::Xmm a = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp3 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp1, a); code.movdqa(tmp2, a); code.movdqa(tmp3, b); code.psrlq(tmp1, 32); code.psrlq(tmp3, 32); code.pmuludq(tmp2, b); code.pmuludq(tmp3, a); code.pmuludq(b, tmp1); code.paddq(b, tmp3); code.psllq(b, 32); code.paddq(tmp2, b); ctx.reg_alloc.DefineValue(inst, tmp2); } void EmitX64::EmitVectorNarrow16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL) && code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512BW)) { const Xbyak::Xmm a = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); code.vpmovwb(result, a); ctx.reg_alloc.DefineValue(inst, result); return; } const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm zeros = ctx.reg_alloc.ScratchXmm(); code.pxor(zeros, zeros); code.pand(a, code.MConst(xword, 0x00FF00FF00FF00FF, 0x00FF00FF00FF00FF)); code.packuswb(a, zeros); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorNarrow32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm zeros = ctx.reg_alloc.ScratchXmm(); // TODO: AVX512F implementation code.pxor(zeros, zeros); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pblendw(a, zeros, 0b10101010); code.packusdw(a, zeros); } else { code.pslld(a, 16); code.psrad(a, 16); code.packssdw(a, zeros); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorNarrow64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm zeros = ctx.reg_alloc.ScratchXmm(); // TODO: AVX512F implementation code.pxor(zeros, zeros); code.shufps(a, zeros, 0b00001000); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorNot(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm xmm_b = ctx.reg_alloc.ScratchXmm(); code.pcmpeqw(xmm_b, xmm_b); code.pxor(xmm_a, xmm_b); ctx.reg_alloc.DefineValue(inst, xmm_a); } void EmitX64::EmitVectorOr(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::por); } void EmitX64::EmitVectorPairedAddLower8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.punpcklqdq(xmm_a, xmm_b); code.movdqa(tmp, xmm_a); code.psllw(xmm_a, 8); code.paddw(xmm_a, tmp); code.pxor(tmp, tmp); code.psrlw(xmm_a, 8); code.packuswb(xmm_a, tmp); ctx.reg_alloc.DefineValue(inst, xmm_a); } void EmitX64::EmitVectorPairedAddLower16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.punpcklqdq(xmm_a, xmm_b); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { code.pxor(tmp, tmp); code.phaddw(xmm_a, tmp); } else { code.movdqa(tmp, xmm_a); code.pslld(xmm_a, 16); code.paddd(xmm_a, tmp); code.pxor(tmp, tmp); code.psrad(xmm_a, 16); code.packssdw(xmm_a, tmp); // Note: packusdw is SSE4.1, hence the arithmetic shift above. } ctx.reg_alloc.DefineValue(inst, xmm_a); } void EmitX64::EmitVectorPairedAddLower32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm xmm_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm xmm_b = ctx.reg_alloc.UseXmm(args[1]); Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.punpcklqdq(xmm_a, xmm_b); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { code.pxor(tmp, tmp); code.phaddd(xmm_a, tmp); } else { code.movdqa(tmp, xmm_a); code.psllq(xmm_a, 32); code.paddq(xmm_a, tmp); code.psrlq(xmm_a, 32); code.pshufd(xmm_a, xmm_a, 0b11011000); } ctx.reg_alloc.DefineValue(inst, xmm_a); } void EmitX64::EmitVectorPairedAdd8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); Xbyak::Xmm d = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.movdqa(d, b); code.psllw(a, 8); code.psllw(b, 8); code.paddw(a, c); code.paddw(b, d); code.psrlw(a, 8); code.psrlw(b, 8); code.packuswb(a, b); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorPairedAdd16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); code.phaddw(a, b); ctx.reg_alloc.DefineValue(inst, a); } else { Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); Xbyak::Xmm d = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.movdqa(d, b); code.pslld(a, 16); code.pslld(b, 16); code.paddd(a, c); code.paddd(b, d); code.psrad(a, 16); code.psrad(b, 16); code.packssdw(a, b); ctx.reg_alloc.DefineValue(inst, a); } } void EmitX64::EmitVectorPairedAdd32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); code.phaddd(a, b); ctx.reg_alloc.DefineValue(inst, a); } else { Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); Xbyak::Xmm d = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.movdqa(d, b); code.psllq(a, 32); code.psllq(b, 32); code.paddq(a, c); code.paddq(b, d); code.shufps(a, b, 0b11011101); ctx.reg_alloc.DefineValue(inst, a); } } void EmitX64::EmitVectorPairedAdd64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm b = ctx.reg_alloc.UseXmm(args[1]); Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.punpcklqdq(a, b); code.punpckhqdq(c, b); code.paddq(a, c); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorPairedAddSignedWiden8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.psllw(a, 8); code.psraw(c, 8); code.psraw(a, 8); code.paddw(a, c); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorPairedAddSignedWiden16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.pslld(a, 16); code.psrad(c, 16); code.psrad(a, 16); code.paddd(a, c); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorPairedAddSignedWiden32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { code.vpsraq(c, a, 32); code.vpsllq(a, a, 32); code.vpsraq(a, a, 32); code.vpaddq(a, a, c); } else { const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.psllq(a, 32); code.movdqa(tmp1, code.MConst(xword, 0x80000000'00000000, 0x80000000'00000000)); code.movdqa(tmp2, tmp1); code.pand(tmp1, a); code.pand(tmp2, c); code.psrlq(a, 32); code.psrlq(c, 32); code.psrad(tmp1, 31); code.psrad(tmp2, 31); code.por(a, tmp1); code.por(c, tmp2); code.paddq(a, c); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorPairedAddUnsignedWiden8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.psllw(a, 8); code.psrlw(c, 8); code.psrlw(a, 8); code.paddw(a, c); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorPairedAddUnsignedWiden16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.pslld(a, 16); code.psrld(c, 16); code.psrld(a, 16); code.paddd(a, c); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorPairedAddUnsignedWiden32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm c = ctx.reg_alloc.ScratchXmm(); code.movdqa(c, a); code.psllq(a, 32); code.psrlq(c, 32); code.psrlq(a, 32); code.paddq(a, c); ctx.reg_alloc.DefineValue(inst, a); } template static void PairedOperation(VectorArray& result, const VectorArray& x, const VectorArray& y, Function fn) { const size_t range = x.size() / 2; for (size_t i = 0; i < range; i++) { result[i] = fn(x[2 * i], x[2 * i + 1]); } for (size_t i = 0; i < range; i++) { result[range + i] = fn(y[2 * i], y[2 * i + 1]); } } template static void PairedMax(VectorArray& result, const VectorArray& x, const VectorArray& y) { PairedOperation(result, x, y, [](auto a, auto b) { return std::max(a, b); }); } template static void PairedMin(VectorArray& result, const VectorArray& x, const VectorArray& y) { PairedOperation(result, x, y, [](auto a, auto b) { return std::min(a, b); }); } void EmitX64::EmitVectorPairedMaxS8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { PairedMax(result, a, b); }); } void EmitX64::EmitVectorPairedMaxS16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { PairedMax(result, a, b); }); } void EmitX64::EmitVectorPairedMaxS32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, x); code.shufps(tmp, y, 0b10001000); code.shufps(x, y, 0b11011101); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pmaxsd(x, tmp); ctx.reg_alloc.DefineValue(inst, x); } else { const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp2, tmp); code.pcmpgtd(tmp2, x); code.pand(tmp, tmp2); code.pandn(tmp2, x); code.por(tmp2, tmp); ctx.reg_alloc.DefineValue(inst, tmp2); } } void EmitX64::EmitVectorPairedMaxU8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { PairedMax(result, a, b); }); } void EmitX64::EmitVectorPairedMaxU16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { PairedMax(result, a, b); }); } void EmitX64::EmitVectorPairedMaxU32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp1, x); code.shufps(tmp1, y, 0b10001000); code.shufps(x, y, 0b11011101); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pmaxud(x, tmp1); ctx.reg_alloc.DefineValue(inst, x); } else { const Xbyak::Xmm tmp3 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp3, code.MConst(xword, 0x8000000080000000, 0x8000000080000000)); const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp2, x); code.pxor(tmp2, tmp3); code.pxor(tmp3, tmp1); code.pcmpgtd(tmp3, tmp2); code.pand(tmp1, tmp3); code.pandn(tmp3, x); code.por(tmp1, tmp3); ctx.reg_alloc.DefineValue(inst, tmp1); } } void EmitX64::EmitVectorPairedMinS8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { PairedMin(result, a, b); }); } void EmitX64::EmitVectorPairedMinS16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { PairedMin(result, a, b); }); } void EmitX64::EmitVectorPairedMinS32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, x); code.shufps(tmp, y, 0b10001000); code.shufps(x, y, 0b11011101); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pminsd(x, tmp); ctx.reg_alloc.DefineValue(inst, x); } else { const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); code.movaps(tmp2, x); code.pcmpgtd(tmp2, tmp); code.pand(tmp, tmp2); code.pandn(tmp2, x); code.por(tmp2, tmp); ctx.reg_alloc.DefineValue(inst, tmp2); } } void EmitX64::EmitVectorPairedMinU8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { PairedMin(result, a, b); }); } void EmitX64::EmitVectorPairedMinU16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { PairedMin(result, a, b); }); } void EmitX64::EmitVectorPairedMinU32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp1, x); code.shufps(tmp1, y, 0b10001000); code.shufps(x, y, 0b11011101); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pminud(x, tmp1); ctx.reg_alloc.DefineValue(inst, x); } else { const Xbyak::Xmm tmp3 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp3, code.MConst(xword, 0x8000000080000000, 0x8000000080000000)); const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp2, tmp1); code.pxor(tmp2, tmp3); code.pxor(tmp3, x); code.pcmpgtd(tmp3, tmp2); code.pand(tmp1, tmp3); code.pandn(tmp3, x); code.por(tmp1, tmp3); ctx.reg_alloc.DefineValue(inst, tmp1); } } template static D PolynomialMultiply(T lhs, T rhs) { constexpr size_t bit_size = Common::BitSize(); const std::bitset operand(lhs); D res = 0; for (size_t i = 0; i < bit_size; i++) { if (operand[i]) { res ^= rhs << i; } } return res; } void EmitX64::EmitVectorPolynomialMultiply8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { std::transform(a.begin(), a.end(), b.begin(), result.begin(), PolynomialMultiply); }); } void EmitX64::EmitVectorPolynomialMultiplyLong8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { for (size_t i = 0; i < result.size(); i++) { result[i] = PolynomialMultiply(a[i], b[i]); } }); } void EmitX64::EmitVectorPolynomialMultiplyLong64(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a, const VectorArray& b) { const auto handle_high_bits = [](u64 lhs, u64 rhs) { constexpr size_t bit_size = Common::BitSize(); u64 result = 0; for (size_t i = 1; i < bit_size; i++) { if (Common::Bit(i, lhs)) { result ^= rhs >> (bit_size - i); } } return result; }; result[0] = PolynomialMultiply(a[0], b[0]); result[1] = handle_high_bits(a[0], b[0]); }); } void EmitX64::EmitVectorPopulationCount(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512_BITALG)) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm data = ctx.reg_alloc.UseScratchXmm(args[0]); code.vpopcntb(data, data); ctx.reg_alloc.DefineValue(inst, data); return; } if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm low_a = ctx.reg_alloc.UseScratchXmm(args[0]); Xbyak::Xmm high_a = ctx.reg_alloc.ScratchXmm(); Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); code.movdqa(high_a, low_a); code.psrlw(high_a, 4); code.movdqa(tmp1, code.MConst(xword, 0x0F0F0F0F0F0F0F0F, 0x0F0F0F0F0F0F0F0F)); code.pand(high_a, tmp1); // High nibbles code.pand(low_a, tmp1); // Low nibbles code.movdqa(tmp1, code.MConst(xword, 0x0302020102010100, 0x0403030203020201)); code.movdqa(tmp2, tmp1); code.pshufb(tmp1, low_a); code.pshufb(tmp2, high_a); code.paddb(tmp1, tmp2); ctx.reg_alloc.DefineValue(inst, tmp1); return; } EmitOneArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { std::transform(a.begin(), a.end(), result.begin(), [](u8 val) { return static_cast(Common::BitCount(val)); }); }); } void EmitX64::EmitVectorReverseBits(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm data = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm high_nibble_reg = ctx.reg_alloc.ScratchXmm(); code.movdqa(high_nibble_reg, code.MConst(xword, 0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0)); code.pand(high_nibble_reg, data); code.pxor(data, high_nibble_reg); code.psrld(high_nibble_reg, 4); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3)) { // High lookup const Xbyak::Xmm high_reversed_reg = ctx.reg_alloc.ScratchXmm(); code.movdqa(high_reversed_reg, code.MConst(xword, 0xE060A020C0408000, 0xF070B030D0509010)); code.pshufb(high_reversed_reg, data); // Low lookup (low nibble equivalent of the above) code.movdqa(data, code.MConst(xword, 0x0E060A020C040800, 0x0F070B030D050901)); code.pshufb(data, high_nibble_reg); code.por(data, high_reversed_reg); } else { code.pslld(data, 4); code.por(data, high_nibble_reg); code.movdqa(high_nibble_reg, code.MConst(xword, 0xCCCCCCCCCCCCCCCC, 0xCCCCCCCCCCCCCCCC)); code.pand(high_nibble_reg, data); code.pxor(data, high_nibble_reg); code.psrld(high_nibble_reg, 2); code.pslld(data, 2); code.por(data, high_nibble_reg); code.movdqa(high_nibble_reg, code.MConst(xword, 0xAAAAAAAAAAAAAAAA, 0xAAAAAAAAAAAAAAAA)); code.pand(high_nibble_reg, data); code.pxor(data, high_nibble_reg); code.psrld(high_nibble_reg, 1); code.paddd(data, data); code.por(data, high_nibble_reg); } ctx.reg_alloc.DefineValue(inst, data); } static void EmitVectorRoundingHalvingAddSigned(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); switch (esize) { case 8: { const Xbyak::Xmm vec_128 = ctx.reg_alloc.ScratchXmm(); code.movdqa(vec_128, code.MConst(xword, 0x8080808080808080, 0x8080808080808080)); code.paddb(a, vec_128); code.paddb(b, vec_128); code.pavgb(a, b); code.paddb(a, vec_128); break; } case 16: { const Xbyak::Xmm vec_32768 = ctx.reg_alloc.ScratchXmm(); code.movdqa(vec_32768, code.MConst(xword, 0x8000800080008000, 0x8000800080008000)); code.paddw(a, vec_32768); code.paddw(b, vec_32768); code.pavgw(a, b); code.paddw(a, vec_32768); break; } case 32: { const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp1, a); code.por(a, b); code.psrad(tmp1, 1); code.psrad(b, 1); code.pslld(a, 31); code.paddd(b, tmp1); code.psrld(a, 31); code.paddd(a, b); break; } } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorRoundingHalvingAddS8(EmitContext& ctx, IR::Inst* inst) { EmitVectorRoundingHalvingAddSigned(8, ctx, inst, code); } void EmitX64::EmitVectorRoundingHalvingAddS16(EmitContext& ctx, IR::Inst* inst) { EmitVectorRoundingHalvingAddSigned(16, ctx, inst, code); } void EmitX64::EmitVectorRoundingHalvingAddS32(EmitContext& ctx, IR::Inst* inst) { EmitVectorRoundingHalvingAddSigned(32, ctx, inst, code); } static void EmitVectorRoundingHalvingAddUnsigned(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { switch (esize) { case 8: EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pavgb); return; case 16: EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::pavgw); return; case 32: { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm b = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp1, a); code.por(a, b); code.psrld(tmp1, 1); code.psrld(b, 1); code.pslld(a, 31); code.paddd(b, tmp1); code.psrld(a, 31); code.paddd(a, b); ctx.reg_alloc.DefineValue(inst, a); break; } } } void EmitX64::EmitVectorRoundingHalvingAddU8(EmitContext& ctx, IR::Inst* inst) { EmitVectorRoundingHalvingAddUnsigned(8, ctx, inst, code); } void EmitX64::EmitVectorRoundingHalvingAddU16(EmitContext& ctx, IR::Inst* inst) { EmitVectorRoundingHalvingAddUnsigned(16, ctx, inst, code); } void EmitX64::EmitVectorRoundingHalvingAddU32(EmitContext& ctx, IR::Inst* inst) { EmitVectorRoundingHalvingAddUnsigned(32, ctx, inst, code); } template static void RoundingShiftLeft(VectorArray& out, const VectorArray& lhs, const VectorArray& rhs) { using signed_type = std::make_signed_t; using unsigned_type = std::make_unsigned_t; constexpr auto bit_size = static_cast(Common::BitSize()); for (size_t i = 0; i < out.size(); i++) { const s64 extended_shift = Common::SignExtend<8>(rhs[i] & 0xFF); if (extended_shift >= 0) { if (extended_shift >= bit_size) { out[i] = 0; } else { out[i] = static_cast(static_cast(lhs[i]) << extended_shift); } } else { if ((std::is_unsigned_v && extended_shift < -bit_size) || (std::is_signed_v && extended_shift <= -bit_size)) { out[i] = 0; } else { const s64 shift_value = -extended_shift - 1; const T shifted = (lhs[i] & (static_cast(1) << shift_value)) >> shift_value; if (extended_shift == -bit_size) { out[i] = shifted; } else { out[i] = (lhs[i] >> -extended_shift) + shifted; } } } } } void EmitX64::EmitVectorRoundingShiftLeftS8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { RoundingShiftLeft(result, lhs, rhs); }); } void EmitX64::EmitVectorRoundingShiftLeftS16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { RoundingShiftLeft(result, lhs, rhs); }); } void EmitX64::EmitVectorRoundingShiftLeftS32(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { RoundingShiftLeft(result, lhs, rhs); }); } void EmitX64::EmitVectorRoundingShiftLeftS64(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { RoundingShiftLeft(result, lhs, rhs); }); } void EmitX64::EmitVectorRoundingShiftLeftU8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { RoundingShiftLeft(result, lhs, rhs); }); } void EmitX64::EmitVectorRoundingShiftLeftU16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { RoundingShiftLeft(result, lhs, rhs); }); } void EmitX64::EmitVectorRoundingShiftLeftU32(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { RoundingShiftLeft(result, lhs, rhs); }); } void EmitX64::EmitVectorRoundingShiftLeftU64(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { RoundingShiftLeft(result, lhs, rhs); }); } static void VectorShuffleImpl(BlockOfCode& code, EmitContext& ctx, IR::Inst* inst, void (Xbyak::CodeGenerator::*fn)(const Xbyak::Mmx&, const Xbyak::Operand&, u8)) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm operand = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); const u8 mask = args[1].GetImmediateU8(); (code.*fn)(result, operand, mask); ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorShuffleHighHalfwords(EmitContext& ctx, IR::Inst* inst) { VectorShuffleImpl(code, ctx, inst, &Xbyak::CodeGenerator::pshufhw); } void EmitX64::EmitVectorShuffleLowHalfwords(EmitContext& ctx, IR::Inst* inst) { VectorShuffleImpl(code, ctx, inst, &Xbyak::CodeGenerator::pshuflw); } void EmitX64::EmitVectorShuffleWords(EmitContext& ctx, IR::Inst* inst) { VectorShuffleImpl(code, ctx, inst, &Xbyak::CodeGenerator::pshufd); } void EmitX64::EmitVectorSignExtend8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); code.pmovsxbw(a, a); ctx.reg_alloc.DefineValue(inst, a); } else { const Xbyak::Xmm a = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); code.pxor(result, result); code.punpcklbw(result, a); code.psraw(result, 8); ctx.reg_alloc.DefineValue(inst, result); } } void EmitX64::EmitVectorSignExtend16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); code.pmovsxwd(a, a); ctx.reg_alloc.DefineValue(inst, a); } else { const Xbyak::Xmm a = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); code.pxor(result, result); code.punpcklwd(result, a); code.psrad(result, 16); ctx.reg_alloc.DefineValue(inst, result); } } void EmitX64::EmitVectorSignExtend32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pmovsxdq(a, a); } else { const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movaps(tmp, a); code.psrad(tmp, 31); code.punpckldq(a, tmp); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorSignExtend64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm data = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Reg64 gpr_tmp = ctx.reg_alloc.ScratchGpr(); code.movq(gpr_tmp, data); code.sar(gpr_tmp, 63); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pinsrq(data, gpr_tmp, 1); } else { const Xbyak::Xmm xmm_tmp = ctx.reg_alloc.ScratchXmm(); code.movq(xmm_tmp, gpr_tmp); code.punpcklqdq(data, xmm_tmp); } ctx.reg_alloc.DefineValue(inst, data); } static void EmitVectorSignedAbsoluteDifference(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); const Xbyak::Xmm mask = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); code.movdqa(mask, x); code.movdqa(tmp1, y); switch (esize) { case 8: code.pcmpgtb(mask, y); code.psubb(tmp1, x); code.psubb(x, y); break; case 16: code.pcmpgtw(mask, y); code.psubw(tmp1, x); code.psubw(x, y); break; case 32: code.pcmpgtd(mask, y); code.psubd(tmp1, x); code.psubd(x, y); break; } code.movdqa(tmp2, mask); code.pand(x, mask); code.pandn(tmp2, tmp1); code.por(x, tmp2); ctx.reg_alloc.DefineValue(inst, x); } void EmitX64::EmitVectorSignedAbsoluteDifference8(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedAbsoluteDifference(8, ctx, inst, code); } void EmitX64::EmitVectorSignedAbsoluteDifference16(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedAbsoluteDifference(16, ctx, inst, code); } void EmitX64::EmitVectorSignedAbsoluteDifference32(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedAbsoluteDifference(32, ctx, inst, code); } static void EmitVectorSignedSaturatedAbs(size_t esize, BlockOfCode& code, EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm data = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm data_test = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm sign = ctx.reg_alloc.ScratchXmm(); const Xbyak::Address mask = [esize, &code] { switch (esize) { case 8: return code.MConst(xword, 0x8080808080808080, 0x8080808080808080); case 16: return code.MConst(xword, 0x8000800080008000, 0x8000800080008000); case 32: return code.MConst(xword, 0x8000000080000000, 0x8000000080000000); case 64: return code.MConst(xword, 0x8000000000000000, 0x8000000000000000); default: UNREACHABLE(); return Xbyak::Address{0}; } }(); const auto vector_equality = [esize, &code](const Xbyak::Xmm& x, const Xbyak::Xmm& y) { switch (esize) { case 8: code.pcmpeqb(x, y); break; case 16: code.pcmpeqw(x, y); break; case 32: code.pcmpeqd(x, y); break; case 64: code.pcmpeqq(x, y); break; } }; // Keep a copy of the initial data for determining whether or not // to set the Q flag code.movdqa(data_test, data); switch (esize) { case 8: VectorAbs8(code, ctx, data); break; case 16: VectorAbs16(code, ctx, data); break; case 32: VectorAbs32(code, ctx, data); break; case 64: VectorAbs64(code, ctx, data); break; } code.movdqa(sign, mask); vector_equality(sign, data); code.pxor(data, sign); // Check if the initial data contained any elements with the value 0x80. // If any exist, then the Q flag needs to be set. const Xbyak::Reg32 bit = ctx.reg_alloc.ScratchGpr().cvt32(); code.movdqa(sign, mask); vector_equality(data_test, sign); code.pmovmskb(bit, data_test); code.or_(code.dword[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], bit); ctx.reg_alloc.DefineValue(inst, data); } void EmitX64::EmitVectorSignedSaturatedAbs8(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedAbs(8, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedAbs16(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedAbs(16, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedAbs32(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedAbs(32, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedAbs64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorSignedSaturatedAbs(64, code, ctx, inst); return; } EmitOneArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& data) { bool qc_flag = false; for (size_t i = 0; i < result.size(); i++) { if (static_cast(data[i]) == 0x8000000000000000) { result[i] = 0x7FFFFFFFFFFFFFFF; qc_flag = true; } else { result[i] = std::abs(data[i]); } } return qc_flag; }); } // Simple generic case for 8, 16, and 32-bit values. 64-bit values // will need to be special-cased as we can't simply use a larger integral size. template static bool EmitSignedSaturatedAccumulateUnsigned(VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { static_assert(std::is_signed_v, "T must be signed."); static_assert(sizeof(T) < 64, "T must be less than 64 bits in size."); bool qc_flag = false; for (size_t i = 0; i < result.size(); i++) { // We treat lhs' members as unsigned, so cast to unsigned before signed to inhibit sign-extension. // We use the unsigned equivalent of T, as we want zero-extension to occur, rather than a plain move. const s64 x = static_cast(static_cast>(lhs[i])); const s64 y = rhs[i]; const s64 sum = x + y; if (sum > std::numeric_limits::max()) { result[i] = std::numeric_limits::max(); qc_flag = true; } else if (sum < std::numeric_limits::min()) { result[i] = std::numeric_limits::min(); qc_flag = true; } else { result[i] = static_cast(sum); } } return qc_flag; } void EmitX64::EmitVectorSignedSaturatedAccumulateUnsigned8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallbackWithSaturation(code, ctx, inst, EmitSignedSaturatedAccumulateUnsigned); } void EmitX64::EmitVectorSignedSaturatedAccumulateUnsigned16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallbackWithSaturation(code, ctx, inst, EmitSignedSaturatedAccumulateUnsigned); } void EmitX64::EmitVectorSignedSaturatedAccumulateUnsigned32(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallbackWithSaturation(code, ctx, inst, EmitSignedSaturatedAccumulateUnsigned); } void EmitX64::EmitVectorSignedSaturatedAccumulateUnsigned64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); code.movdqa(xmm0, y); ctx.reg_alloc.Release(y); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { code.vpaddq(result, x, xmm0); } else { code.movdqa(result, x); code.paddq(result, xmm0); } if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX512VL)) { // xmm0 = majority(~y, x, res) // y x res xmm0 // 0 0 0 0 // 0 0 1 1 // 0 1 0 1 // 0 1 1 1 // 1 0 0 0 // 1 0 1 0 // 1 1 0 0 // 1 1 1 1 code.vpternlogd(xmm0, x, result, 0b10001110); code.vpsraq(xmm0, xmm0, 63); code.movdqa(tmp, xmm0); } else if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { code.vpor(tmp, x, result); code.pand(x, result); code.pblendvb(tmp, x); code.psrad(tmp, 31); code.pshufd(tmp, tmp, 0b11110101); code.movdqa(xmm0, tmp); } else { code.movdqa(tmp, x); code.por(tmp, result); code.pand(x, result); code.pand(x, xmm0); code.pandn(xmm0, tmp); code.por(xmm0, x); code.psrad(xmm0, 31); code.pshufd(xmm0, xmm0, 0b11110101); code.movdqa(tmp, xmm0); } const Xbyak::Reg32 mask = ctx.reg_alloc.ScratchGpr().cvt32(); code.pmovmskb(mask, xmm0); code.or_(code.dword[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], mask); code.psrlq(tmp, 1); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pblendvb(result, tmp); } else { code.pandn(xmm0, result); code.por(xmm0, tmp); code.movdqa(result, xmm0); } ctx.reg_alloc.DefineValue(inst, result); } void EmitX64::EmitVectorSignedSaturatedDoublingMultiplyReturnHigh16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp, x); code.pmulhw(tmp, y); code.paddw(tmp, tmp); code.pmullw(y, x); code.psrlw(y, 15); code.por(y, tmp); code.movdqa(x, code.MConst(xword, 0x8000800080008000, 0x8000800080008000)); code.pcmpeqw(x, y); code.movdqa(tmp, x); code.pxor(x, y); // Check if any saturation occurred (i.e. if any halfwords in x were // 0x8000 before saturating const Xbyak::Reg32 mask = ctx.reg_alloc.ScratchGpr().cvt32(); code.pmovmskb(mask, tmp); code.or_(code.dword[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], mask); ctx.reg_alloc.DefineValue(inst, x); } void EmitX64::EmitVectorSignedSaturatedDoublingMultiplyReturnHigh32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseScratchXmm(args[1]); const Xbyak::Xmm tmp1 = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp2 = ctx.reg_alloc.ScratchXmm(); code.movdqa(tmp1, x); code.punpckldq(tmp1, y); code.movdqa(tmp2, y); code.punpckldq(tmp2, x); code.pmuldq(tmp2, tmp1); code.paddq(tmp2, tmp2); code.movdqa(tmp1, x); code.punpckhdq(tmp1, y); code.punpckhdq(y, x); code.pmuldq(y, tmp1); code.paddq(y, y); code.pshufd(tmp1, tmp2, 0b11101101); code.pshufd(x, y, 0b11101101); code.punpcklqdq(tmp1, x); code.movdqa(x, code.MConst(xword, 0x8000000080000000, 0x8000000080000000)); code.pcmpeqd(x, tmp1); code.movdqa(tmp2, x); code.pxor(x, tmp1); // Check if any saturation occurred (i.e. if any words in x were // 0x80000000 before saturating const Xbyak::Reg32 mask = ctx.reg_alloc.ScratchGpr().cvt32(); code.pmovmskb(mask, tmp2); code.or_(code.dword[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], mask); ctx.reg_alloc.DefineValue(inst, x); } static void EmitVectorSignedSaturatedNarrowToSigned(size_t original_esize, BlockOfCode& code, EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm src = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm dest = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm reconstructed = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm sign = ctx.reg_alloc.ScratchXmm(); code.movdqa(dest, src); switch (original_esize) { case 16: code.packsswb(dest, dest); code.movdqa(sign, src); code.psraw(sign, 15); code.packsswb(sign, sign); code.movdqa(reconstructed, dest); code.punpcklbw(reconstructed, sign); break; case 32: code.packssdw(dest, dest); code.movdqa(reconstructed, dest); code.movdqa(sign, dest); code.psraw(sign, 15); code.punpcklwd(reconstructed, sign); break; default: UNREACHABLE(); break; } const Xbyak::Reg32 bit = ctx.reg_alloc.ScratchGpr().cvt32(); code.pcmpeqd(reconstructed, src); code.movmskps(bit, reconstructed); code.xor_(bit, 0b1111); code.or_(code.dword[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], bit); ctx.reg_alloc.DefineValue(inst, dest); } void EmitX64::EmitVectorSignedSaturatedNarrowToSigned16(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedNarrowToSigned(16, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedNarrowToSigned32(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedNarrowToSigned(32, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedNarrowToSigned64(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { bool qc_flag = false; for (size_t i = 0; i < a.size(); ++i) { const s64 saturated = std::clamp(a[i], -s64(0x80000000), s64(0x7FFFFFFF)); result[i] = static_cast(saturated); qc_flag |= saturated != a[i]; } return qc_flag; }); } static void EmitVectorSignedSaturatedNarrowToUnsigned(size_t original_esize, BlockOfCode& code, EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm src = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm dest = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm reconstructed = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm zero = ctx.reg_alloc.ScratchXmm(); code.movdqa(dest, src); code.pxor(zero, zero); switch (original_esize) { case 16: code.packuswb(dest, dest); code.movdqa(reconstructed, dest); code.punpcklbw(reconstructed, zero); break; case 32: ASSERT(code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)); code.packusdw(dest, dest); // SSE4.1 code.movdqa(reconstructed, dest); code.punpcklwd(reconstructed, zero); break; default: UNREACHABLE(); break; } const Xbyak::Reg32 bit = ctx.reg_alloc.ScratchGpr().cvt32(); code.pcmpeqd(reconstructed, src); code.movmskps(bit, reconstructed); code.xor_(bit, 0b1111); code.or_(code.dword[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], bit); ctx.reg_alloc.DefineValue(inst, dest); } void EmitX64::EmitVectorSignedSaturatedNarrowToUnsigned16(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedNarrowToUnsigned(16, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedNarrowToUnsigned32(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorSignedSaturatedNarrowToUnsigned(32, code, ctx, inst); return; } EmitOneArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { bool qc_flag = false; for (size_t i = 0; i < a.size(); ++i) { const s32 saturated = std::clamp(a[i], 0, 0xFFFF); result[i] = static_cast(saturated); qc_flag |= saturated != a[i]; } return qc_flag; }); } void EmitX64::EmitVectorSignedSaturatedNarrowToUnsigned64(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { bool qc_flag = false; for (size_t i = 0; i < a.size(); ++i) { const s64 saturated = std::clamp(a[i], 0, 0xFFFFFFFF); result[i] = static_cast(saturated); qc_flag |= saturated != a[i]; } return qc_flag; }); } static void EmitVectorSignedSaturatedNeg(size_t esize, BlockOfCode& code, EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm data = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm zero = ctx.reg_alloc.ScratchXmm(); const Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm(); const Xbyak::Address mask = [esize, &code] { switch (esize) { case 8: return code.MConst(xword, 0x8080808080808080, 0x8080808080808080); case 16: return code.MConst(xword, 0x8000800080008000, 0x8000800080008000); case 32: return code.MConst(xword, 0x8000000080000000, 0x8000000080000000); case 64: return code.MConst(xword, 0x8000000000000000, 0x8000000000000000); default: UNREACHABLE(); return Xbyak::Address{0}; } }(); const auto vector_equality = [esize, &code](const Xbyak::Xmm& x, const auto& y) { switch (esize) { case 8: code.pcmpeqb(x, y); break; case 16: code.pcmpeqw(x, y); break; case 32: code.pcmpeqd(x, y); break; case 64: code.pcmpeqq(x, y); break; } }; code.movdqa(tmp, data); vector_equality(tmp, mask); // Perform negation code.pxor(zero, zero); switch (esize) { case 8: code.psubsb(zero, data); break; case 16: code.psubsw(zero, data); break; case 32: code.psubd(zero, data); code.pxor(zero, tmp); break; case 64: code.psubq(zero, data); code.pxor(zero, tmp); break; } // Check if any elements matched the mask prior to performing saturation. If so, set the Q bit. const Xbyak::Reg32 bit = ctx.reg_alloc.ScratchGpr().cvt32(); code.pmovmskb(bit, tmp); code.or_(code.dword[code.r15 + code.GetJitStateInfo().offsetof_fpsr_qc], bit); ctx.reg_alloc.DefineValue(inst, zero); } void EmitX64::EmitVectorSignedSaturatedNeg8(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedNeg(8, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedNeg16(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedNeg(16, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedNeg32(EmitContext& ctx, IR::Inst* inst) { EmitVectorSignedSaturatedNeg(32, code, ctx, inst); } void EmitX64::EmitVectorSignedSaturatedNeg64(EmitContext& ctx, IR::Inst* inst) { if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { EmitVectorSignedSaturatedNeg(64, code, ctx, inst); return; } EmitOneArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& data) { bool qc_flag = false; for (size_t i = 0; i < result.size(); i++) { if (static_cast(data[i]) == 0x8000000000000000) { result[i] = 0x7FFFFFFFFFFFFFFF; qc_flag = true; } else { result[i] = -data[i]; } } return qc_flag; }); } void EmitX64::EmitVectorSub8(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::psubb); } void EmitX64::EmitVectorSub16(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::psubw); } void EmitX64::EmitVectorSub32(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::psubd); } void EmitX64::EmitVectorSub64(EmitContext& ctx, IR::Inst* inst) { EmitVectorOperation(code, ctx, inst, &Xbyak::CodeGenerator::psubq); } void EmitX64::EmitVectorTable(EmitContext&, IR::Inst* inst) { // Do nothing. We *want* to hold on to the refcount for our arguments, so VectorTableLookup can use our arguments. ASSERT_MSG(inst->UseCount() == 1, "Table cannot be used multiple times"); } void EmitX64::EmitVectorTableLookup(EmitContext& ctx, IR::Inst* inst) { ASSERT(inst->GetArg(1).GetInst()->GetOpcode() == IR::Opcode::VectorTable); auto args = ctx.reg_alloc.GetArgumentInfo(inst); auto table = ctx.reg_alloc.GetArgumentInfo(inst->GetArg(1).GetInst()); const size_t table_size = std::count_if(table.begin(), table.end(), [](const auto& elem){ return !elem.IsVoid(); }); const bool is_defaults_zero = !inst->GetArg(0).IsImmediate() && inst->GetArg(0).GetInst()->GetOpcode() == IR::Opcode::ZeroVector; // TODO: AVX512VL implementation when available (VPERMB / VPERMI2B / VPERMT2B) if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSSE3) && is_defaults_zero && table_size == 1) { const Xbyak::Xmm indicies = ctx.reg_alloc.UseScratchXmm(args[2]); const Xbyak::Xmm xmm_table0 = ctx.reg_alloc.UseScratchXmm(table[0]); code.paddusb(indicies, code.MConst(xword, 0x7070707070707070, 0x7070707070707070)); code.pshufb(xmm_table0, indicies); ctx.reg_alloc.DefineValue(inst, xmm_table0); return; } if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41) && table_size == 1) { const Xbyak::Xmm indicies = ctx.reg_alloc.UseXmm(args[2]); const Xbyak::Xmm defaults = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm xmm_table0 = ctx.reg_alloc.UseScratchXmm(table[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { code.vpaddusb(xmm0, indicies, code.MConst(xword, 0x7070707070707070, 0x7070707070707070)); } else { code.movaps(xmm0, indicies); code.paddusb(xmm0, code.MConst(xword, 0x7070707070707070, 0x7070707070707070)); } code.pshufb(xmm_table0, indicies); code.pblendvb(xmm_table0, defaults); ctx.reg_alloc.DefineValue(inst, xmm_table0); return; } if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41) && is_defaults_zero && table_size == 2) { const Xbyak::Xmm indicies = ctx.reg_alloc.UseScratchXmm(args[2]); const Xbyak::Xmm xmm_table0 = ctx.reg_alloc.UseScratchXmm(table[0]); const Xbyak::Xmm xmm_table1 = ctx.reg_alloc.UseScratchXmm(table[1]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { code.vpaddusb(xmm0, indicies, code.MConst(xword, 0x7070707070707070, 0x7070707070707070)); } else { code.movaps(xmm0, indicies); code.paddusb(xmm0, code.MConst(xword, 0x7070707070707070, 0x7070707070707070)); } code.paddusb(indicies, code.MConst(xword, 0x6060606060606060, 0x6060606060606060)); code.pshufb(xmm_table0, xmm0); code.pshufb(xmm_table1, indicies); code.pblendvb(xmm_table0, xmm_table1); ctx.reg_alloc.DefineValue(inst, xmm_table0); return; } if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { const Xbyak::Xmm indicies = ctx.reg_alloc.UseXmm(args[2]); const Xbyak::Xmm result = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm masked = ctx.reg_alloc.ScratchXmm(); code.movaps(masked, code.MConst(xword, 0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0)); code.pand(masked, indicies); for (size_t i = 0; i < table_size; ++i) { const Xbyak::Xmm xmm_table = ctx.reg_alloc.UseScratchXmm(table[i]); const u64 table_index = Common::Replicate(i * 16, 8); if (table_index == 0) { code.pxor(xmm0, xmm0); code.pcmpeqb(xmm0, masked); } else if (code.DoesCpuSupport(Xbyak::util::Cpu::tAVX)) { code.vpcmpeqb(xmm0, masked, code.MConst(xword, table_index, table_index)); } else { code.movaps(xmm0, code.MConst(xword, table_index, table_index)); code.pcmpeqb(xmm0, masked); } code.pshufb(xmm_table, indicies); code.pblendvb(result, xmm_table); ctx.reg_alloc.Release(xmm_table); } ctx.reg_alloc.DefineValue(inst, result); return; } const u32 stack_space = static_cast((table_size + 2) * 16); code.sub(rsp, stack_space + ABI_SHADOW_SPACE); for (size_t i = 0; i < table_size; ++i) { const Xbyak::Xmm table_value = ctx.reg_alloc.UseXmm(table[i]); code.movaps(xword[rsp + ABI_SHADOW_SPACE + i * 16], table_value); ctx.reg_alloc.Release(table_value); } const Xbyak::Xmm defaults = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm indicies = ctx.reg_alloc.UseXmm(args[2]); const Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm(); ctx.reg_alloc.EndOfAllocScope(); ctx.reg_alloc.HostCall(nullptr); code.lea(code.ABI_PARAM1, ptr[rsp + ABI_SHADOW_SPACE]); code.lea(code.ABI_PARAM2, ptr[rsp + ABI_SHADOW_SPACE + (table_size + 0) * 16]); code.lea(code.ABI_PARAM3, ptr[rsp + ABI_SHADOW_SPACE + (table_size + 1) * 16]); code.mov(code.ABI_PARAM4.cvt32(), table_size); code.movaps(xword[code.ABI_PARAM2], defaults); code.movaps(xword[code.ABI_PARAM3], indicies); code.CallFunction(static_cast*, VectorArray&, const VectorArray&, size_t)>( [](const VectorArray* table, VectorArray& result, const VectorArray& indicies, size_t table_size) { for (size_t i = 0; i < result.size(); ++i) { const size_t index = indicies[i] / table[0].size(); const size_t elem = indicies[i] % table[0].size(); if (index < table_size) { result[i] = table[index][elem]; } } } )); code.movaps(result, xword[rsp + ABI_SHADOW_SPACE + (table_size + 0) * 16]); code.add(rsp, stack_space + ABI_SHADOW_SPACE); ctx.reg_alloc.DefineValue(inst, result); } static void EmitVectorUnsignedAbsoluteDifference(size_t esize, EmitContext& ctx, IR::Inst* inst, BlockOfCode& code) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm temp = ctx.reg_alloc.ScratchXmm(); switch (esize) { case 8: { const Xbyak::Xmm x = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseScratchXmm(args[1]); code.movdqa(temp, x); code.psubusb(temp, y); code.psubusb(y, x); code.por(temp, y); break; } case 16: { const Xbyak::Xmm x = ctx.reg_alloc.UseXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseScratchXmm(args[1]); code.movdqa(temp, x); code.psubusw(temp, y); code.psubusw(y, x); code.por(temp, y); break; } case 32: if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseXmm(args[1]); code.movdqa(temp, x); code.pminud(x, y); code.pmaxud(temp, y); code.psubd(temp, x); } else { const Xbyak::Xmm x = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm y = ctx.reg_alloc.UseScratchXmm(args[1]); code.movdqa(temp, code.MConst(xword, 0x8000000080000000, 0x8000000080000000)); code.pxor(x, temp); code.pxor(y, temp); code.movdqa(temp, x); code.psubd(temp, y); code.pcmpgtd(y, x); code.psrld(y, 1); code.pxor(temp, y); code.psubd(temp, y); } break; } ctx.reg_alloc.DefineValue(inst, temp); } void EmitX64::EmitVectorUnsignedAbsoluteDifference8(EmitContext& ctx, IR::Inst* inst) { EmitVectorUnsignedAbsoluteDifference(8, ctx, inst, code); } void EmitX64::EmitVectorUnsignedAbsoluteDifference16(EmitContext& ctx, IR::Inst* inst) { EmitVectorUnsignedAbsoluteDifference(16, ctx, inst, code); } void EmitX64::EmitVectorUnsignedAbsoluteDifference32(EmitContext& ctx, IR::Inst* inst) { EmitVectorUnsignedAbsoluteDifference(32, ctx, inst, code); } void EmitX64::EmitVectorUnsignedRecipEstimate(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { for (size_t i = 0; i < result.size(); i++) { if ((a[i] & 0x80000000) == 0) { result[i] = 0xFFFFFFFF; continue; } const u32 input = Common::Bits<23, 31>(a[i]); const u32 estimate = Common::RecipEstimate(input); result[i] = (0b100000000 | estimate) << 23; } }); } void EmitX64::EmitVectorUnsignedRecipSqrtEstimate(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallback(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { for (size_t i = 0; i < result.size(); i++) { if ((a[i] & 0xC0000000) == 0) { result[i] = 0xFFFFFFFF; continue; } const u32 input = Common::Bits<23, 31>(a[i]); const u32 estimate = Common::RecipSqrtEstimate(input); result[i] = (0b100000000 | estimate) << 23; } }); } // Simple generic case for 8, 16, and 32-bit values. 64-bit values // will need to be special-cased as we can't simply use a larger integral size. template > bool EmitVectorUnsignedSaturatedAccumulateSigned(VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { static_assert(std::is_signed_v, "T must be signed."); static_assert(sizeof(T) < 64, "T must be less than 64 bits in size."); bool qc_flag = false; for (size_t i = 0; i < result.size(); i++) { // We treat rhs' members as unsigned, so cast to unsigned before signed to inhibit sign-extension. // We use the unsigned equivalent of T, as we want zero-extension to occur, rather than a plain move. const s64 x = s64{lhs[i]}; const s64 y = static_cast(static_cast>(rhs[i])); const s64 sum = x + y; if (sum > std::numeric_limits::max()) { result[i] = std::numeric_limits::max(); qc_flag = true; } else if (sum < 0) { result[i] = std::numeric_limits::min(); qc_flag = true; } else { result[i] = static_cast(sum); } } return qc_flag; } void EmitX64::EmitVectorUnsignedSaturatedAccumulateSigned8(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallbackWithSaturation(code, ctx, inst, EmitVectorUnsignedSaturatedAccumulateSigned); } void EmitX64::EmitVectorUnsignedSaturatedAccumulateSigned16(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallbackWithSaturation(code, ctx, inst, EmitVectorUnsignedSaturatedAccumulateSigned); } void EmitX64::EmitVectorUnsignedSaturatedAccumulateSigned32(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallbackWithSaturation(code, ctx, inst, EmitVectorUnsignedSaturatedAccumulateSigned); } void EmitX64::EmitVectorUnsignedSaturatedAccumulateSigned64(EmitContext& ctx, IR::Inst* inst) { EmitTwoArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& lhs, const VectorArray& rhs) { bool qc_flag = false; for (size_t i = 0; i < result.size(); i++) { const u64 x = lhs[i]; const u64 y = rhs[i]; const u64 res = x + y; // Check sign bits to determine if an overflow occurred. if ((~x & y & ~res) & 0x8000000000000000) { result[i] = UINT64_MAX; qc_flag = true; } else if ((x & ~y & res) & 0x8000000000000000) { result[i] = 0; qc_flag = true; } else { result[i] = res; } } return qc_flag; }); } void EmitX64::EmitVectorUnsignedSaturatedNarrow16(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { bool qc_flag = false; for (size_t i = 0; i < a.size(); ++i) { const u16 saturated = std::clamp(a[i], 0, 0xFF); result[i] = static_cast(saturated); qc_flag |= saturated != a[i]; } return qc_flag; }); } void EmitX64::EmitVectorUnsignedSaturatedNarrow32(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { bool qc_flag = false; for (size_t i = 0; i < a.size(); ++i) { const u32 saturated = std::clamp(a[i], 0, 0xFFFF); result[i] = static_cast(saturated); qc_flag |= saturated != a[i]; } return qc_flag; }); } void EmitX64::EmitVectorUnsignedSaturatedNarrow64(EmitContext& ctx, IR::Inst* inst) { EmitOneArgumentFallbackWithSaturation(code, ctx, inst, [](VectorArray& result, const VectorArray& a) { bool qc_flag = false; for (size_t i = 0; i < a.size(); ++i) { const u64 saturated = std::clamp(a[i], 0, 0xFFFFFFFF); result[i] = static_cast(saturated); qc_flag |= saturated != a[i]; } return qc_flag; }); } void EmitX64::EmitVectorZeroExtend8(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pmovzxbw(a, a); } else { const Xbyak::Xmm zeros = ctx.reg_alloc.ScratchXmm(); code.pxor(zeros, zeros); code.punpcklbw(a, zeros); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorZeroExtend16(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pmovzxwd(a, a); } else { const Xbyak::Xmm zeros = ctx.reg_alloc.ScratchXmm(); code.pxor(zeros, zeros); code.punpcklwd(a, zeros); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorZeroExtend32(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) { code.pmovzxdq(a, a); } else { const Xbyak::Xmm zeros = ctx.reg_alloc.ScratchXmm(); code.pxor(zeros, zeros); code.punpckldq(a, zeros); } ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorZeroExtend64(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); const Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); const Xbyak::Xmm zeros = ctx.reg_alloc.ScratchXmm(); code.pxor(zeros, zeros); code.punpcklqdq(a, zeros); ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitVectorZeroUpper(EmitContext& ctx, IR::Inst* inst) { auto args = ctx.reg_alloc.GetArgumentInfo(inst); Xbyak::Xmm a = ctx.reg_alloc.UseScratchXmm(args[0]); code.movq(a, a); // TODO: !IsLastUse ctx.reg_alloc.DefineValue(inst, a); } void EmitX64::EmitZeroVector(EmitContext& ctx, IR::Inst* inst) { Xbyak::Xmm a = ctx.reg_alloc.ScratchXmm(); code.pxor(a, a); ctx.reg_alloc.DefineValue(inst, a); } } // namespace Dynarmic::BackendX64