// This file is part of AsmJit project // // See asmjit.h or LICENSE.md for license and copyright information // SPDX-License-Identifier: Zlib #include #if !defined(ASMJIT_NO_X86) #include #include #include #include #include "asmjit_test_misc.h" #include "asmjit_test_perf.h" using namespace asmjit; enum class InstForm { kReg, kMem }; // Generates a long sequence of GP instructions. template static void generateGpSequenceInternal( Emitter& cc, InstForm form, const x86::Gp& a, const x86::Gp& b, const x86::Gp& c, const x86::Gp& d) { cc.mov(a, 0xAAAAAAAA); cc.mov(b, 0xBBBBBBBB); cc.mov(c, 0xCCCCCCCC); cc.mov(d, 0xFFFFFFFF); if (form == InstForm::kReg) { cc.adc(a, b); cc.adc(b, c); cc.adc(c, d); cc.add(a, b); cc.add(b, c); cc.add(c, d); cc.and_(a, b); cc.and_(b, c); cc.and_(c, d); cc.bsf(a, b); cc.bsf(b, c); cc.bsf(c, d); cc.bsr(a, b); cc.bsr(b, c); cc.bsr(c, d); cc.bswap(a); cc.bswap(b); cc.bswap(c); cc.bt(a, b); cc.bt(b, c); cc.bt(c, d); cc.btc(a, b); cc.btc(b, c); cc.btc(c, d); cc.btr(a, b); cc.btr(b, c); cc.btr(c, d); cc.bts(a, b); cc.bts(b, c); cc.bts(c, d); cc.cmp(a, b); cc.cmovc(a, b); cc.cmp(b, c); cc.cmovc(b, c); cc.cmp(c, d); cc.cmovc(c, d); cc.dec(a); cc.dec(b); cc.dec(c); cc.imul(a, b); cc.imul(b, c); cc.imul(c, d); cc.movsx(a, b.r8Lo()); cc.movsx(b, c.r8Lo()); cc.movsx(c, d.r8Lo()); cc.movzx(a, b.r8Lo()); cc.movzx(b, c.r8Lo()); cc.movzx(c, d.r8Lo()); cc.neg(a); cc.neg(b); cc.neg(c); cc.not_(a); cc.not_(b); cc.not_(c); cc.or_(a, b); cc.or_(b, c); cc.or_(c, d); cc.sbb(a, b); cc.sbb(b, c); cc.sbb(c, d); cc.sub(a, b); cc.sub(b, c); cc.sub(c, d); cc.test(a, b); cc.test(b, c); cc.test(c, d); cc.xchg(a, b); cc.xchg(b, c); cc.xchg(c, d); cc.xor_(a, b); cc.xor_(b, c); cc.xor_(c, d); cc.rcl(a, c.r8Lo()); cc.rcl(b, c.r8Lo()); cc.rcl(d, c.r8Lo()); cc.rcr(a, c.r8Lo()); cc.rcr(b, c.r8Lo()); cc.rcr(d, c.r8Lo()); cc.rol(a, c.r8Lo()); cc.rol(b, c.r8Lo()); cc.rol(d, c.r8Lo()); cc.ror(a, c.r8Lo()); cc.ror(b, c.r8Lo()); cc.ror(d, c.r8Lo()); cc.shl(a, c.r8Lo()); cc.shl(b, c.r8Lo()); cc.shl(d, c.r8Lo()); cc.shr(a, c.r8Lo()); cc.shr(b, c.r8Lo()); cc.shr(d, c.r8Lo()); cc.sar(a, c.r8Lo()); cc.sar(b, c.r8Lo()); cc.sar(d, c.r8Lo()); cc.shld(a, b, c.r8Lo()); cc.shld(b, d, c.r8Lo()); cc.shld(d, a, c.r8Lo()); cc.shrd(a, b, c.r8Lo()); cc.shrd(b, d, c.r8Lo()); cc.shrd(d, a, c.r8Lo()); cc.adcx(a, b); cc.adox(a, b); cc.adcx(b, c); cc.adox(b, c); cc.adcx(c, d); cc.adox(c, d); cc.andn(a, b, c); cc.andn(b, c, d); cc.andn(c, d, a); cc.bextr(a, b, c); cc.bextr(b, c, d); cc.bextr(c, d, a); cc.blsi(a, b); cc.blsi(b, c); cc.blsi(c, d); cc.blsmsk(a, b); cc.blsmsk(b, c); cc.blsmsk(c, d); cc.blsr(a, b); cc.blsr(b, c); cc.blsr(c, d); cc.bzhi(a, b, c); cc.bzhi(b, c, d); cc.bzhi(c, d, a); cc.lzcnt(a, b); cc.lzcnt(b, c); cc.lzcnt(c, d); cc.pdep(a, b, c); cc.pdep(b, c, d); cc.pdep(c, d, a); cc.pext(a, b, c); cc.pext(b, c, d); cc.pext(c, d, a); cc.popcnt(a, b); cc.popcnt(b, c); cc.popcnt(c, d); cc.rorx(a, b, 8); cc.rorx(b, c, 8); cc.rorx(c, d, 8); cc.sarx(a, b, c); cc.sarx(b, c, d); cc.sarx(c, d, a); cc.shlx(a, b, c); cc.shlx(b, c, d); cc.shlx(c, d, a); cc.shrx(a, b, c); cc.shrx(b, c, d); cc.shrx(c, d, a); cc.tzcnt(a, b); cc.tzcnt(b, c); cc.tzcnt(c, d); } else { uint32_t regSize = cc.registerSize(); x86::Mem m = x86::ptr(c, 0, regSize); x86::Mem m8 = x86::byte_ptr(c); cc.adc(a, m); cc.adc(b, m); cc.adc(c, m); cc.add(a, m); cc.add(b, m); cc.add(c, m); cc.and_(a, m); cc.and_(b, m); cc.and_(c, m); cc.bsf(a, m); cc.bsf(b, m); cc.bsf(c, m); cc.bsr(a, m); cc.bsr(b, m); cc.bsr(c, m); cc.bt(m, a); cc.bt(m, b); cc.bt(m, c); cc.btc(m, a); cc.btc(m, b); cc.btc(m, c); cc.btr(m, a); cc.btr(m, b); cc.btr(m, c); cc.bts(m, a); cc.bts(m, b); cc.bts(m, c); cc.cmp(a, m); cc.cmovc(a, m); cc.cmp(b, m); cc.cmovc(b, m); cc.cmp(c, m); cc.cmovc(c, m); cc.dec(m); cc.movsx(a, m8); cc.movsx(b, m8); cc.movsx(c, m8); cc.movzx(a, m8); cc.movzx(b, m8); cc.movzx(c, m8); cc.neg(m); cc.not_(m); cc.or_(a, m); cc.or_(b, m); cc.or_(c, m); cc.sbb(a, m); cc.sbb(b, m); cc.sbb(c, m); cc.sub(a, m); cc.sub(b, m); cc.sub(c, m); cc.test(m, a); cc.test(m, b); cc.test(m, c); cc.xchg(a, m); cc.xchg(b, m); cc.xchg(c, m); cc.xor_(a, m); cc.xor_(b, m); cc.xor_(c, m); cc.rcl(m, c.r8Lo()); cc.rcr(m, c.r8Lo()); cc.rol(m, c.r8Lo()); cc.ror(m, c.r8Lo()); cc.shl(m, c.r8Lo()); cc.shr(m, c.r8Lo()); cc.sar(m, c.r8Lo()); cc.shld(m, b, c.r8Lo()); cc.shld(m, d, c.r8Lo()); cc.shld(m, a, c.r8Lo()); cc.shrd(m, b, c.r8Lo()); cc.shrd(m, d, c.r8Lo()); cc.shrd(m, a, c.r8Lo()); cc.adcx(a, m); cc.adox(a, m); cc.adcx(b, m); cc.adox(b, m); cc.adcx(c, m); cc.adox(c, m); cc.andn(a, b, m); cc.andn(b, c, m); cc.andn(c, d, m); cc.bextr(a, m, c); cc.bextr(b, m, d); cc.bextr(c, m, a); cc.blsi(a, m); cc.blsi(b, m); cc.blsi(c, m); cc.blsmsk(a, m); cc.blsmsk(b, m); cc.blsmsk(c, m); cc.blsr(a, m); cc.blsr(b, m); cc.blsr(c, m); cc.bzhi(a, m, c); cc.bzhi(b, m, d); cc.bzhi(c, m, a); cc.lzcnt(a, m); cc.lzcnt(b, m); cc.lzcnt(c, m); cc.pdep(a, b, m); cc.pdep(b, c, m); cc.pdep(c, d, m); cc.pext(a, b, m); cc.pext(b, c, m); cc.pext(c, d, m); cc.popcnt(a, m); cc.popcnt(b, m); cc.popcnt(c, m); cc.rorx(a, m, 8); cc.rorx(b, m, 8); cc.rorx(c, m, 8); cc.sarx(a, m, c); cc.sarx(b, m, d); cc.sarx(c, m, a); cc.shlx(a, m, c); cc.shlx(b, m, d); cc.shlx(c, m, a); cc.shrx(a, m, c); cc.shrx(b, m, d); cc.shrx(c, m, a); cc.tzcnt(a, m); cc.tzcnt(b, m); cc.tzcnt(c, m); } } static void generateGpSequence(BaseEmitter& emitter, InstForm form, bool emitPrologEpilog) { using namespace asmjit::x86; if (emitter.isAssembler()) { Assembler& cc = *emitter.as(); x86::Gp a = cc.zax(); x86::Gp b = cc.zbx(); x86::Gp c = cc.zcx(); x86::Gp d = cc.zdx(); if (emitPrologEpilog) { FuncDetail func; func.init(FuncSignature::build(), cc.environment()); FuncFrame frame; frame.init(func); frame.addDirtyRegs(a, b, c, d); frame.finalize(); cc.emitProlog(frame); generateGpSequenceInternal(cc, form, a, b, c, d); cc.emitEpilog(frame); } else { generateGpSequenceInternal(cc, form, a, b, c, d); } } #ifndef ASMJIT_NO_BUILDER else if (emitter.isBuilder()) { Builder& cc = *emitter.as(); x86::Gp a = cc.zax(); x86::Gp b = cc.zbx(); x86::Gp c = cc.zcx(); x86::Gp d = cc.zdx(); if (emitPrologEpilog) { FuncDetail func; func.init(FuncSignature::build(), cc.environment()); FuncFrame frame; frame.init(func); frame.addDirtyRegs(a, b, c, d); frame.finalize(); cc.emitProlog(frame); generateGpSequenceInternal(cc, form, a, b, c, d); cc.emitEpilog(frame); } else { generateGpSequenceInternal(cc, form, a, b, c, d); } } #endif #ifndef ASMJIT_NO_COMPILER else if (emitter.isCompiler()) { Compiler& cc = *emitter.as(); Gp a = cc.newIntPtr("a"); Gp b = cc.newIntPtr("b"); Gp c = cc.newIntPtr("c"); Gp d = cc.newIntPtr("d"); cc.addFunc(FuncSignature::build()); generateGpSequenceInternal(cc, form, a, b, c, d); cc.endFunc(); } #endif } // Generates a long sequence of SSE instructions using only registers. template static void generateSseSequenceInternal( Emitter& cc, InstForm form, const x86::Gp& gp, const x86::Xmm& xmmA, const x86::Xmm& xmmB, const x86::Xmm& xmmC, const x86::Xmm& xmmD) { x86::Gp gpd = gp.r32(); x86::Gp gpq = gp.r64(); x86::Gp gpz = cc.is32Bit() ? gpd : gpq; cc.xor_(gpd, gpd); cc.xorps(xmmA, xmmA); cc.xorps(xmmB, xmmB); cc.xorps(xmmC, xmmC); cc.xorps(xmmD, xmmD); if (form == InstForm::kReg) { // SSE. cc.addps(xmmA, xmmB); cc.addss(xmmA, xmmB); cc.andnps(xmmA, xmmB); cc.andps(xmmA, xmmB); cc.cmpps(xmmA, xmmB, 0); cc.cmpss(xmmA, xmmB, 0); cc.comiss(xmmA, xmmB); cc.cvtsi2ss(xmmA, gpd); cc.cvtsi2ss(xmmA, gpz); cc.cvtss2si(gpd, xmmB); cc.cvtss2si(gpz, xmmB); cc.cvttss2si(gpd, xmmB); cc.cvttss2si(gpz, xmmB); cc.divps(xmmA, xmmB); cc.divss(xmmA, xmmB); cc.maxps(xmmA, xmmB); cc.maxss(xmmA, xmmB); cc.minps(xmmA, xmmB); cc.minss(xmmA, xmmB); cc.movaps(xmmA, xmmB); cc.movd(gpd, xmmB); cc.movd(xmmA, gpd); cc.movq(xmmA, xmmB); cc.movhlps(xmmA, xmmB); cc.movlhps(xmmA, xmmB); cc.movups(xmmA, xmmB); cc.mulps(xmmA, xmmB); cc.mulss(xmmA, xmmB); cc.orps(xmmA, xmmB); cc.rcpps(xmmA, xmmB); cc.rcpss(xmmA, xmmB); cc.psadbw(xmmA, xmmB); cc.rsqrtps(xmmA, xmmB); cc.rsqrtss(xmmA, xmmB); cc.sfence(); cc.shufps(xmmA, xmmB, 0); cc.sqrtps(xmmA, xmmB); cc.sqrtss(xmmA, xmmB); cc.subps(xmmA, xmmB); cc.subss(xmmA, xmmB); cc.ucomiss(xmmA, xmmB); cc.unpckhps(xmmA, xmmB); cc.unpcklps(xmmA, xmmB); cc.xorps(xmmA, xmmB); // SSE2. cc.addpd(xmmA, xmmB); cc.addsd(xmmA, xmmB); cc.andnpd(xmmA, xmmB); cc.andpd(xmmA, xmmB); cc.cmppd(xmmA, xmmB, 0); cc.cmpsd(xmmA, xmmB, 0); cc.comisd(xmmA, xmmB); cc.cvtdq2pd(xmmA, xmmB); cc.cvtdq2ps(xmmA, xmmB); cc.cvtpd2dq(xmmA, xmmB); cc.cvtpd2ps(xmmA, xmmB); cc.cvtps2dq(xmmA, xmmB); cc.cvtps2pd(xmmA, xmmB); cc.cvtsd2si(gpd, xmmB); cc.cvtsd2si(gpz, xmmB); cc.cvtsd2ss(xmmA, xmmB); cc.cvtsi2sd(xmmA, gpd); cc.cvtsi2sd(xmmA, gpz); cc.cvtss2sd(xmmA, xmmB); cc.cvtss2si(gpd, xmmB); cc.cvtss2si(gpz, xmmB); cc.cvttpd2dq(xmmA, xmmB); cc.cvttps2dq(xmmA, xmmB); cc.cvttsd2si(gpd, xmmB); cc.cvttsd2si(gpz, xmmB); cc.divpd(xmmA, xmmB); cc.divsd(xmmA, xmmB); cc.maxpd(xmmA, xmmB); cc.maxsd(xmmA, xmmB); cc.minpd(xmmA, xmmB); cc.minsd(xmmA, xmmB); cc.movdqa(xmmA, xmmB); cc.movdqu(xmmA, xmmB); cc.movmskps(gpd, xmmB); cc.movmskpd(gpd, xmmB); cc.movsd(xmmA, xmmB); cc.mulpd(xmmA, xmmB); cc.mulsd(xmmA, xmmB); cc.orpd(xmmA, xmmB); cc.packsswb(xmmA, xmmB); cc.packssdw(xmmA, xmmB); cc.packuswb(xmmA, xmmB); cc.paddb(xmmA, xmmB); cc.paddw(xmmA, xmmB); cc.paddd(xmmA, xmmB); cc.paddq(xmmA, xmmB); cc.paddsb(xmmA, xmmB); cc.paddsw(xmmA, xmmB); cc.paddusb(xmmA, xmmB); cc.paddusw(xmmA, xmmB); cc.pand(xmmA, xmmB); cc.pandn(xmmA, xmmB); cc.pavgb(xmmA, xmmB); cc.pavgw(xmmA, xmmB); cc.pcmpeqb(xmmA, xmmB); cc.pcmpeqw(xmmA, xmmB); cc.pcmpeqd(xmmA, xmmB); cc.pcmpgtb(xmmA, xmmB); cc.pcmpgtw(xmmA, xmmB); cc.pcmpgtd(xmmA, xmmB); cc.pmaxsw(xmmA, xmmB); cc.pmaxub(xmmA, xmmB); cc.pminsw(xmmA, xmmB); cc.pminub(xmmA, xmmB); cc.pmovmskb(gpd, xmmB); cc.pmulhw(xmmA, xmmB); cc.pmulhuw(xmmA, xmmB); cc.pmullw(xmmA, xmmB); cc.pmuludq(xmmA, xmmB); cc.por(xmmA, xmmB); cc.pslld(xmmA, xmmB); cc.pslld(xmmA, 0); cc.psllq(xmmA, xmmB); cc.psllq(xmmA, 0); cc.psllw(xmmA, xmmB); cc.psllw(xmmA, 0); cc.pslldq(xmmA, 0); cc.psrad(xmmA, xmmB); cc.psrad(xmmA, 0); cc.psraw(xmmA, xmmB); cc.psraw(xmmA, 0); cc.psubb(xmmA, xmmB); cc.psubw(xmmA, xmmB); cc.psubd(xmmA, xmmB); cc.psubq(xmmA, xmmB); cc.pmaddwd(xmmA, xmmB); cc.pshufd(xmmA, xmmB, 0); cc.pshufhw(xmmA, xmmB, 0); cc.pshuflw(xmmA, xmmB, 0); cc.psrld(xmmA, xmmB); cc.psrld(xmmA, 0); cc.psrlq(xmmA, xmmB); cc.psrlq(xmmA, 0); cc.psrldq(xmmA, 0); cc.psrlw(xmmA, xmmB); cc.psrlw(xmmA, 0); cc.psubsb(xmmA, xmmB); cc.psubsw(xmmA, xmmB); cc.psubusb(xmmA, xmmB); cc.psubusw(xmmA, xmmB); cc.punpckhbw(xmmA, xmmB); cc.punpckhwd(xmmA, xmmB); cc.punpckhdq(xmmA, xmmB); cc.punpckhqdq(xmmA, xmmB); cc.punpcklbw(xmmA, xmmB); cc.punpcklwd(xmmA, xmmB); cc.punpckldq(xmmA, xmmB); cc.punpcklqdq(xmmA, xmmB); cc.pxor(xmmA, xmmB); cc.sqrtpd(xmmA, xmmB); cc.sqrtsd(xmmA, xmmB); cc.subpd(xmmA, xmmB); cc.subsd(xmmA, xmmB); cc.ucomisd(xmmA, xmmB); cc.unpckhpd(xmmA, xmmB); cc.unpcklpd(xmmA, xmmB); cc.xorpd(xmmA, xmmB); // SSE3. cc.addsubpd(xmmA, xmmB); cc.addsubps(xmmA, xmmB); cc.haddpd(xmmA, xmmB); cc.haddps(xmmA, xmmB); cc.hsubpd(xmmA, xmmB); cc.hsubps(xmmA, xmmB); cc.movddup(xmmA, xmmB); cc.movshdup(xmmA, xmmB); cc.movsldup(xmmA, xmmB); // SSSE3. cc.psignb(xmmA, xmmB); cc.psignw(xmmA, xmmB); cc.psignd(xmmA, xmmB); cc.phaddw(xmmA, xmmB); cc.phaddd(xmmA, xmmB); cc.phaddsw(xmmA, xmmB); cc.phsubw(xmmA, xmmB); cc.phsubd(xmmA, xmmB); cc.phsubsw(xmmA, xmmB); cc.pmaddubsw(xmmA, xmmB); cc.pabsb(xmmA, xmmB); cc.pabsw(xmmA, xmmB); cc.pabsd(xmmA, xmmB); cc.pmulhrsw(xmmA, xmmB); cc.pshufb(xmmA, xmmB); cc.palignr(xmmA, xmmB, 0); // SSE4.1. cc.blendpd(xmmA, xmmB, 0); cc.blendps(xmmA, xmmB, 0); cc.blendvpd(xmmA, xmmB, xmmA); cc.blendvps(xmmA, xmmB, xmmA); cc.dppd(xmmA, xmmB, 0); cc.dpps(xmmA, xmmB, 0); cc.extractps(gpd, xmmB, 0); cc.insertps(xmmA, xmmB, 0); cc.mpsadbw(xmmA, xmmB, 0); cc.packusdw(xmmA, xmmB); cc.pblendvb(xmmA, xmmB, xmmA); cc.pblendw(xmmA, xmmB, 0); cc.pcmpeqq(xmmA, xmmB); cc.pextrb(gpd, xmmB, 0); cc.pextrd(gpd, xmmB, 0); if (cc.is64Bit()) cc.pextrq(gpq, xmmB, 0); cc.pextrw(gpd, xmmB, 0); cc.phminposuw(xmmA, xmmB); cc.pinsrb(xmmA, gpd, 0); cc.pinsrd(xmmA, gpd, 0); cc.pinsrw(xmmA, gpd, 0); cc.pmaxuw(xmmA, xmmB); cc.pmaxsb(xmmA, xmmB); cc.pmaxsd(xmmA, xmmB); cc.pmaxud(xmmA, xmmB); cc.pminsb(xmmA, xmmB); cc.pminuw(xmmA, xmmB); cc.pminud(xmmA, xmmB); cc.pminsd(xmmA, xmmB); cc.pmovsxbw(xmmA, xmmB); cc.pmovsxbd(xmmA, xmmB); cc.pmovsxbq(xmmA, xmmB); cc.pmovsxwd(xmmA, xmmB); cc.pmovsxwq(xmmA, xmmB); cc.pmovsxdq(xmmA, xmmB); cc.pmovzxbw(xmmA, xmmB); cc.pmovzxbd(xmmA, xmmB); cc.pmovzxbq(xmmA, xmmB); cc.pmovzxwd(xmmA, xmmB); cc.pmovzxwq(xmmA, xmmB); cc.pmovzxdq(xmmA, xmmB); cc.pmuldq(xmmA, xmmB); cc.pmulld(xmmA, xmmB); cc.ptest(xmmA, xmmB); cc.roundps(xmmA, xmmB, 0); cc.roundss(xmmA, xmmB, 0); cc.roundpd(xmmA, xmmB, 0); cc.roundsd(xmmA, xmmB, 0); } else { x86::Mem m = x86::ptr(gpz); cc.addps(xmmA, m); cc.addss(xmmA, m); cc.andnps(xmmA, m); cc.andps(xmmA, m); cc.cmpps(xmmA, m, 0); cc.cmpss(xmmA, m, 0); cc.comiss(xmmA, m); cc.cvtpi2ps(xmmA, m); cc.cvtsi2ss(xmmA, m); cc.cvtss2si(gpd, m); cc.cvtss2si(gpz, m); cc.cvttss2si(gpd, m); cc.cvttss2si(gpz, m); cc.divps(xmmA, m); cc.divss(xmmA, m); cc.maxps(xmmA, m); cc.maxss(xmmA, m); cc.minps(xmmA, m); cc.minss(xmmA, m); cc.movaps(xmmA, m); cc.movaps(m, xmmB); cc.movd(m, xmmB); cc.movd(xmmA, m); cc.movq(m, xmmB); cc.movq(xmmA, m); cc.movhps(xmmA, m); cc.movhps(m, xmmB); cc.movlps(xmmA, m); cc.movlps(m, xmmB); cc.movntps(m, xmmB); cc.movss(xmmA, m); cc.movss(m, xmmB); cc.movups(xmmA, m); cc.movups(m, xmmB); cc.mulps(xmmA, m); cc.mulss(xmmA, m); cc.orps(xmmA, m); cc.rcpps(xmmA, m); cc.rcpss(xmmA, m); cc.psadbw(xmmA, m); cc.rsqrtps(xmmA, m); cc.rsqrtss(xmmA, m); cc.shufps(xmmA, m, 0); cc.sqrtps(xmmA, m); cc.sqrtss(xmmA, m); cc.stmxcsr(m); cc.subps(xmmA, m); cc.subss(xmmA, m); cc.ucomiss(xmmA, m); cc.unpckhps(xmmA, m); cc.unpcklps(xmmA, m); cc.xorps(xmmA, m); // SSE2. cc.addpd(xmmA, m); cc.addsd(xmmA, m); cc.andnpd(xmmA, m); cc.andpd(xmmA, m); cc.cmppd(xmmA, m, 0); cc.cmpsd(xmmA, m, 0); cc.comisd(xmmA, m); cc.cvtdq2pd(xmmA, m); cc.cvtdq2ps(xmmA, m); cc.cvtpd2dq(xmmA, m); cc.cvtpd2ps(xmmA, m); cc.cvtpi2pd(xmmA, m); cc.cvtps2dq(xmmA, m); cc.cvtps2pd(xmmA, m); cc.cvtsd2si(gpd, m); cc.cvtsd2si(gpz, m); cc.cvtsd2ss(xmmA, m); cc.cvtsi2sd(xmmA, m); cc.cvtss2sd(xmmA, m); cc.cvtss2si(gpd, m); cc.cvtss2si(gpz, m); cc.cvttpd2dq(xmmA, m); cc.cvttps2dq(xmmA, m); cc.cvttsd2si(gpd, m); cc.cvttsd2si(gpz, m); cc.divpd(xmmA, m); cc.divsd(xmmA, m); cc.maxpd(xmmA, m); cc.maxsd(xmmA, m); cc.minpd(xmmA, m); cc.minsd(xmmA, m); cc.movdqa(xmmA, m); cc.movdqa(m, xmmB); cc.movdqu(xmmA, m); cc.movdqu(m, xmmB); cc.movsd(xmmA, m); cc.movsd(m, xmmB); cc.movapd(xmmA, m); cc.movapd(m, xmmB); cc.movhpd(xmmA, m); cc.movhpd(m, xmmB); cc.movlpd(xmmA, m); cc.movlpd(m, xmmB); cc.movntdq(m, xmmB); cc.movntpd(m, xmmB); cc.movupd(xmmA, m); cc.movupd(m, xmmB); cc.mulpd(xmmA, m); cc.mulsd(xmmA, m); cc.orpd(xmmA, m); cc.packsswb(xmmA, m); cc.packssdw(xmmA, m); cc.packuswb(xmmA, m); cc.paddb(xmmA, m); cc.paddw(xmmA, m); cc.paddd(xmmA, m); cc.paddq(xmmA, m); cc.paddsb(xmmA, m); cc.paddsw(xmmA, m); cc.paddusb(xmmA, m); cc.paddusw(xmmA, m); cc.pand(xmmA, m); cc.pandn(xmmA, m); cc.pavgb(xmmA, m); cc.pavgw(xmmA, m); cc.pcmpeqb(xmmA, m); cc.pcmpeqw(xmmA, m); cc.pcmpeqd(xmmA, m); cc.pcmpgtb(xmmA, m); cc.pcmpgtw(xmmA, m); cc.pcmpgtd(xmmA, m); cc.pmaxsw(xmmA, m); cc.pmaxub(xmmA, m); cc.pminsw(xmmA, m); cc.pminub(xmmA, m); cc.pmulhw(xmmA, m); cc.pmulhuw(xmmA, m); cc.pmullw(xmmA, m); cc.pmuludq(xmmA, m); cc.por(xmmA, m); cc.pslld(xmmA, m); cc.psllq(xmmA, m); cc.psllw(xmmA, m); cc.psrad(xmmA, m); cc.psraw(xmmA, m); cc.psubb(xmmA, m); cc.psubw(xmmA, m); cc.psubd(xmmA, m); cc.psubq(xmmA, m); cc.pmaddwd(xmmA, m); cc.pshufd(xmmA, m, 0); cc.pshufhw(xmmA, m, 0); cc.pshuflw(xmmA, m, 0); cc.psrld(xmmA, m); cc.psrlq(xmmA, m); cc.psrlw(xmmA, m); cc.psubsb(xmmA, m); cc.psubsw(xmmA, m); cc.psubusb(xmmA, m); cc.psubusw(xmmA, m); cc.punpckhbw(xmmA, m); cc.punpckhwd(xmmA, m); cc.punpckhdq(xmmA, m); cc.punpckhqdq(xmmA, m); cc.punpcklbw(xmmA, m); cc.punpcklwd(xmmA, m); cc.punpckldq(xmmA, m); cc.punpcklqdq(xmmA, m); cc.pxor(xmmA, m); cc.sqrtpd(xmmA, m); cc.sqrtsd(xmmA, m); cc.subpd(xmmA, m); cc.subsd(xmmA, m); cc.ucomisd(xmmA, m); cc.unpckhpd(xmmA, m); cc.unpcklpd(xmmA, m); cc.xorpd(xmmA, m); // SSE3. cc.addsubpd(xmmA, m); cc.addsubps(xmmA, m); cc.haddpd(xmmA, m); cc.haddps(xmmA, m); cc.hsubpd(xmmA, m); cc.hsubps(xmmA, m); cc.lddqu(xmmA, m); cc.movddup(xmmA, m); cc.movshdup(xmmA, m); cc.movsldup(xmmA, m); // SSSE3. cc.psignb(xmmA, m); cc.psignw(xmmA, m); cc.psignd(xmmA, m); cc.phaddw(xmmA, m); cc.phaddd(xmmA, m); cc.phaddsw(xmmA, m); cc.phsubw(xmmA, m); cc.phsubd(xmmA, m); cc.phsubsw(xmmA, m); cc.pmaddubsw(xmmA, m); cc.pabsb(xmmA, m); cc.pabsw(xmmA, m); cc.pabsd(xmmA, m); cc.pmulhrsw(xmmA, m); cc.pshufb(xmmA, m); cc.palignr(xmmA, m, 0); // SSE4.1. cc.blendpd(xmmA, m, 0); cc.blendps(xmmA, m, 0); cc.blendvpd(xmmA, m, xmmA); cc.blendvps(xmmA, m, xmmA); cc.dppd(xmmA, m, 0); cc.dpps(xmmA, m, 0); cc.extractps(m, xmmB, 0); cc.insertps(xmmA, m, 0); cc.movntdqa(xmmA, m); cc.mpsadbw(xmmA, m, 0); cc.packusdw(xmmA, m); cc.pblendvb(xmmA, m, xmmA); cc.pblendw(xmmA, m, 0); cc.pcmpeqq(xmmA, m); cc.pextrb(m, xmmB, 0); cc.pextrd(m, xmmB, 0); if (cc.is64Bit()) cc.pextrq(m, xmmB, 0); cc.pextrw(m, xmmB, 0); cc.phminposuw(xmmA, m); cc.pinsrb(xmmA, m, 0); cc.pinsrd(xmmA, m, 0); cc.pinsrw(xmmA, m, 0); cc.pmaxuw(xmmA, m); cc.pmaxsb(xmmA, m); cc.pmaxsd(xmmA, m); cc.pmaxud(xmmA, m); cc.pminsb(xmmA, m); cc.pminuw(xmmA, m); cc.pminud(xmmA, m); cc.pminsd(xmmA, m); cc.pmovsxbw(xmmA, m); cc.pmovsxbd(xmmA, m); cc.pmovsxbq(xmmA, m); cc.pmovsxwd(xmmA, m); cc.pmovsxwq(xmmA, m); cc.pmovsxdq(xmmA, m); cc.pmovzxbw(xmmA, m); cc.pmovzxbd(xmmA, m); cc.pmovzxbq(xmmA, m); cc.pmovzxwd(xmmA, m); cc.pmovzxwq(xmmA, m); cc.pmovzxdq(xmmA, m); cc.pmuldq(xmmA, m); cc.pmulld(xmmA, m); cc.ptest(xmmA, m); cc.roundps(xmmA, m, 0); cc.roundss(xmmA, m, 0); cc.roundpd(xmmA, m, 0); cc.roundsd(xmmA, m, 0); // SSE4.2. cc.pcmpgtq(xmmA, m); } } static void generateSseSequence(BaseEmitter& emitter, InstForm form, bool emitPrologEpilog) { using namespace asmjit::x86; if (emitter.isAssembler()) { Assembler& cc = *emitter.as(); if (emitPrologEpilog) { FuncDetail func; func.init(FuncSignature::build(), cc.environment()); FuncFrame frame; frame.init(func); frame.addDirtyRegs(eax, xmm0, xmm1, xmm2, xmm3); frame.finalize(); cc.emitProlog(frame); generateSseSequenceInternal(cc, form, eax, xmm0, xmm1, xmm2, xmm3); cc.emitEpilog(frame); } else { generateSseSequenceInternal(cc, form, eax, xmm0, xmm1, xmm2, xmm3); } } #ifndef ASMJIT_NO_BUILDER else if (emitter.isBuilder()) { Builder& cc = *emitter.as(); if (emitPrologEpilog) { FuncDetail func; func.init(FuncSignature::build(), cc.environment()); FuncFrame frame; frame.init(func); frame.addDirtyRegs(eax, xmm0, xmm1, xmm2, xmm3); frame.finalize(); cc.emitProlog(frame); generateSseSequenceInternal(cc, form, eax, xmm0, xmm1, xmm2, xmm3); cc.emitEpilog(frame); } else { generateSseSequenceInternal(cc, form, eax, xmm0, xmm1, xmm2, xmm3); } } #endif #ifndef ASMJIT_NO_COMPILER else if (emitter.isCompiler()) { Compiler& cc = *emitter.as(); Gp gp = cc.newGpz("gp"); Xmm a = cc.newXmm("a"); Xmm b = cc.newXmm("b"); Xmm c = cc.newXmm("c"); Xmm d = cc.newXmm("d"); cc.addFunc(FuncSignature::build()); generateSseSequenceInternal(cc, form, gp, a, b, c, d); cc.endFunc(); } #endif } // Generates a long sequence of AVX instructions. template static void generateAvxSequenceInternalRegOnly( Emitter& cc, const x86::Gp& gp, const x86::Vec& vecA, const x86::Vec& vecB, const x86::Vec& vecC, const x86::Vec& vecD) { x86::Gp gpd = gp.r32(); x86::Gp gpq = gp.r64(); x86::Gp gpz = cc.is32Bit() ? gpd : gpq; x86::Xmm xmmA = vecA.xmm(); x86::Xmm xmmB = vecB.xmm(); x86::Xmm xmmC = vecC.xmm(); x86::Xmm xmmD = vecD.xmm(); x86::Ymm ymmA = vecA.ymm(); x86::Ymm ymmB = vecB.ymm(); x86::Ymm ymmC = vecC.ymm(); cc.xor_(gpd, gpd); cc.vxorps(xmmA, xmmA, xmmA); cc.vxorps(xmmB, xmmB, xmmB); cc.vxorps(xmmC, xmmC, xmmC); cc.vxorps(xmmD, xmmD, xmmD); cc.vaddpd(xmmA, xmmB, xmmC); cc.vaddpd(ymmA, ymmB, ymmC); cc.vaddps(xmmA, xmmB, xmmC); cc.vaddps(ymmA, ymmB, ymmC); cc.vaddsd(xmmA, xmmB, xmmC); cc.vaddss(xmmA, xmmB, xmmC); cc.vaddsubpd(xmmA, xmmB, xmmC); cc.vaddsubpd(ymmA, ymmB, ymmC); cc.vaddsubps(xmmA, xmmB, xmmC); cc.vaddsubps(ymmA, ymmB, ymmC); cc.vandpd(xmmA, xmmB, xmmC); cc.vandpd(ymmA, ymmB, ymmC); cc.vandps(xmmA, xmmB, xmmC); cc.vandps(ymmA, ymmB, ymmC); cc.vandnpd(xmmA, xmmB, xmmC); cc.vandnpd(ymmA, ymmB, ymmC); cc.vandnps(xmmA, xmmB, xmmC); cc.vandnps(ymmA, ymmB, ymmC); cc.vblendpd(xmmA, xmmB, xmmC, 0); cc.vblendpd(ymmA, ymmB, ymmC, 0); cc.vblendps(xmmA, xmmB, xmmC, 0); cc.vblendps(ymmA, ymmB, ymmC, 0); cc.vblendvpd(xmmA, xmmB, xmmC, xmmA); cc.vblendvpd(ymmA, ymmB, ymmC, ymmA); cc.vcmppd(xmmA, xmmB, xmmC, 0); cc.vcmppd(ymmA, ymmB, ymmC, 0); cc.vcmpps(xmmA, xmmB, xmmC, 0); cc.vcmpps(ymmA, ymmB, ymmC, 0); cc.vcmpsd(xmmA, xmmB, xmmC, 0); cc.vcmpss(xmmA, xmmB, xmmC, 0); cc.vcomisd(xmmA, xmmB); cc.vcomiss(xmmA, xmmB); cc.vcvtdq2pd(xmmA, xmmB); cc.vcvtdq2pd(ymmA, xmmB); cc.vcvtdq2ps(xmmA, xmmB); cc.vcvtdq2ps(ymmA, ymmB); cc.vcvtpd2dq(xmmA, xmmB); cc.vcvtpd2dq(xmmA, ymmB); cc.vcvtpd2ps(xmmA, xmmB); cc.vcvtpd2ps(xmmA, ymmB); cc.vcvtps2dq(xmmA, xmmB); cc.vcvtps2dq(ymmA, ymmB); cc.vcvtps2pd(xmmA, xmmB); cc.vcvtps2pd(ymmA, xmmB); cc.vcvtsd2si(gpd, xmmB); cc.vcvtsd2si(gpz, xmmB); cc.vcvtsd2ss(xmmA, xmmB, xmmC); cc.vcvtsi2sd(xmmA, xmmB, gpd); cc.vcvtsi2sd(xmmA, xmmB, gpz); cc.vcvtsi2ss(xmmA, xmmB, gpd); cc.vcvtsi2ss(xmmA, xmmB, gpz); cc.vcvtss2sd(xmmA, xmmB, xmmC); cc.vcvtss2si(gpd, xmmB); cc.vcvttpd2dq(xmmA, xmmB); cc.vcvttpd2dq(xmmA, ymmB); cc.vcvttps2dq(xmmA, xmmB); cc.vcvttps2dq(ymmA, ymmB); cc.vcvttsd2si(gpd, xmmB); cc.vcvttss2si(gpz, xmmB); cc.vdivpd(xmmA, xmmB, xmmC); cc.vdivpd(ymmA, ymmB, ymmC); cc.vdivps(xmmA, xmmB, xmmC); cc.vdivps(ymmA, ymmB, ymmC); cc.vdivsd(xmmA, xmmB, xmmC); cc.vdivss(xmmA, xmmB, xmmC); cc.vdppd(xmmA, xmmB, xmmC, 0); cc.vdpps(xmmA, xmmB, xmmC, 0); cc.vdpps(ymmA, ymmB, ymmC, 0); cc.vextractf128(xmmA, ymmB, 0); cc.vextractps(gpd, xmmB, 0); cc.vhaddpd(xmmA, xmmB, xmmC); cc.vhaddpd(ymmA, ymmB, ymmC); cc.vhaddps(xmmA, xmmB, xmmC); cc.vhaddps(ymmA, ymmB, ymmC); cc.vhsubpd(xmmA, xmmB, xmmC); cc.vhsubpd(ymmA, ymmB, ymmC); cc.vhsubps(xmmA, xmmB, xmmC); cc.vhsubps(ymmA, ymmB, ymmC); cc.vinsertf128(ymmA, ymmB, xmmC, 0); cc.vinsertps(xmmA, xmmB, xmmC, 0); cc.vmaxpd(xmmA, xmmB, xmmC); cc.vmaxpd(ymmA, ymmB, ymmC); cc.vmaxps(xmmA, xmmB, xmmC); cc.vmaxps(ymmA, ymmB, ymmC); cc.vmaxsd(xmmA, xmmB, xmmC); cc.vmaxss(xmmA, xmmB, xmmC); cc.vminpd(xmmA, xmmB, xmmC); cc.vminpd(ymmA, ymmB, ymmC); cc.vminps(xmmA, xmmB, xmmC); cc.vminps(ymmA, ymmB, ymmC); cc.vminsd(xmmA, xmmB, xmmC); cc.vminss(xmmA, xmmB, xmmC); cc.vmovapd(xmmA, xmmB); cc.vmovapd(ymmA, ymmB); cc.vmovaps(xmmA, xmmB); cc.vmovaps(ymmA, ymmB); cc.vmovd(xmmA, gpd); cc.vmovd(gpd, xmmB); cc.vmovddup(xmmA, xmmB); cc.vmovddup(ymmA, ymmB); cc.vmovdqa(xmmA, xmmB); cc.vmovdqa(ymmA, ymmB); cc.vmovdqu(xmmA, xmmB); cc.vmovdqu(ymmA, ymmB); cc.vmovhlps(xmmA, xmmB, xmmC); cc.vmovlhps(xmmA, xmmB, xmmC); cc.vmovmskpd(gpd, xmmB); cc.vmovmskpd(gpd, ymmB); cc.vmovmskps(gpd, xmmB); cc.vmovmskps(gpd, ymmB); cc.vmovsd(xmmA, xmmB, xmmC); cc.vmovshdup(xmmA, xmmB); cc.vmovshdup(ymmA, ymmB); cc.vmovsldup(xmmA, xmmB); cc.vmovsldup(ymmA, ymmB); cc.vmovss(xmmA, xmmB, xmmC); cc.vmovupd(xmmA, xmmB); cc.vmovupd(ymmA, ymmB); cc.vmovups(xmmA, xmmB); cc.vmovups(ymmA, ymmB); cc.vmpsadbw(xmmA, xmmB, xmmC, 0); cc.vmulpd(xmmA, xmmB, xmmC); cc.vmulpd(ymmA, ymmB, ymmC); cc.vmulps(xmmA, xmmB, xmmC); cc.vmulps(ymmA, ymmB, ymmC); cc.vmulsd(xmmA, xmmB, xmmC); cc.vmulss(xmmA, xmmB, xmmC); cc.vorpd(xmmA, xmmB, xmmC); cc.vorpd(ymmA, ymmB, ymmC); cc.vorps(xmmA, xmmB, xmmC); cc.vorps(ymmA, ymmB, ymmC); cc.vpabsb(xmmA, xmmB); cc.vpabsd(xmmA, xmmB); cc.vpabsw(xmmA, xmmB); cc.vpackssdw(xmmA, xmmB, xmmC); cc.vpacksswb(xmmA, xmmB, xmmC); cc.vpackusdw(xmmA, xmmB, xmmC); cc.vpackuswb(xmmA, xmmB, xmmC); cc.vpaddb(xmmA, xmmB, xmmC); cc.vpaddd(xmmA, xmmB, xmmC); cc.vpaddq(xmmA, xmmB, xmmC); cc.vpaddw(xmmA, xmmB, xmmC); cc.vpaddsb(xmmA, xmmB, xmmC); cc.vpaddsw(xmmA, xmmB, xmmC); cc.vpaddusb(xmmA, xmmB, xmmC); cc.vpaddusw(xmmA, xmmB, xmmC); cc.vpalignr(xmmA, xmmB, xmmC, 0); cc.vpand(xmmA, xmmB, xmmC); cc.vpandn(xmmA, xmmB, xmmC); cc.vpavgb(xmmA, xmmB, xmmC); cc.vpavgw(xmmA, xmmB, xmmC); cc.vpblendvb(xmmA, xmmB, xmmC, xmmA); cc.vpblendw(xmmA, xmmB, xmmC, 0); cc.vpcmpeqb(xmmA, xmmB, xmmC); cc.vpcmpeqd(xmmA, xmmB, xmmC); cc.vpcmpeqq(xmmA, xmmB, xmmC); cc.vpcmpeqw(xmmA, xmmB, xmmC); cc.vpcmpgtb(xmmA, xmmB, xmmC); cc.vpcmpgtd(xmmA, xmmB, xmmC); cc.vpcmpgtq(xmmA, xmmB, xmmC); cc.vpcmpgtw(xmmA, xmmB, xmmC); cc.vpermilpd(xmmA, xmmB, xmmC); cc.vpermilpd(ymmA, ymmB, ymmC); cc.vpermilpd(xmmA, xmmB, 0); cc.vpermilpd(ymmA, ymmB, 0); cc.vpermilps(xmmA, xmmB, xmmC); cc.vpermilps(ymmA, ymmB, ymmC); cc.vpermilps(xmmA, xmmB, 0); cc.vpermilps(ymmA, ymmB, 0); cc.vperm2f128(ymmA, ymmB, ymmC, 0); cc.vpextrb(gpd, xmmB, 0); cc.vpextrd(gpd, xmmB, 0); if (cc.is64Bit()) cc.vpextrq(gpq, xmmB, 0); cc.vpextrw(gpd, xmmB, 0); cc.vphaddd(xmmA, xmmB, xmmC); cc.vphaddsw(xmmA, xmmB, xmmC); cc.vphaddw(xmmA, xmmB, xmmC); cc.vphminposuw(xmmA, xmmB); cc.vphsubd(xmmA, xmmB, xmmC); cc.vphsubsw(xmmA, xmmB, xmmC); cc.vphsubw(xmmA, xmmB, xmmC); cc.vpinsrb(xmmA, xmmB, gpd, 0); cc.vpinsrd(xmmA, xmmB, gpd, 0); cc.vpinsrw(xmmA, xmmB, gpd, 0); cc.vpmaddubsw(xmmA, xmmB, xmmC); cc.vpmaddwd(xmmA, xmmB, xmmC); cc.vpmaxsb(xmmA, xmmB, xmmC); cc.vpmaxsd(xmmA, xmmB, xmmC); cc.vpmaxsw(xmmA, xmmB, xmmC); cc.vpmaxub(xmmA, xmmB, xmmC); cc.vpmaxud(xmmA, xmmB, xmmC); cc.vpmaxuw(xmmA, xmmB, xmmC); cc.vpminsb(xmmA, xmmB, xmmC); cc.vpminsd(xmmA, xmmB, xmmC); cc.vpminsw(xmmA, xmmB, xmmC); cc.vpminub(xmmA, xmmB, xmmC); cc.vpminud(xmmA, xmmB, xmmC); cc.vpminuw(xmmA, xmmB, xmmC); cc.vpmovmskb(gpd, xmmB); cc.vpmovsxbd(xmmA, xmmB); cc.vpmovsxbq(xmmA, xmmB); cc.vpmovsxbw(xmmA, xmmB); cc.vpmovsxdq(xmmA, xmmB); cc.vpmovsxwd(xmmA, xmmB); cc.vpmovsxwq(xmmA, xmmB); cc.vpmovzxbd(xmmA, xmmB); cc.vpmovzxbq(xmmA, xmmB); cc.vpmovzxbw(xmmA, xmmB); cc.vpmovzxdq(xmmA, xmmB); cc.vpmovzxwd(xmmA, xmmB); cc.vpmovzxwq(xmmA, xmmB); cc.vpmuldq(xmmA, xmmB, xmmC); cc.vpmulhrsw(xmmA, xmmB, xmmC); cc.vpmulhuw(xmmA, xmmB, xmmC); cc.vpmulhw(xmmA, xmmB, xmmC); cc.vpmulld(xmmA, xmmB, xmmC); cc.vpmullw(xmmA, xmmB, xmmC); cc.vpmuludq(xmmA, xmmB, xmmC); cc.vpor(xmmA, xmmB, xmmC); cc.vpsadbw(xmmA, xmmB, xmmC); cc.vpshufb(xmmA, xmmB, xmmC); cc.vpshufd(xmmA, xmmB, 0); cc.vpshufhw(xmmA, xmmB, 0); cc.vpshuflw(xmmA, xmmB, 0); cc.vpsignb(xmmA, xmmB, xmmC); cc.vpsignd(xmmA, xmmB, xmmC); cc.vpsignw(xmmA, xmmB, xmmC); cc.vpslld(xmmA, xmmB, xmmC); cc.vpslld(xmmA, xmmB, 0); cc.vpslldq(xmmA, xmmB, 0); cc.vpsllq(xmmA, xmmB, xmmC); cc.vpsllq(xmmA, xmmB, 0); cc.vpsllw(xmmA, xmmB, xmmC); cc.vpsllw(xmmA, xmmB, 0); cc.vpsrad(xmmA, xmmB, xmmC); cc.vpsrad(xmmA, xmmB, 0); cc.vpsraw(xmmA, xmmB, xmmC); cc.vpsraw(xmmA, xmmB, 0); cc.vpsrld(xmmA, xmmB, xmmC); cc.vpsrld(xmmA, xmmB, 0); cc.vpsrldq(xmmA, xmmB, 0); cc.vpsrlq(xmmA, xmmB, xmmC); cc.vpsrlq(xmmA, xmmB, 0); cc.vpsrlw(xmmA, xmmB, xmmC); cc.vpsrlw(xmmA, xmmB, 0); cc.vpsubb(xmmA, xmmB, xmmC); cc.vpsubd(xmmA, xmmB, xmmC); cc.vpsubq(xmmA, xmmB, xmmC); cc.vpsubw(xmmA, xmmB, xmmC); cc.vpsubsb(xmmA, xmmB, xmmC); cc.vpsubsw(xmmA, xmmB, xmmC); cc.vpsubusb(xmmA, xmmB, xmmC); cc.vpsubusw(xmmA, xmmB, xmmC); cc.vptest(xmmA, xmmB); cc.vptest(ymmA, ymmB); cc.vpunpckhbw(xmmA, xmmB, xmmC); cc.vpunpckhdq(xmmA, xmmB, xmmC); cc.vpunpckhqdq(xmmA, xmmB, xmmC); cc.vpunpckhwd(xmmA, xmmB, xmmC); cc.vpunpcklbw(xmmA, xmmB, xmmC); cc.vpunpckldq(xmmA, xmmB, xmmC); cc.vpunpcklqdq(xmmA, xmmB, xmmC); cc.vpunpcklwd(xmmA, xmmB, xmmC); cc.vpxor(xmmA, xmmB, xmmC); cc.vrcpps(xmmA, xmmB); cc.vrcpps(ymmA, ymmB); cc.vrcpss(xmmA, xmmB, xmmC); cc.vrsqrtps(xmmA, xmmB); cc.vrsqrtps(ymmA, ymmB); cc.vrsqrtss(xmmA, xmmB, xmmC); cc.vroundpd(xmmA, xmmB, 0); cc.vroundpd(ymmA, ymmB, 0); cc.vroundps(xmmA, xmmB, 0); cc.vroundps(ymmA, ymmB, 0); cc.vroundsd(xmmA, xmmB, xmmC, 0); cc.vroundss(xmmA, xmmB, xmmC, 0); cc.vshufpd(xmmA, xmmB, xmmC, 0); cc.vshufpd(ymmA, ymmB, ymmC, 0); cc.vshufps(xmmA, xmmB, xmmC, 0); cc.vshufps(ymmA, ymmB, ymmC, 0); cc.vsqrtpd(xmmA, xmmB); cc.vsqrtpd(ymmA, ymmB); cc.vsqrtps(xmmA, xmmB); cc.vsqrtps(ymmA, ymmB); cc.vsqrtsd(xmmA, xmmB, xmmC); cc.vsqrtss(xmmA, xmmB, xmmC); cc.vsubpd(xmmA, xmmB, xmmC); cc.vsubpd(ymmA, ymmB, ymmC); cc.vsubps(xmmA, xmmB, xmmC); cc.vsubps(ymmA, ymmB, ymmC); cc.vsubsd(xmmA, xmmB, xmmC); cc.vsubss(xmmA, xmmB, xmmC); cc.vtestps(xmmA, xmmB); cc.vtestps(ymmA, ymmB); cc.vtestpd(xmmA, xmmB); cc.vtestpd(ymmA, ymmB); cc.vucomisd(xmmA, xmmB); cc.vucomiss(xmmA, xmmB); cc.vunpckhpd(xmmA, xmmB, xmmC); cc.vunpckhpd(ymmA, ymmB, ymmC); cc.vunpckhps(xmmA, xmmB, xmmC); cc.vunpckhps(ymmA, ymmB, ymmC); cc.vunpcklpd(xmmA, xmmB, xmmC); cc.vunpcklpd(ymmA, ymmB, ymmC); cc.vunpcklps(xmmA, xmmB, xmmC); cc.vunpcklps(ymmA, ymmB, ymmC); cc.vxorpd(xmmA, xmmB, xmmC); cc.vxorpd(ymmA, ymmB, ymmC); cc.vxorps(xmmA, xmmB, xmmC); cc.vxorps(ymmA, ymmB, ymmC); // AVX+AESNI. cc.vaesdec(xmmA, xmmB, xmmC); cc.vaesdeclast(xmmA, xmmB, xmmC); cc.vaesenc(xmmA, xmmB, xmmC); cc.vaesenclast(xmmA, xmmB, xmmC); cc.vaesimc(xmmA, xmmB); cc.vaeskeygenassist(xmmA, xmmB, 0); // AVX+PCLMULQDQ. cc.vpclmulqdq(xmmA, xmmB, xmmC, 0); // AVX2. cc.vbroadcastsd(ymmA, xmmB); cc.vbroadcastss(xmmA, xmmB); cc.vbroadcastss(ymmA, xmmB); cc.vextracti128(xmmA, ymmB, 0); cc.vinserti128(ymmA, ymmB, xmmC, 0); cc.vmpsadbw(ymmA, ymmB, ymmC, 0); cc.vpabsb(ymmA, ymmB); cc.vpabsd(ymmA, ymmB); cc.vpabsw(ymmA, ymmB); cc.vpackssdw(ymmA, ymmB, ymmC); cc.vpacksswb(ymmA, ymmB, ymmC); cc.vpackusdw(ymmA, ymmB, ymmC); cc.vpackuswb(ymmA, ymmB, ymmC); cc.vpaddb(ymmA, ymmB, ymmC); cc.vpaddd(ymmA, ymmB, ymmC); cc.vpaddq(ymmA, ymmB, ymmC); cc.vpaddw(ymmA, ymmB, ymmC); cc.vpaddsb(ymmA, ymmB, ymmC); cc.vpaddsw(ymmA, ymmB, ymmC); cc.vpaddusb(ymmA, ymmB, ymmC); cc.vpaddusw(ymmA, ymmB, ymmC); cc.vpalignr(ymmA, ymmB, ymmC, 0); cc.vpand(ymmA, ymmB, ymmC); cc.vpandn(ymmA, ymmB, ymmC); cc.vpavgb(ymmA, ymmB, ymmC); cc.vpavgw(ymmA, ymmB, ymmC); cc.vpblendd(xmmA, xmmB, xmmC, 0); cc.vpblendd(ymmA, ymmB, ymmC, 0); cc.vpblendvb(ymmA, ymmB, ymmC, ymmA); cc.vpblendw(ymmA, ymmB, ymmC, 0); cc.vpbroadcastb(xmmA, xmmB); cc.vpbroadcastb(ymmA, xmmB); cc.vpbroadcastd(xmmA, xmmB); cc.vpbroadcastd(ymmA, xmmB); cc.vpbroadcastq(xmmA, xmmB); cc.vpbroadcastq(ymmA, xmmB); cc.vpbroadcastw(xmmA, xmmB); cc.vpbroadcastw(ymmA, xmmB); cc.vpcmpeqb(ymmA, ymmB, ymmC); cc.vpcmpeqd(ymmA, ymmB, ymmC); cc.vpcmpeqq(ymmA, ymmB, ymmC); cc.vpcmpeqw(ymmA, ymmB, ymmC); cc.vpcmpgtb(ymmA, ymmB, ymmC); cc.vpcmpgtd(ymmA, ymmB, ymmC); cc.vpcmpgtq(ymmA, ymmB, ymmC); cc.vpcmpgtw(ymmA, ymmB, ymmC); cc.vperm2i128(ymmA, ymmB, ymmC, 0); cc.vpermd(ymmA, ymmB, ymmC); cc.vpermps(ymmA, ymmB, ymmC); cc.vpermpd(ymmA, ymmB, 0); cc.vpermq(ymmA, ymmB, 0); cc.vpmovmskb(gpd, ymmB); cc.vpmovsxbd(ymmA, xmmB); cc.vpmovsxbq(ymmA, xmmB); cc.vpmovsxbw(ymmA, xmmB); cc.vpmovsxdq(ymmA, xmmB); cc.vpmovsxwd(ymmA, xmmB); cc.vpmovsxwq(ymmA, xmmB); cc.vpmovzxbd(ymmA, xmmB); cc.vpmovzxbq(ymmA, xmmB); cc.vpmovzxbw(ymmA, xmmB); cc.vpmovzxdq(ymmA, xmmB); cc.vpmovzxwd(ymmA, xmmB); cc.vpmovzxwq(ymmA, xmmB); cc.vpshufd(ymmA, ymmB, 0); cc.vpshufhw(ymmA, ymmB, 0); cc.vpshuflw(ymmA, ymmB, 0); cc.vpslld(ymmA, ymmB, 0); cc.vpslldq(ymmA, ymmB, 0); cc.vpsllq(ymmA, ymmB, 0); cc.vpsllw(ymmA, ymmB, 0); cc.vpsrad(ymmA, ymmB, 0); cc.vpsraw(ymmA, ymmB, 0); cc.vpsrld(ymmA, ymmB, 0); cc.vpsrldq(ymmA, ymmB, 0); cc.vpsrlq(ymmA, ymmB, 0); cc.vpsrlw(ymmA, ymmB, 0); cc.vphaddd(ymmA, ymmB, ymmC); cc.vphaddsw(ymmA, ymmB, ymmC); cc.vphaddw(ymmA, ymmB, ymmC); cc.vphsubd(ymmA, ymmB, ymmC); cc.vphsubsw(ymmA, ymmB, ymmC); cc.vphsubw(ymmA, ymmB, ymmC); cc.vpmaddubsw(ymmA, ymmB, ymmC); cc.vpmaddwd(ymmA, ymmB, ymmC); cc.vpmaxsb(ymmA, ymmB, ymmC); cc.vpmaxsd(ymmA, ymmB, ymmC); cc.vpmaxsw(ymmA, ymmB, ymmC); cc.vpmaxub(ymmA, ymmB, ymmC); cc.vpmaxud(ymmA, ymmB, ymmC); cc.vpmaxuw(ymmA, ymmB, ymmC); cc.vpminsb(ymmA, ymmB, ymmC); cc.vpminsd(ymmA, ymmB, ymmC); cc.vpminsw(ymmA, ymmB, ymmC); cc.vpminub(ymmA, ymmB, ymmC); cc.vpminud(ymmA, ymmB, ymmC); cc.vpminuw(ymmA, ymmB, ymmC); cc.vpmuldq(ymmA, ymmB, ymmC); cc.vpmulhrsw(ymmA, ymmB, ymmC); cc.vpmulhuw(ymmA, ymmB, ymmC); cc.vpmulhw(ymmA, ymmB, ymmC); cc.vpmulld(ymmA, ymmB, ymmC); cc.vpmullw(ymmA, ymmB, ymmC); cc.vpmuludq(ymmA, ymmB, ymmC); cc.vpor(ymmA, ymmB, ymmC); cc.vpsadbw(ymmA, ymmB, ymmC); cc.vpshufb(ymmA, ymmB, ymmC); cc.vpsignb(ymmA, ymmB, ymmC); cc.vpsignd(ymmA, ymmB, ymmC); cc.vpsignw(ymmA, ymmB, ymmC); cc.vpslld(ymmA, ymmB, xmmC); cc.vpsllq(ymmA, ymmB, xmmC); cc.vpsllvd(xmmA, xmmB, xmmC); cc.vpsllvd(ymmA, ymmB, ymmC); cc.vpsllvq(xmmA, xmmB, xmmC); cc.vpsllvq(ymmA, ymmB, ymmC); cc.vpsllw(ymmA, ymmB, xmmC); cc.vpsrad(ymmA, ymmB, xmmC); cc.vpsravd(xmmA, xmmB, xmmC); cc.vpsravd(ymmA, ymmB, ymmC); cc.vpsraw(ymmA, ymmB, xmmC); cc.vpsrld(ymmA, ymmB, xmmC); cc.vpsrlq(ymmA, ymmB, xmmC); cc.vpsrlvd(xmmA, xmmB, xmmC); cc.vpsrlvd(ymmA, ymmB, ymmC); cc.vpsrlvq(xmmA, xmmB, xmmC); cc.vpsrlvq(ymmA, ymmB, ymmC); cc.vpsrlw(ymmA, ymmB, xmmC); cc.vpsubb(ymmA, ymmB, ymmC); cc.vpsubd(ymmA, ymmB, ymmC); cc.vpsubq(ymmA, ymmB, ymmC); cc.vpsubsb(ymmA, ymmB, ymmC); cc.vpsubsw(ymmA, ymmB, ymmC); cc.vpsubusb(ymmA, ymmB, ymmC); cc.vpsubusw(ymmA, ymmB, ymmC); cc.vpsubw(ymmA, ymmB, ymmC); cc.vpunpckhbw(ymmA, ymmB, ymmC); cc.vpunpckhdq(ymmA, ymmB, ymmC); cc.vpunpckhqdq(ymmA, ymmB, ymmC); cc.vpunpckhwd(ymmA, ymmB, ymmC); cc.vpunpcklbw(ymmA, ymmB, ymmC); cc.vpunpckldq(ymmA, ymmB, ymmC); cc.vpunpcklqdq(ymmA, ymmB, ymmC); cc.vpunpcklwd(ymmA, ymmB, ymmC); cc.vpxor(ymmA, ymmB, ymmC); // FMA. cc.vfmadd132pd(xmmA, xmmB, xmmC); cc.vfmadd132pd(ymmA, ymmB, ymmC); cc.vfmadd132ps(xmmA, xmmB, xmmC); cc.vfmadd132ps(ymmA, ymmB, ymmC); cc.vfmadd132sd(xmmA, xmmB, xmmC); cc.vfmadd132ss(xmmA, xmmB, xmmC); cc.vfmadd213pd(xmmA, xmmB, xmmC); cc.vfmadd213pd(ymmA, ymmB, ymmC); cc.vfmadd213ps(xmmA, xmmB, xmmC); cc.vfmadd213ps(ymmA, ymmB, ymmC); cc.vfmadd213sd(xmmA, xmmB, xmmC); cc.vfmadd213ss(xmmA, xmmB, xmmC); cc.vfmadd231pd(xmmA, xmmB, xmmC); cc.vfmadd231pd(ymmA, ymmB, ymmC); cc.vfmadd231ps(xmmA, xmmB, xmmC); cc.vfmadd231ps(ymmA, ymmB, ymmC); cc.vfmadd231sd(xmmA, xmmB, xmmC); cc.vfmadd231ss(xmmA, xmmB, xmmC); cc.vfmaddsub132pd(xmmA, xmmB, xmmC); cc.vfmaddsub132pd(ymmA, ymmB, ymmC); cc.vfmaddsub132ps(xmmA, xmmB, xmmC); cc.vfmaddsub132ps(ymmA, ymmB, ymmC); cc.vfmaddsub213pd(xmmA, xmmB, xmmC); cc.vfmaddsub213pd(ymmA, ymmB, ymmC); cc.vfmaddsub213ps(xmmA, xmmB, xmmC); cc.vfmaddsub213ps(ymmA, ymmB, ymmC); cc.vfmaddsub231pd(xmmA, xmmB, xmmC); cc.vfmaddsub231pd(ymmA, ymmB, ymmC); cc.vfmaddsub231ps(xmmA, xmmB, xmmC); cc.vfmaddsub231ps(ymmA, ymmB, ymmC); cc.vfmsub132pd(xmmA, xmmB, xmmC); cc.vfmsub132pd(ymmA, ymmB, ymmC); cc.vfmsub132ps(xmmA, xmmB, xmmC); cc.vfmsub132ps(ymmA, ymmB, ymmC); cc.vfmsub132sd(xmmA, xmmB, xmmC); cc.vfmsub132ss(xmmA, xmmB, xmmC); cc.vfmsub213pd(xmmA, xmmB, xmmC); cc.vfmsub213pd(ymmA, ymmB, ymmC); cc.vfmsub213ps(xmmA, xmmB, xmmC); cc.vfmsub213ps(ymmA, ymmB, ymmC); cc.vfmsub213sd(xmmA, xmmB, xmmC); cc.vfmsub213ss(xmmA, xmmB, xmmC); cc.vfmsub231pd(xmmA, xmmB, xmmC); cc.vfmsub231pd(ymmA, ymmB, ymmC); cc.vfmsub231ps(xmmA, xmmB, xmmC); cc.vfmsub231ps(ymmA, ymmB, ymmC); cc.vfmsub231sd(xmmA, xmmB, xmmC); cc.vfmsub231ss(xmmA, xmmB, xmmC); cc.vfmsubadd132pd(xmmA, xmmB, xmmC); cc.vfmsubadd132pd(ymmA, ymmB, ymmC); cc.vfmsubadd132ps(xmmA, xmmB, xmmC); cc.vfmsubadd132ps(ymmA, ymmB, ymmC); cc.vfmsubadd213pd(xmmA, xmmB, xmmC); cc.vfmsubadd213pd(ymmA, ymmB, ymmC); cc.vfmsubadd213ps(xmmA, xmmB, xmmC); cc.vfmsubadd213ps(ymmA, ymmB, ymmC); cc.vfmsubadd231pd(xmmA, xmmB, xmmC); cc.vfmsubadd231pd(ymmA, ymmB, ymmC); cc.vfmsubadd231ps(xmmA, xmmB, xmmC); cc.vfmsubadd231ps(ymmA, ymmB, ymmC); cc.vfnmadd132pd(xmmA, xmmB, xmmC); cc.vfnmadd132pd(ymmA, ymmB, ymmC); cc.vfnmadd132ps(xmmA, xmmB, xmmC); cc.vfnmadd132ps(ymmA, ymmB, ymmC); cc.vfnmadd132sd(xmmA, xmmB, xmmC); cc.vfnmadd132ss(xmmA, xmmB, xmmC); cc.vfnmadd213pd(xmmA, xmmB, xmmC); cc.vfnmadd213pd(ymmA, ymmB, ymmC); cc.vfnmadd213ps(xmmA, xmmB, xmmC); cc.vfnmadd213ps(ymmA, ymmB, ymmC); cc.vfnmadd213sd(xmmA, xmmB, xmmC); cc.vfnmadd213ss(xmmA, xmmB, xmmC); cc.vfnmadd231pd(xmmA, xmmB, xmmC); cc.vfnmadd231pd(ymmA, ymmB, ymmC); cc.vfnmadd231ps(xmmA, xmmB, xmmC); cc.vfnmadd231ps(ymmA, ymmB, ymmC); cc.vfnmadd231sd(xmmA, xmmB, xmmC); cc.vfnmadd231ss(xmmA, xmmB, xmmC); cc.vfnmsub132pd(xmmA, xmmB, xmmC); cc.vfnmsub132pd(ymmA, ymmB, ymmC); cc.vfnmsub132ps(xmmA, xmmB, xmmC); cc.vfnmsub132ps(ymmA, ymmB, ymmC); cc.vfnmsub132sd(xmmA, xmmB, xmmC); cc.vfnmsub132ss(xmmA, xmmB, xmmC); cc.vfnmsub213pd(xmmA, xmmB, xmmC); cc.vfnmsub213pd(ymmA, ymmB, ymmC); cc.vfnmsub213ps(xmmA, xmmB, xmmC); cc.vfnmsub213ps(ymmA, ymmB, ymmC); cc.vfnmsub213sd(xmmA, xmmB, xmmC); cc.vfnmsub213ss(xmmA, xmmB, xmmC); cc.vfnmsub231pd(xmmA, xmmB, xmmC); cc.vfnmsub231pd(ymmA, ymmB, ymmC); cc.vfnmsub231ps(xmmA, xmmB, xmmC); cc.vfnmsub231ps(ymmA, ymmB, ymmC); cc.vfnmsub231sd(xmmA, xmmB, xmmC); cc.vfnmsub231ss(xmmA, xmmB, xmmC); } // Generates a long sequence of AVX instructions. template static void generateAvxSequenceInternalRegMem( Emitter& cc, const x86::Gp& gp, const x86::Vec& vecA, const x86::Vec& vecB, const x86::Vec& vecC, const x86::Vec& vecD) { x86::Gp gpd = gp.r32(); x86::Gp gpq = gp.r64(); x86::Gp gpz = cc.is32Bit() ? gpd : gpq; x86::Xmm xmmA = vecA.xmm(); x86::Xmm xmmB = vecB.xmm(); x86::Xmm xmmC = vecC.xmm(); x86::Xmm xmmD = vecD.xmm(); x86::Ymm ymmA = vecA.ymm(); x86::Ymm ymmB = vecB.ymm(); x86::Ymm ymmC = vecC.ymm(); x86::Ymm ymmD = vecD.ymm(); x86::Mem m = x86::ptr(gpz); x86::Mem m128 = x86::xmmword_ptr(gpz); x86::Mem m256 = x86::xmmword_ptr(gpz); x86::Mem vx_ptr = x86::ptr(gpz, xmmD); x86::Mem vy_ptr = x86::ptr(gpz, ymmD); cc.xor_(gpd, gpd); cc.vxorps(xmmA, xmmA, xmmA); cc.vxorps(xmmB, xmmB, xmmB); cc.vxorps(xmmC, xmmC, xmmC); cc.vxorps(xmmD, xmmD, xmmD); cc.vaddpd(xmmA, xmmB, m); cc.vaddpd(ymmA, ymmB, m); cc.vaddps(xmmA, xmmB, m); cc.vaddps(ymmA, ymmB, m); cc.vaddsd(xmmA, xmmB, m); cc.vaddss(xmmA, xmmB, m); cc.vaddsubpd(xmmA, xmmB, m); cc.vaddsubpd(ymmA, ymmB, m); cc.vaddsubps(xmmA, xmmB, m); cc.vaddsubps(ymmA, ymmB, m); cc.vandpd(xmmA, xmmB, m); cc.vandpd(ymmA, ymmB, m); cc.vandps(xmmA, xmmB, m); cc.vandps(ymmA, ymmB, m); cc.vandnpd(xmmA, xmmB, m); cc.vandnpd(ymmA, ymmB, m); cc.vandnps(xmmA, xmmB, m); cc.vandnps(ymmA, ymmB, m); cc.vblendpd(xmmA, xmmB, m, 0); cc.vblendpd(ymmA, ymmB, m, 0); cc.vblendps(xmmA, xmmB, m, 0); cc.vblendps(ymmA, ymmB, m, 0); cc.vblendvpd(xmmA, xmmB, m, xmmA); cc.vblendvpd(ymmA, ymmB, m, ymmA); cc.vbroadcastf128(ymmA, m); cc.vbroadcastsd(ymmA, m); cc.vbroadcastss(xmmA, m); cc.vbroadcastss(ymmA, m); cc.vcmppd(xmmA, xmmB, m, 0); cc.vcmppd(ymmA, ymmB, m, 0); cc.vcmpps(xmmA, xmmB, m, 0); cc.vcmpps(ymmA, ymmB, m, 0); cc.vcmpsd(xmmA, xmmB, m, 0); cc.vcmpss(xmmA, xmmB, m, 0); cc.vcomisd(xmmA, m); cc.vcomiss(xmmA, m); cc.vcvtdq2pd(xmmA, m); cc.vcvtdq2pd(ymmA, m); cc.vcvtdq2ps(xmmA, m); cc.vcvtdq2ps(ymmA, m); cc.vcvtpd2dq(xmmA, m128); cc.vcvtpd2dq(xmmA, m256); cc.vcvtpd2ps(xmmA, m128); cc.vcvtpd2ps(xmmA, m256); cc.vcvtps2dq(xmmA, m); cc.vcvtps2dq(ymmA, m); cc.vcvtps2pd(xmmA, m); cc.vcvtps2pd(ymmA, m); cc.vcvtsd2si(gpd, m); cc.vcvtsd2ss(xmmA, xmmB, m); cc.vcvtsi2sd(xmmA, xmmB, m); cc.vcvtsi2ss(xmmA, xmmB, m); cc.vcvtss2sd(xmmA, xmmB, m); cc.vcvtss2si(gpd, m); cc.vcvttpd2dq(xmmA, m128); cc.vcvttpd2dq(xmmA, m256); cc.vcvttps2dq(xmmA, m); cc.vcvttps2dq(ymmA, m); cc.vcvttsd2si(gpd, m); cc.vcvttss2si(gpd, m); cc.vdivpd(xmmA, xmmB, m); cc.vdivpd(ymmA, ymmB, m); cc.vdivps(xmmA, xmmB, m); cc.vdivps(ymmA, ymmB, m); cc.vdivsd(xmmA, xmmB, m); cc.vdivss(xmmA, xmmB, m); cc.vdppd(xmmA, xmmB, m, 0); cc.vdpps(xmmA, xmmB, m, 0); cc.vdpps(ymmA, ymmB, m, 0); cc.vextractf128(m, ymmB, 0); cc.vextractps(m, xmmB, 0); cc.vhaddpd(xmmA, xmmB, m); cc.vhaddpd(ymmA, ymmB, m); cc.vhaddps(xmmA, xmmB, m); cc.vhaddps(ymmA, ymmB, m); cc.vhsubpd(xmmA, xmmB, m); cc.vhsubpd(ymmA, ymmB, m); cc.vhsubps(xmmA, xmmB, m); cc.vhsubps(ymmA, ymmB, m); cc.vinsertf128(ymmA, ymmB, m, 0); cc.vinsertps(xmmA, xmmB, m, 0); cc.vlddqu(xmmA, m); cc.vlddqu(ymmA, m); cc.vmaskmovps(xmmA, xmmB, m); cc.vmaskmovps(ymmA, ymmB, m); cc.vmaskmovps(m, xmmB, xmmC); cc.vmaskmovps(m, ymmB, ymmC); cc.vmaskmovpd(xmmA, xmmB, m); cc.vmaskmovpd(ymmA, ymmB, m); cc.vmaskmovpd(m, xmmB, xmmC); cc.vmaskmovpd(m, ymmB, ymmC); cc.vmaxpd(xmmA, xmmB, m); cc.vmaxpd(ymmA, ymmB, m); cc.vmaxps(xmmA, xmmB, m); cc.vmaxps(ymmA, ymmB, m); cc.vmaxsd(xmmA, xmmB, m); cc.vmaxss(xmmA, xmmB, m); cc.vminpd(xmmA, xmmB, m); cc.vminpd(ymmA, ymmB, m); cc.vminps(xmmA, xmmB, m); cc.vminps(ymmA, ymmB, m); cc.vminsd(xmmA, xmmB, m); cc.vminss(xmmA, xmmB, m); cc.vmovapd(xmmA, m); cc.vmovapd(m, xmmB); cc.vmovapd(ymmA, m); cc.vmovapd(m, ymmB); cc.vmovaps(xmmA, m); cc.vmovaps(m, xmmB); cc.vmovaps(ymmA, m); cc.vmovaps(m, ymmB); cc.vmovd(xmmA, m); cc.vmovd(m, xmmB); cc.vmovddup(xmmA, m); cc.vmovddup(ymmA, m); cc.vmovdqa(xmmA, m); cc.vmovdqa(m, xmmB); cc.vmovdqa(ymmA, m); cc.vmovdqa(m, ymmB); cc.vmovdqu(xmmA, m); cc.vmovdqu(m, xmmB); cc.vmovdqu(ymmA, m); cc.vmovdqu(m, ymmB); cc.vmovhpd(xmmA, xmmB, m); cc.vmovhps(xmmA, xmmB, m); cc.vmovhps(m, xmmB); cc.vmovlpd(xmmA, xmmB, m); cc.vmovlpd(m, xmmB); cc.vmovlps(xmmA, xmmB, m); cc.vmovlps(m, xmmB); cc.vmovntdq(m, xmmB); cc.vmovntdq(m, ymmB); cc.vmovntdqa(xmmA, m); cc.vmovntpd(m, xmmB); cc.vmovntpd(m, ymmB); cc.vmovntps(m, xmmB); cc.vmovntps(m, ymmB); cc.vmovsd(xmmA, m); cc.vmovsd(m, xmmB); cc.vmovshdup(xmmA, m); cc.vmovshdup(ymmA, m); cc.vmovsldup(xmmA, m); cc.vmovsldup(ymmA, m); cc.vmovss(xmmA, m); cc.vmovss(m, xmmB); cc.vmovupd(xmmA, m); cc.vmovupd(m, xmmB); cc.vmovupd(ymmA, m); cc.vmovupd(m, ymmB); cc.vmovups(xmmA, m); cc.vmovups(m, xmmB); cc.vmovups(ymmA, m); cc.vmovups(m, ymmB); cc.vmpsadbw(xmmA, xmmB, m, 0); cc.vmulpd(xmmA, xmmB, m); cc.vmulpd(ymmA, ymmB, m); cc.vmulps(xmmA, xmmB, m); cc.vmulps(ymmA, ymmB, m); cc.vmulsd(xmmA, xmmB, m); cc.vmulss(xmmA, xmmB, m); cc.vorpd(xmmA, xmmB, m); cc.vorpd(ymmA, ymmB, m); cc.vorps(xmmA, xmmB, m); cc.vorps(ymmA, ymmB, m); cc.vpabsb(xmmA, m); cc.vpabsd(xmmA, m); cc.vpabsw(xmmA, m); cc.vpackssdw(xmmA, xmmB, m); cc.vpacksswb(xmmA, xmmB, m); cc.vpackusdw(xmmA, xmmB, m); cc.vpackuswb(xmmA, xmmB, m); cc.vpaddb(xmmA, xmmB, m); cc.vpaddd(xmmA, xmmB, m); cc.vpaddq(xmmA, xmmB, m); cc.vpaddw(xmmA, xmmB, m); cc.vpaddsb(xmmA, xmmB, m); cc.vpaddsw(xmmA, xmmB, m); cc.vpaddusb(xmmA, xmmB, m); cc.vpaddusw(xmmA, xmmB, m); cc.vpalignr(xmmA, xmmB, m, 0); cc.vpand(xmmA, xmmB, m); cc.vpandn(xmmA, xmmB, m); cc.vpavgb(xmmA, xmmB, m); cc.vpavgw(xmmA, xmmB, m); cc.vpblendvb(xmmA, xmmB, m, xmmA); cc.vpblendw(xmmA, xmmB, m, 0); cc.vpcmpeqb(xmmA, xmmB, m); cc.vpcmpeqd(xmmA, xmmB, m); cc.vpcmpeqq(xmmA, xmmB, m); cc.vpcmpeqw(xmmA, xmmB, m); cc.vpcmpgtb(xmmA, xmmB, m); cc.vpcmpgtd(xmmA, xmmB, m); cc.vpcmpgtq(xmmA, xmmB, m); cc.vpcmpgtw(xmmA, xmmB, m); cc.vpermilpd(xmmA, xmmB, m); cc.vpermilpd(ymmA, ymmB, m); cc.vpermilpd(xmmA, m, 0); cc.vpermilpd(ymmA, m, 0); cc.vpermilps(xmmA, xmmB, m); cc.vpermilps(ymmA, ymmB, m); cc.vpermilps(xmmA, m, 0); cc.vpermilps(ymmA, m, 0); cc.vperm2f128(ymmA, ymmB, m, 0); cc.vpextrb(m, xmmB, 0); cc.vpextrd(m, xmmB, 0); if (cc.is64Bit()) cc.vpextrq(m, xmmB, 0); cc.vpextrw(m, xmmB, 0); cc.vphaddd(xmmA, xmmB, m); cc.vphaddsw(xmmA, xmmB, m); cc.vphaddw(xmmA, xmmB, m); cc.vphminposuw(xmmA, m); cc.vphsubd(xmmA, xmmB, m); cc.vphsubsw(xmmA, xmmB, m); cc.vphsubw(xmmA, xmmB, m); cc.vpinsrb(xmmA, xmmB, m, 0); cc.vpinsrd(xmmA, xmmB, m, 0); cc.vpinsrw(xmmA, xmmB, m, 0); cc.vpmaddubsw(xmmA, xmmB, m); cc.vpmaddwd(xmmA, xmmB, m); cc.vpmaxsb(xmmA, xmmB, m); cc.vpmaxsd(xmmA, xmmB, m); cc.vpmaxsw(xmmA, xmmB, m); cc.vpmaxub(xmmA, xmmB, m); cc.vpmaxud(xmmA, xmmB, m); cc.vpmaxuw(xmmA, xmmB, m); cc.vpminsb(xmmA, xmmB, m); cc.vpminsd(xmmA, xmmB, m); cc.vpminsw(xmmA, xmmB, m); cc.vpminub(xmmA, xmmB, m); cc.vpminud(xmmA, xmmB, m); cc.vpminuw(xmmA, xmmB, m); cc.vpmovsxbd(xmmA, m); cc.vpmovsxbq(xmmA, m); cc.vpmovsxbw(xmmA, m); cc.vpmovsxdq(xmmA, m); cc.vpmovsxwd(xmmA, m); cc.vpmovsxwq(xmmA, m); cc.vpmovzxbd(xmmA, m); cc.vpmovzxbq(xmmA, m); cc.vpmovzxbw(xmmA, m); cc.vpmovzxdq(xmmA, m); cc.vpmovzxwd(xmmA, m); cc.vpmovzxwq(xmmA, m); cc.vpmuldq(xmmA, xmmB, m); cc.vpmulhrsw(xmmA, xmmB, m); cc.vpmulhuw(xmmA, xmmB, m); cc.vpmulhw(xmmA, xmmB, m); cc.vpmulld(xmmA, xmmB, m); cc.vpmullw(xmmA, xmmB, m); cc.vpmuludq(xmmA, xmmB, m); cc.vpor(xmmA, xmmB, m); cc.vpsadbw(xmmA, xmmB, m); cc.vpshufb(xmmA, xmmB, m); cc.vpshufd(xmmA, m, 0); cc.vpshufhw(xmmA, m, 0); cc.vpshuflw(xmmA, m, 0); cc.vpsignb(xmmA, xmmB, m); cc.vpsignd(xmmA, xmmB, m); cc.vpsignw(xmmA, xmmB, m); cc.vpslld(xmmA, xmmB, m); cc.vpsllq(xmmA, xmmB, m); cc.vpsllw(xmmA, xmmB, m); cc.vpsrad(xmmA, xmmB, m); cc.vpsraw(xmmA, xmmB, m); cc.vpsrld(xmmA, xmmB, m); cc.vpsrlq(xmmA, xmmB, m); cc.vpsrlw(xmmA, xmmB, m); cc.vpsubb(xmmA, xmmB, m); cc.vpsubd(xmmA, xmmB, m); cc.vpsubq(xmmA, xmmB, m); cc.vpsubw(xmmA, xmmB, m); cc.vpsubsb(xmmA, xmmB, m); cc.vpsubsw(xmmA, xmmB, m); cc.vpsubusb(xmmA, xmmB, m); cc.vpsubusw(xmmA, xmmB, m); cc.vptest(xmmA, m); cc.vptest(ymmA, m); cc.vpunpckhbw(xmmA, xmmB, m); cc.vpunpckhdq(xmmA, xmmB, m); cc.vpunpckhqdq(xmmA, xmmB, m); cc.vpunpckhwd(xmmA, xmmB, m); cc.vpunpcklbw(xmmA, xmmB, m); cc.vpunpckldq(xmmA, xmmB, m); cc.vpunpcklqdq(xmmA, xmmB, m); cc.vpunpcklwd(xmmA, xmmB, m); cc.vpxor(xmmA, xmmB, m); cc.vrcpps(xmmA, m); cc.vrcpps(ymmA, m); cc.vrcpss(xmmA, xmmB, m); cc.vrsqrtps(xmmA, m); cc.vrsqrtps(ymmA, m); cc.vrsqrtss(xmmA, xmmB, m); cc.vroundpd(xmmA, m, 0); cc.vroundpd(ymmA, m, 0); cc.vroundps(xmmA, m, 0); cc.vroundps(ymmA, m, 0); cc.vroundsd(xmmA, xmmB, m, 0); cc.vroundss(xmmA, xmmB, m, 0); cc.vshufpd(xmmA, xmmB, m, 0); cc.vshufpd(ymmA, ymmB, m, 0); cc.vshufps(xmmA, xmmB, m, 0); cc.vshufps(ymmA, ymmB, m, 0); cc.vsqrtpd(xmmA, m); cc.vsqrtpd(ymmA, m); cc.vsqrtps(xmmA, m); cc.vsqrtps(ymmA, m); cc.vsqrtsd(xmmA, xmmB, m); cc.vsqrtss(xmmA, xmmB, m); cc.vsubpd(xmmA, xmmB, m); cc.vsubpd(ymmA, ymmB, m); cc.vsubps(xmmA, xmmB, m); cc.vsubps(ymmA, ymmB, m); cc.vsubsd(xmmA, xmmB, m); cc.vsubss(xmmA, xmmB, m); cc.vtestps(xmmA, m); cc.vtestps(ymmA, m); cc.vtestpd(xmmA, m); cc.vtestpd(ymmA, m); cc.vucomisd(xmmA, m); cc.vucomiss(xmmA, m); cc.vunpckhpd(xmmA, xmmB, m); cc.vunpckhpd(ymmA, ymmB, m); cc.vunpckhps(xmmA, xmmB, m); cc.vunpckhps(ymmA, ymmB, m); cc.vunpcklpd(xmmA, xmmB, m); cc.vunpcklpd(ymmA, ymmB, m); cc.vunpcklps(xmmA, xmmB, m); cc.vunpcklps(ymmA, ymmB, m); cc.vxorpd(xmmA, xmmB, m); cc.vxorpd(ymmA, ymmB, m); cc.vxorps(xmmA, xmmB, m); cc.vxorps(ymmA, ymmB, m); // AVX+AESNI. cc.vaesdec(xmmA, xmmB, m); cc.vaesdeclast(xmmA, xmmB, m); cc.vaesenc(xmmA, xmmB, m); cc.vaesenclast(xmmA, xmmB, m); cc.vaesimc(xmmA, m); cc.vaeskeygenassist(xmmA, m, 0); // AVX+PCLMULQDQ. cc.vpclmulqdq(xmmA, xmmB, m, 0); // AVX2. cc.vbroadcasti128(ymmA, m); cc.vextracti128(m, ymmB, 0); cc.vgatherdpd(xmmA, vx_ptr, xmmC); cc.vgatherdpd(ymmA, vx_ptr, ymmC); cc.vgatherdps(xmmA, vx_ptr, xmmC); cc.vgatherdps(ymmA, vy_ptr, ymmC); cc.vgatherqpd(xmmA, vx_ptr, xmmC); cc.vgatherqpd(ymmA, vy_ptr, ymmC); cc.vgatherqps(xmmA, vx_ptr, xmmC); cc.vgatherqps(xmmA, vy_ptr, xmmC); cc.vinserti128(ymmA, ymmB, m, 0); cc.vmovntdqa(ymmA, m); cc.vmpsadbw(ymmA, ymmB, m, 0); cc.vpabsb(ymmA, m); cc.vpabsd(ymmA, m); cc.vpabsw(ymmA, m); cc.vpackssdw(ymmA, ymmB, m); cc.vpacksswb(ymmA, ymmB, m); cc.vpackusdw(ymmA, ymmB, m); cc.vpackuswb(ymmA, ymmB, m); cc.vpaddb(ymmA, ymmB, m); cc.vpaddd(ymmA, ymmB, m); cc.vpaddq(ymmA, ymmB, m); cc.vpaddw(ymmA, ymmB, m); cc.vpaddsb(ymmA, ymmB, m); cc.vpaddsw(ymmA, ymmB, m); cc.vpaddusb(ymmA, ymmB, m); cc.vpaddusw(ymmA, ymmB, m); cc.vpalignr(ymmA, ymmB, m, 0); cc.vpand(ymmA, ymmB, m); cc.vpandn(ymmA, ymmB, m); cc.vpavgb(ymmA, ymmB, m); cc.vpavgw(ymmA, ymmB, m); cc.vpblendd(xmmA, xmmB, m, 0); cc.vpblendd(ymmA, ymmB, m, 0); cc.vpblendvb(ymmA, ymmB, m, ymmA); cc.vpblendw(ymmA, ymmB, m, 0); cc.vpbroadcastb(xmmA, m); cc.vpbroadcastb(ymmA, m); cc.vpbroadcastd(xmmA, m); cc.vpbroadcastd(ymmA, m); cc.vpbroadcastq(xmmA, m); cc.vpbroadcastq(ymmA, m); cc.vpbroadcastw(xmmA, m); cc.vpbroadcastw(ymmA, m); cc.vpcmpeqb(ymmA, ymmB, m); cc.vpcmpeqd(ymmA, ymmB, m); cc.vpcmpeqq(ymmA, ymmB, m); cc.vpcmpeqw(ymmA, ymmB, m); cc.vpcmpgtb(ymmA, ymmB, m); cc.vpcmpgtd(ymmA, ymmB, m); cc.vpcmpgtq(ymmA, ymmB, m); cc.vpcmpgtw(ymmA, ymmB, m); cc.vperm2i128(ymmA, ymmB, m, 0); cc.vpermd(ymmA, ymmB, m); cc.vpermps(ymmA, ymmB, m); cc.vpermpd(ymmA, m, 0); cc.vpermq(ymmA, m, 0); cc.vpgatherdd(xmmA, vx_ptr, xmmC); cc.vpgatherdd(ymmA, vy_ptr, ymmC); cc.vpgatherdq(xmmA, vx_ptr, xmmC); cc.vpgatherdq(ymmA, vx_ptr, ymmC); cc.vpgatherqd(xmmA, vx_ptr, xmmC); cc.vpgatherqd(xmmA, vy_ptr, xmmC); cc.vpgatherqq(xmmA, vx_ptr, xmmC); cc.vpgatherqq(ymmA, vy_ptr, ymmC); cc.vpmovsxbd(ymmA, m); cc.vpmovsxbq(ymmA, m); cc.vpmovsxbw(ymmA, m); cc.vpmovsxdq(ymmA, m); cc.vpmovsxwd(ymmA, m); cc.vpmovsxwq(ymmA, m); cc.vpmovzxbd(ymmA, m); cc.vpmovzxbq(ymmA, m); cc.vpmovzxbw(ymmA, m); cc.vpmovzxdq(ymmA, m); cc.vpmovzxwd(ymmA, m); cc.vpmovzxwq(ymmA, m); cc.vpshufd(ymmA, m, 0); cc.vpshufhw(ymmA, m, 0); cc.vpshuflw(ymmA, m, 0); cc.vphaddd(ymmA, ymmB, m); cc.vphaddsw(ymmA, ymmB, m); cc.vphaddw(ymmA, ymmB, m); cc.vphsubd(ymmA, ymmB, m); cc.vphsubsw(ymmA, ymmB, m); cc.vphsubw(ymmA, ymmB, m); cc.vpmaddubsw(ymmA, ymmB, m); cc.vpmaddwd(ymmA, ymmB, m); cc.vpmaskmovd(m, xmmB, xmmC); cc.vpmaskmovd(m, ymmB, ymmC); cc.vpmaskmovd(xmmA, xmmB, m); cc.vpmaskmovd(ymmA, ymmB, m); cc.vpmaskmovq(m, xmmB, xmmC); cc.vpmaskmovq(m, ymmB, ymmC); cc.vpmaskmovq(xmmA, xmmB, m); cc.vpmaskmovq(ymmA, ymmB, m); cc.vpmaxsb(ymmA, ymmB, m); cc.vpmaxsd(ymmA, ymmB, m); cc.vpmaxsw(ymmA, ymmB, m); cc.vpmaxub(ymmA, ymmB, m); cc.vpmaxud(ymmA, ymmB, m); cc.vpmaxuw(ymmA, ymmB, m); cc.vpminsb(ymmA, ymmB, m); cc.vpminsd(ymmA, ymmB, m); cc.vpminsw(ymmA, ymmB, m); cc.vpminub(ymmA, ymmB, m); cc.vpminud(ymmA, ymmB, m); cc.vpminuw(ymmA, ymmB, m); cc.vpmuldq(ymmA, ymmB, m); cc.vpmulhrsw(ymmA, ymmB, m); cc.vpmulhuw(ymmA, ymmB, m); cc.vpmulhw(ymmA, ymmB, m); cc.vpmulld(ymmA, ymmB, m); cc.vpmullw(ymmA, ymmB, m); cc.vpmuludq(ymmA, ymmB, m); cc.vpor(ymmA, ymmB, m); cc.vpsadbw(ymmA, ymmB, m); cc.vpshufb(ymmA, ymmB, m); cc.vpsignb(ymmA, ymmB, m); cc.vpsignd(ymmA, ymmB, m); cc.vpsignw(ymmA, ymmB, m); cc.vpslld(ymmA, ymmB, m); cc.vpsllq(ymmA, ymmB, m); cc.vpsllvd(xmmA, xmmB, m); cc.vpsllvd(ymmA, ymmB, m); cc.vpsllvq(xmmA, xmmB, m); cc.vpsllvq(ymmA, ymmB, m); cc.vpsllw(ymmA, ymmB, m); cc.vpsrad(ymmA, ymmB, m); cc.vpsravd(xmmA, xmmB, m); cc.vpsravd(ymmA, ymmB, m); cc.vpsraw(ymmA, ymmB, m); cc.vpsrld(ymmA, ymmB, m); cc.vpsrlq(ymmA, ymmB, m); cc.vpsrlvd(xmmA, xmmB, m); cc.vpsrlvd(ymmA, ymmB, m); cc.vpsrlvq(xmmA, xmmB, m); cc.vpsrlvq(ymmA, ymmB, m); cc.vpsrlw(ymmA, ymmB, m); cc.vpsubb(ymmA, ymmB, m); cc.vpsubd(ymmA, ymmB, m); cc.vpsubq(ymmA, ymmB, m); cc.vpsubsb(ymmA, ymmB, m); cc.vpsubsw(ymmA, ymmB, m); cc.vpsubusb(ymmA, ymmB, m); cc.vpsubusw(ymmA, ymmB, m); cc.vpsubw(ymmA, ymmB, m); cc.vpunpckhbw(ymmA, ymmB, m); cc.vpunpckhdq(ymmA, ymmB, m); cc.vpunpckhqdq(ymmA, ymmB, m); cc.vpunpckhwd(ymmA, ymmB, m); cc.vpunpcklbw(ymmA, ymmB, m); cc.vpunpckldq(ymmA, ymmB, m); cc.vpunpcklqdq(ymmA, ymmB, m); cc.vpunpcklwd(ymmA, ymmB, m); cc.vpxor(ymmA, ymmB, m); } // Generates a long sequence of AVX instructions. template static void generateAvxSequenceInternal( Emitter& cc, InstForm form, const x86::Gp& gp, const x86::Vec& vecA, const x86::Vec& vecB, const x86::Vec& vecC, const x86::Vec& vecD) { if (form == InstForm::kReg) generateAvxSequenceInternalRegOnly(cc, gp, vecA, vecB, vecC, vecD); else generateAvxSequenceInternalRegMem(cc, gp, vecA, vecB, vecC, vecD); } static void generateAvxSequence(BaseEmitter& emitter, InstForm form, bool emitPrologEpilog) { using namespace asmjit::x86; if (emitter.isAssembler()) { Assembler& cc = *emitter.as(); if (emitPrologEpilog) { FuncDetail func; func.init(FuncSignature::build(), cc.environment()); FuncFrame frame; frame.init(func); frame.addDirtyRegs(eax, ymm0, ymm1, ymm2, ymm3); frame.finalize(); cc.emitProlog(frame); generateAvxSequenceInternal(cc, form, eax, ymm0, ymm1, ymm2, ymm3); cc.emitEpilog(frame); } else { generateAvxSequenceInternal(cc, form, eax, ymm0, ymm1, ymm2, ymm3); } } #ifndef ASMJIT_NO_BUILDER else if (emitter.isBuilder()) { Builder& cc = *emitter.as(); if (emitPrologEpilog) { FuncDetail func; func.init(FuncSignature::build(), cc.environment()); FuncFrame frame; frame.init(func); frame.addDirtyRegs(eax, ymm0, ymm1, ymm2, ymm3); frame.finalize(); cc.emitProlog(frame); generateAvxSequenceInternal(cc, form, eax, ymm0, ymm1, ymm2, ymm3); cc.emitEpilog(frame); } else { generateAvxSequenceInternal(cc, form, eax, ymm0, ymm1, ymm2, ymm3); } } #endif #ifndef ASMJIT_NO_COMPILER else if (emitter.isCompiler()) { Compiler& cc = *emitter.as(); Gp gp = cc.newGpz("gp"); Ymm a = cc.newYmm("a"); Ymm b = cc.newYmm("b"); Ymm c = cc.newYmm("c"); Ymm d = cc.newYmm("d"); cc.addFunc(FuncSignature::build()); generateAvxSequenceInternal(cc, form, gp, a, b, c, d); cc.endFunc(); } #endif } // Generates a long sequence of AVX512 instructions. template static void generateAvx512SequenceInternalRegOnly( Emitter& cc, const x86::Gp& gp, const x86::KReg& kA, const x86::KReg& kB, const x86::KReg& kC, const x86::Vec& vecA, const x86::Vec& vecB, const x86::Vec& vecC, const x86::Vec& vecD) { x86::Gp gpd = gp.r32(); x86::Gp gpq = gp.r64(); x86::Gp gpz = cc.is32Bit() ? gpd : gpq; x86::Xmm xmmA = vecA.xmm(); x86::Xmm xmmB = vecB.xmm(); x86::Xmm xmmC = vecC.xmm(); x86::Xmm xmmD = vecD.xmm(); x86::Ymm ymmA = vecA.ymm(); x86::Ymm ymmB = vecB.ymm(); x86::Ymm ymmC = vecC.ymm(); x86::Zmm zmmA = vecA.zmm(); x86::Zmm zmmB = vecB.zmm(); x86::Zmm zmmC = vecC.zmm(); cc.xor_(gpd, gpd); cc.vxorps(xmmA, xmmA, xmmA); cc.vxorps(xmmB, xmmB, xmmB); cc.vxorps(xmmC, xmmC, xmmC); cc.vxorps(xmmD, xmmD, xmmD); cc.kaddb(kA, kB, kC); cc.kaddd(kA, kB, kC); cc.kaddq(kA, kB, kC); cc.kaddw(kA, kB, kC); cc.kandb(kA, kB, kC); cc.kandd(kA, kB, kC); cc.kandnb(kA, kB, kC); cc.kandnd(kA, kB, kC); cc.kandnq(kA, kB, kC); cc.kandnw(kA, kB, kC); cc.kandq(kA, kB, kC); cc.kandw(kA, kB, kC); cc.kmovb(kA, kB); cc.kmovb(kA, gpd); cc.kmovb(gpd, kB); cc.kmovd(kA, kB); cc.kmovd(kA, gpd); cc.kmovd(gpd, kB); cc.kmovq(kA, kB); if (cc.is64Bit()) cc.kmovq(kA, gpq); if (cc.is64Bit()) cc.kmovq(gpq, kB); cc.kmovw(kA, kB); cc.kmovw(kA, gpd); cc.kmovw(gpd, kB); cc.knotb(kA, kB); cc.knotd(kA, kB); cc.knotq(kA, kB); cc.knotw(kA, kB); cc.korb(kA, kB, kC); cc.kord(kA, kB, kC); cc.korq(kA, kB, kC); cc.kortestb(kA, kB); cc.kortestd(kA, kB); cc.kortestq(kA, kB); cc.kortestw(kA, kB); cc.korw(kA, kB, kC); cc.kshiftlb(kA, kB, 0); cc.kshiftld(kA, kB, 0); cc.kshiftlq(kA, kB, 0); cc.kshiftlw(kA, kB, 0); cc.kshiftrb(kA, kB, 0); cc.kshiftrd(kA, kB, 0); cc.kshiftrq(kA, kB, 0); cc.kshiftrw(kA, kB, 0); cc.ktestb(kA, kB); cc.ktestd(kA, kB); cc.ktestq(kA, kB); cc.ktestw(kA, kB); cc.kunpckbw(kA, kB, kC); cc.kunpckdq(kA, kB, kC); cc.kunpckwd(kA, kB, kC); cc.kxnorb(kA, kB, kC); cc.kxnord(kA, kB, kC); cc.kxnorq(kA, kB, kC); cc.kxnorw(kA, kB, kC); cc.kxorb(kA, kB, kC); cc.kxord(kA, kB, kC); cc.kxorq(kA, kB, kC); cc.kxorw(kA, kB, kC); cc.nop(); cc.evex().vaddpd(xmmA, xmmB, xmmC); cc.evex().vaddpd(ymmA, ymmB, ymmC); cc.evex().vaddpd(zmmA, zmmB, zmmC); cc.evex().vaddps(xmmA, xmmB, xmmC); cc.evex().vaddps(ymmA, ymmB, ymmC); cc.evex().vaddps(zmmA, zmmB, zmmC); cc.evex().vaddsd(xmmA, xmmB, xmmC); cc.evex().vaddss(xmmA, xmmB, xmmC); cc.evex().valignd(xmmA, xmmB, xmmC, 0); cc.evex().valignd(ymmA, ymmB, ymmC, 0); cc.evex().valignd(zmmA, zmmB, zmmC, 0); cc.evex().valignq(xmmA, xmmB, xmmC, 0); cc.evex().valignq(ymmA, ymmB, ymmC, 0); cc.evex().valignq(zmmA, zmmB, zmmC, 0); cc.evex().vandnpd(xmmA, xmmB, xmmC); cc.evex().vandnpd(ymmA, ymmB, ymmC); cc.evex().vandnpd(zmmA, zmmB, zmmC); cc.evex().vandnps(xmmA, xmmB, xmmC); cc.evex().vandnps(ymmA, ymmB, ymmC); cc.evex().vandnps(zmmA, zmmB, zmmC); cc.evex().vandpd(xmmA, xmmB, xmmC); cc.evex().vandpd(ymmA, ymmB, ymmC); cc.evex().vandpd(zmmA, zmmB, zmmC); cc.evex().vandps(xmmA, xmmB, xmmC); cc.evex().vandps(ymmA, ymmB, ymmC); cc.evex().vandps(zmmA, zmmB, zmmC); cc.evex().vblendmpd(xmmA, xmmB, xmmC); cc.evex().vblendmpd(ymmA, ymmB, ymmC); cc.evex().vblendmpd(zmmA, zmmB, zmmC); cc.evex().vblendmps(xmmA, xmmB, xmmC); cc.evex().vblendmps(ymmA, ymmB, ymmC); cc.evex().vblendmps(zmmA, zmmB, zmmC); cc.evex().vbroadcastf32x2(ymmA, xmmB); cc.evex().vbroadcastf32x2(zmmA, xmmB); cc.evex().vbroadcasti32x2(xmmA, xmmB); cc.evex().vbroadcasti32x2(ymmA, xmmB); cc.evex().vbroadcasti32x2(zmmA, xmmB); cc.evex().vbroadcastsd(ymmA, xmmB); cc.evex().vbroadcastsd(zmmA, xmmB); cc.evex().vbroadcastss(xmmA, xmmB); cc.evex().vbroadcastss(ymmA, xmmB); cc.evex().vbroadcastss(zmmA, xmmB); cc.evex().vcmppd(kA, xmmB, xmmC, 0); cc.evex().vcmppd(kA, ymmB, ymmC, 0); cc.evex().vcmppd(kA, zmmB, zmmC, 0); cc.evex().vcmpps(kA, xmmB, xmmC, 0); cc.evex().vcmpps(kA, ymmB, ymmC, 0); cc.evex().vcmpps(kA, zmmB, zmmC, 0); cc.evex().vcmpsd(kA, xmmB, xmmC, 0); cc.evex().vcmpss(kA, xmmB, xmmC, 0); cc.evex().vcomisd(xmmA, xmmB); cc.evex().vcomiss(xmmA, xmmB); cc.evex().vcompresspd(xmmA, xmmB); cc.evex().vcompresspd(ymmA, ymmB); cc.evex().vcompresspd(zmmA, zmmB); cc.evex().vcompressps(xmmA, xmmB); cc.evex().vcompressps(ymmA, ymmB); cc.evex().vcompressps(zmmA, zmmB); cc.evex().vcvtdq2pd(xmmA, xmmB); cc.evex().vcvtdq2pd(ymmA, xmmB); cc.evex().vcvtdq2pd(zmmA, ymmB); cc.evex().vcvtdq2ps(xmmA, xmmB); cc.evex().vcvtdq2ps(ymmA, ymmB); cc.evex().vcvtdq2ps(zmmA, zmmB); cc.evex().vcvtpd2dq(xmmA, xmmB); cc.evex().vcvtpd2dq(xmmA, ymmB); cc.evex().vcvtpd2dq(ymmA, zmmB); cc.evex().vcvtpd2qq(xmmA, xmmB); cc.evex().vcvtpd2qq(ymmA, ymmB); cc.evex().vcvtpd2qq(zmmA, zmmB); cc.evex().vcvtpd2udq(xmmA, xmmB); cc.evex().vcvtpd2udq(xmmA, ymmB); cc.evex().vcvtpd2udq(ymmA, zmmB); cc.evex().vcvtpd2uqq(xmmA, xmmB); cc.evex().vcvtpd2uqq(ymmA, ymmB); cc.evex().vcvtpd2uqq(zmmA, zmmB); cc.evex().vcvtph2ps(xmmA, xmmB); cc.evex().vcvtph2ps(ymmA, xmmB); cc.evex().vcvtph2ps(zmmA, ymmB); cc.evex().vcvtps2dq(xmmA, xmmB); cc.evex().vcvtps2dq(ymmA, ymmB); cc.evex().vcvtps2dq(zmmA, zmmB); cc.evex().vcvtps2pd(xmmA, xmmB); cc.evex().vcvtps2pd(ymmA, xmmB); cc.evex().vcvtps2pd(zmmA, ymmB); cc.evex().vcvtps2ph(xmmA, xmmB, 0); cc.evex().vcvtps2ph(xmmA, ymmB, 0); cc.evex().vcvtps2ph(ymmA, zmmB, 0); cc.evex().vcvtps2qq(xmmA, xmmB); cc.evex().vcvtps2qq(ymmA, xmmB); cc.evex().vcvtps2qq(zmmA, ymmB); cc.evex().vcvtps2udq(xmmA, xmmB); cc.evex().vcvtps2udq(ymmA, ymmB); cc.evex().vcvtps2udq(zmmA, zmmB); cc.evex().vcvtps2uqq(xmmA, xmmB); cc.evex().vcvtps2uqq(ymmA, xmmB); cc.evex().vcvtps2uqq(zmmA, ymmB); cc.evex().vcvtqq2pd(xmmA, xmmB); cc.evex().vcvtqq2pd(ymmA, ymmB); cc.evex().vcvtqq2pd(zmmA, zmmB); cc.evex().vcvtqq2ps(xmmA, xmmB); cc.evex().vcvtqq2ps(xmmA, ymmB); cc.evex().vcvtqq2ps(ymmA, zmmB); cc.evex().vcvtsd2si(gpd, xmmB); cc.evex().vcvtsd2si(gpz, xmmB); cc.evex().vcvtsd2ss(xmmA, xmmB, xmmC); cc.evex().vcvtsd2usi(gpd, xmmB); cc.evex().vcvtsd2usi(gpz, xmmB); cc.evex().vcvtsi2sd(xmmA, xmmB, gpd); cc.evex().vcvtsi2sd(xmmA, xmmB, gpz); cc.evex().vcvtsi2ss(xmmA, xmmB, gpd); cc.evex().vcvtsi2ss(xmmA, xmmB, gpz); cc.evex().vcvtss2sd(xmmA, xmmB, xmmC); cc.evex().vcvtss2si(gpd, xmmB); cc.evex().vcvtss2si(gpz, xmmB); cc.evex().vcvtss2usi(gpd, xmmB); cc.evex().vcvtss2usi(gpz, xmmB); cc.evex().vcvttpd2dq(xmmA, xmmB); cc.evex().vcvttpd2dq(xmmA, ymmB); cc.evex().vcvttpd2dq(ymmA, zmmB); cc.evex().vcvttpd2qq(xmmA, xmmB); cc.evex().vcvttpd2qq(ymmA, ymmB); cc.evex().vcvttpd2qq(zmmA, zmmB); cc.evex().vcvttpd2udq(xmmA, xmmB); cc.evex().vcvttpd2udq(xmmA, ymmB); cc.evex().vcvttpd2udq(ymmA, zmmB); cc.evex().vcvttpd2uqq(xmmA, xmmB); cc.evex().vcvttpd2uqq(ymmA, ymmB); cc.evex().vcvttpd2uqq(zmmA, zmmB); cc.evex().vcvttps2dq(xmmA, xmmB); cc.evex().vcvttps2dq(ymmA, ymmB); cc.evex().vcvttps2dq(zmmA, zmmB); cc.evex().vcvttps2qq(xmmA, xmmB); cc.evex().vcvttps2qq(ymmA, xmmB); cc.evex().vcvttps2qq(zmmA, ymmB); cc.evex().vcvttps2udq(xmmA, xmmB); cc.evex().vcvttps2udq(ymmA, ymmB); cc.evex().vcvttps2udq(zmmA, zmmB); cc.evex().vcvttps2uqq(xmmA, xmmB); cc.evex().vcvttps2uqq(ymmA, xmmB); cc.evex().vcvttps2uqq(zmmA, ymmB); cc.evex().vcvttsd2si(gpd, xmmB); cc.evex().vcvttsd2si(gpz, xmmB); cc.evex().vcvttsd2usi(gpd, xmmB); cc.evex().vcvttsd2usi(gpz, xmmB); cc.evex().vcvttss2si(gpd, xmmB); cc.evex().vcvttss2si(gpz, xmmB); cc.evex().vcvttss2usi(gpd, xmmB); cc.evex().vcvttss2usi(gpz, xmmB); cc.evex().vcvtudq2pd(xmmA, xmmB); cc.evex().vcvtudq2pd(ymmA, xmmB); cc.evex().vcvtudq2pd(zmmA, ymmB); cc.evex().vcvtudq2ps(xmmA, xmmB); cc.evex().vcvtudq2ps(ymmA, ymmB); cc.evex().vcvtudq2ps(zmmA, zmmB); cc.evex().vcvtuqq2pd(xmmA, xmmB); cc.evex().vcvtuqq2pd(ymmA, ymmB); cc.evex().vcvtuqq2pd(zmmA, zmmB); cc.evex().vcvtuqq2ps(xmmA, xmmB); cc.evex().vcvtuqq2ps(xmmA, ymmB); cc.evex().vcvtuqq2ps(ymmA, zmmB); cc.evex().vcvtusi2sd(xmmA, xmmB, gpd); cc.evex().vcvtusi2sd(xmmA, xmmB, gpz); cc.evex().vcvtusi2ss(xmmA, xmmB, gpd); cc.evex().vcvtusi2ss(xmmA, xmmB, gpz); cc.evex().vdbpsadbw(xmmA, xmmB, xmmC, 0); cc.evex().vdbpsadbw(ymmA, ymmB, ymmC, 0); cc.evex().vdbpsadbw(zmmA, zmmB, zmmC, 0); cc.evex().vdivpd(xmmA, xmmB, xmmC); cc.evex().vdivpd(ymmA, ymmB, ymmC); cc.evex().vdivpd(zmmA, zmmB, zmmC); cc.evex().vdivps(xmmA, xmmB, xmmC); cc.evex().vdivps(ymmA, ymmB, ymmC); cc.evex().vdivps(zmmA, zmmB, zmmC); cc.evex().vdivsd(xmmA, xmmB, xmmC); cc.evex().vdivss(xmmA, xmmB, xmmC); cc.evex().vexp2pd(zmmA, zmmB); cc.evex().vexp2ps(zmmA, zmmB); cc.evex().vexpandpd(xmmA, xmmB); cc.evex().vexpandpd(ymmA, ymmB); cc.evex().vexpandpd(zmmA, zmmB); cc.evex().vexpandps(xmmA, xmmB); cc.evex().vexpandps(ymmA, ymmB); cc.evex().vexpandps(zmmA, zmmB); cc.evex().vextractf32x4(xmmA, ymmB, 0); cc.evex().vextractf32x4(xmmA, zmmB, 0); cc.evex().vextractf32x8(ymmA, zmmB, 0); cc.evex().vextractf64x2(xmmA, ymmB, 0); cc.evex().vextractf64x2(xmmA, zmmB, 0); cc.evex().vextractf64x4(ymmA, zmmB, 0); cc.evex().vextracti32x4(xmmA, ymmB, 0); cc.evex().vextracti32x4(xmmA, zmmB, 0); cc.evex().vextracti32x8(ymmA, zmmB, 0); cc.evex().vextracti64x2(xmmA, ymmB, 0); cc.evex().vextracti64x2(xmmA, zmmB, 0); cc.evex().vextracti64x4(ymmA, zmmB, 0); cc.evex().vextractps(gpd, xmmB, 0); cc.evex().vfixupimmpd(xmmA, xmmB, xmmC, 0); cc.evex().vfixupimmpd(ymmA, ymmB, ymmC, 0); cc.evex().vfixupimmpd(zmmA, zmmB, zmmC, 0); cc.evex().vfixupimmps(xmmA, xmmB, xmmC, 0); cc.evex().vfixupimmps(ymmA, ymmB, ymmC, 0); cc.evex().vfixupimmps(zmmA, zmmB, zmmC, 0); cc.evex().vfixupimmsd(xmmA, xmmB, xmmC, 0); cc.evex().vfixupimmss(xmmA, xmmB, xmmC, 0); cc.evex().vfmadd132pd(xmmA, xmmB, xmmC); cc.evex().vfmadd132pd(ymmA, ymmB, ymmC); cc.evex().vfmadd132pd(zmmA, zmmB, zmmC); cc.evex().vfmadd132ps(xmmA, xmmB, xmmC); cc.evex().vfmadd132ps(ymmA, ymmB, ymmC); cc.evex().vfmadd132ps(zmmA, zmmB, zmmC); cc.evex().vfmadd132sd(xmmA, xmmB, xmmC); cc.evex().vfmadd132ss(xmmA, xmmB, xmmC); cc.evex().vfmadd213pd(xmmA, xmmB, xmmC); cc.evex().vfmadd213pd(ymmA, ymmB, ymmC); cc.evex().vfmadd213pd(zmmA, zmmB, zmmC); cc.evex().vfmadd213ps(xmmA, xmmB, xmmC); cc.evex().vfmadd213ps(ymmA, ymmB, ymmC); cc.evex().vfmadd213ps(zmmA, zmmB, zmmC); cc.evex().vfmadd213sd(xmmA, xmmB, xmmC); cc.evex().vfmadd213ss(xmmA, xmmB, xmmC); cc.evex().vfmadd231pd(xmmA, xmmB, xmmC); cc.evex().vfmadd231pd(ymmA, ymmB, ymmC); cc.evex().vfmadd231pd(zmmA, zmmB, zmmC); cc.evex().vfmadd231ps(xmmA, xmmB, xmmC); cc.evex().vfmadd231ps(ymmA, ymmB, ymmC); cc.evex().vfmadd231ps(zmmA, zmmB, zmmC); cc.evex().vfmadd231sd(xmmA, xmmB, xmmC); cc.evex().vfmadd231ss(xmmA, xmmB, xmmC); cc.evex().vfmaddsub132pd(xmmA, xmmB, xmmC); cc.evex().vfmaddsub132pd(ymmA, ymmB, ymmC); cc.evex().vfmaddsub132pd(zmmA, zmmB, zmmC); cc.evex().vfmaddsub132ps(xmmA, xmmB, xmmC); cc.evex().vfmaddsub132ps(ymmA, ymmB, ymmC); cc.evex().vfmaddsub132ps(zmmA, zmmB, zmmC); cc.evex().vfmaddsub213pd(xmmA, xmmB, xmmC); cc.evex().vfmaddsub213pd(ymmA, ymmB, ymmC); cc.evex().vfmaddsub213pd(zmmA, zmmB, zmmC); cc.evex().vfmaddsub213ps(xmmA, xmmB, xmmC); cc.evex().vfmaddsub213ps(ymmA, ymmB, ymmC); cc.evex().vfmaddsub213ps(zmmA, zmmB, zmmC); cc.evex().vfmaddsub231pd(xmmA, xmmB, xmmC); cc.evex().vfmaddsub231pd(ymmA, ymmB, ymmC); cc.evex().vfmaddsub231pd(zmmA, zmmB, zmmC); cc.evex().vfmaddsub231ps(xmmA, xmmB, xmmC); cc.evex().vfmaddsub231ps(ymmA, ymmB, ymmC); cc.evex().vfmaddsub231ps(zmmA, zmmB, zmmC); cc.evex().vfmsub132pd(xmmA, xmmB, xmmC); cc.evex().vfmsub132pd(ymmA, ymmB, ymmC); cc.evex().vfmsub132pd(zmmA, zmmB, zmmC); cc.evex().vfmsub132ps(xmmA, xmmB, xmmC); cc.evex().vfmsub132ps(ymmA, ymmB, ymmC); cc.evex().vfmsub132ps(zmmA, zmmB, zmmC); cc.evex().vfmsub132sd(xmmA, xmmB, xmmC); cc.evex().vfmsub132ss(xmmA, xmmB, xmmC); cc.evex().vfmsub213pd(xmmA, xmmB, xmmC); cc.evex().vfmsub213pd(ymmA, ymmB, ymmC); cc.evex().vfmsub213pd(zmmA, zmmB, zmmC); cc.evex().vfmsub213ps(xmmA, xmmB, xmmC); cc.evex().vfmsub213ps(ymmA, ymmB, ymmC); cc.evex().vfmsub213ps(zmmA, zmmB, zmmC); cc.evex().vfmsub213sd(xmmA, xmmB, xmmC); cc.evex().vfmsub213ss(xmmA, xmmB, xmmC); cc.evex().vfmsub231pd(xmmA, xmmB, xmmC); cc.evex().vfmsub231pd(ymmA, ymmB, ymmC); cc.evex().vfmsub231pd(zmmA, zmmB, zmmC); cc.evex().vfmsub231ps(xmmA, xmmB, xmmC); cc.evex().vfmsub231ps(ymmA, ymmB, ymmC); cc.evex().vfmsub231ps(zmmA, zmmB, zmmC); cc.evex().vfmsub231sd(xmmA, xmmB, xmmC); cc.evex().vfmsub231ss(xmmA, xmmB, xmmC); cc.evex().vfmsubadd132pd(xmmA, xmmB, xmmC); cc.evex().vfmsubadd132pd(ymmA, ymmB, ymmC); cc.evex().vfmsubadd132pd(zmmA, zmmB, zmmC); cc.evex().vfmsubadd132ps(xmmA, xmmB, xmmC); cc.evex().vfmsubadd132ps(ymmA, ymmB, ymmC); cc.evex().vfmsubadd132ps(zmmA, zmmB, zmmC); cc.evex().vfmsubadd213pd(xmmA, xmmB, xmmC); cc.evex().vfmsubadd213pd(ymmA, ymmB, ymmC); cc.evex().vfmsubadd213pd(zmmA, zmmB, zmmC); cc.evex().vfmsubadd213ps(xmmA, xmmB, xmmC); cc.evex().vfmsubadd213ps(ymmA, ymmB, ymmC); cc.evex().vfmsubadd213ps(zmmA, zmmB, zmmC); cc.evex().vfmsubadd231pd(xmmA, xmmB, xmmC); cc.evex().vfmsubadd231pd(ymmA, ymmB, ymmC); cc.evex().vfmsubadd231pd(zmmA, zmmB, zmmC); cc.evex().vfmsubadd231ps(xmmA, xmmB, xmmC); cc.evex().vfmsubadd231ps(ymmA, ymmB, ymmC); cc.evex().vfmsubadd231ps(zmmA, zmmB, zmmC); cc.evex().vfnmadd132pd(xmmA, xmmB, xmmC); cc.evex().vfnmadd132pd(ymmA, ymmB, ymmC); cc.evex().vfnmadd132pd(zmmA, zmmB, zmmC); cc.evex().vfnmadd132ps(xmmA, xmmB, xmmC); cc.evex().vfnmadd132ps(ymmA, ymmB, ymmC); cc.evex().vfnmadd132ps(zmmA, zmmB, zmmC); cc.evex().vfnmadd132sd(xmmA, xmmB, xmmC); cc.evex().vfnmadd132ss(xmmA, xmmB, xmmC); cc.evex().vfnmadd213pd(xmmA, xmmB, xmmC); cc.evex().vfnmadd213pd(ymmA, ymmB, ymmC); cc.evex().vfnmadd213pd(zmmA, zmmB, zmmC); cc.evex().vfnmadd213ps(xmmA, xmmB, xmmC); cc.evex().vfnmadd213ps(ymmA, ymmB, ymmC); cc.evex().vfnmadd213ps(zmmA, zmmB, zmmC); cc.evex().vfnmadd213sd(xmmA, xmmB, xmmC); cc.evex().vfnmadd213ss(xmmA, xmmB, xmmC); cc.evex().vfnmadd231pd(xmmA, xmmB, xmmC); cc.evex().vfnmadd231pd(ymmA, ymmB, ymmC); cc.evex().vfnmadd231pd(zmmA, zmmB, zmmC); cc.evex().vfnmadd231ps(xmmA, xmmB, xmmC); cc.evex().vfnmadd231ps(ymmA, ymmB, ymmC); cc.evex().vfnmadd231ps(zmmA, zmmB, zmmC); cc.evex().vfnmadd231sd(xmmA, xmmB, xmmC); cc.evex().vfnmadd231ss(xmmA, xmmB, xmmC); cc.evex().vfnmsub132pd(xmmA, xmmB, xmmC); cc.evex().vfnmsub132pd(ymmA, ymmB, ymmC); cc.evex().vfnmsub132pd(zmmA, zmmB, zmmC); cc.evex().vfnmsub132ps(xmmA, xmmB, xmmC); cc.evex().vfnmsub132ps(ymmA, ymmB, ymmC); cc.evex().vfnmsub132ps(zmmA, zmmB, zmmC); cc.evex().vfnmsub132sd(xmmA, xmmB, xmmC); cc.evex().vfnmsub132ss(xmmA, xmmB, xmmC); cc.evex().vfnmsub213pd(xmmA, xmmB, xmmC); cc.evex().vfnmsub213pd(ymmA, ymmB, ymmC); cc.evex().vfnmsub213pd(zmmA, zmmB, zmmC); cc.evex().vfnmsub213ps(xmmA, xmmB, xmmC); cc.evex().vfnmsub213ps(ymmA, ymmB, ymmC); cc.evex().vfnmsub213ps(zmmA, zmmB, zmmC); cc.evex().vfnmsub213sd(xmmA, xmmB, xmmC); cc.evex().vfnmsub213ss(xmmA, xmmB, xmmC); cc.evex().vfnmsub231pd(xmmA, xmmB, xmmC); cc.evex().vfnmsub231pd(ymmA, ymmB, ymmC); cc.evex().vfnmsub231pd(zmmA, zmmB, zmmC); cc.evex().vfnmsub231ps(xmmA, xmmB, xmmC); cc.evex().vfnmsub231ps(ymmA, ymmB, ymmC); cc.evex().vfnmsub231ps(zmmA, zmmB, zmmC); cc.evex().vfnmsub231sd(xmmA, xmmB, xmmC); cc.evex().vfnmsub231ss(xmmA, xmmB, xmmC); cc.evex().vfpclasspd(kA, xmmB, 0); cc.evex().vfpclasspd(kA, ymmB, 0); cc.evex().vfpclasspd(kA, zmmB, 0); cc.evex().vfpclassps(kA, xmmB, 0); cc.evex().vfpclassps(kA, ymmB, 0); cc.evex().vfpclassps(kA, zmmB, 0); cc.evex().vfpclasssd(kA, xmmB, 0); cc.evex().vfpclassss(kA, xmmB, 0); cc.evex().vgetexppd(xmmA, xmmB); cc.evex().vgetexppd(ymmA, ymmB); cc.evex().vgetexppd(zmmA, zmmB); cc.evex().vgetexpps(xmmA, xmmB); cc.evex().vgetexpps(ymmA, ymmB); cc.evex().vgetexpps(zmmA, zmmB); cc.evex().vgetexpsd(xmmA, xmmB, xmmC); cc.evex().vgetexpss(xmmA, xmmB, xmmC); cc.evex().vgetmantpd(xmmA, xmmB, 0); cc.evex().vgetmantpd(ymmA, ymmB, 0); cc.evex().vgetmantpd(zmmA, zmmB, 0); cc.evex().vgetmantps(xmmA, xmmB, 0); cc.evex().vgetmantps(ymmA, ymmB, 0); cc.evex().vgetmantps(zmmA, zmmB, 0); cc.evex().vgetmantsd(xmmA, xmmB, xmmC, 0); cc.evex().vgetmantss(xmmA, xmmB, xmmC, 0); cc.evex().vinsertf32x4(ymmA, ymmB, xmmC, 0); cc.evex().vinsertf32x4(zmmA, zmmB, xmmC, 0); cc.evex().vinsertf32x8(zmmA, zmmB, ymmC, 0); cc.evex().vinsertf64x2(ymmA, ymmB, xmmC, 0); cc.evex().vinsertf64x2(zmmA, zmmB, xmmC, 0); cc.evex().vinsertf64x4(zmmA, zmmB, ymmC, 0); cc.evex().vinserti32x4(ymmA, ymmB, xmmC, 0); cc.evex().vinserti32x4(zmmA, zmmB, xmmC, 0); cc.evex().vinserti32x8(zmmA, zmmB, ymmC, 0); cc.evex().vinserti64x2(ymmA, ymmB, xmmC, 0); cc.evex().vinserti64x2(zmmA, zmmB, xmmC, 0); cc.evex().vinserti64x4(zmmA, zmmB, ymmC, 0); cc.evex().vinsertps(xmmA, xmmB, xmmC, 0); cc.evex().vmaxpd(xmmA, xmmB, xmmC); cc.evex().vmaxpd(ymmA, ymmB, ymmC); cc.evex().vmaxpd(zmmA, zmmB, zmmC); cc.evex().vmaxps(xmmA, xmmB, xmmC); cc.evex().vmaxps(ymmA, ymmB, ymmC); cc.evex().vmaxps(zmmA, zmmB, zmmC); cc.evex().vmaxsd(xmmA, xmmB, xmmC); cc.evex().vmaxss(xmmA, xmmB, xmmC); cc.evex().vminpd(xmmA, xmmB, xmmC); cc.evex().vminpd(ymmA, ymmB, ymmC); cc.evex().vminpd(zmmA, zmmB, zmmC); cc.evex().vminps(xmmA, xmmB, xmmC); cc.evex().vminps(ymmA, ymmB, ymmC); cc.evex().vminps(zmmA, zmmB, zmmC); cc.evex().vminsd(xmmA, xmmB, xmmC); cc.evex().vminss(xmmA, xmmB, xmmC); cc.evex().vmovapd(xmmA, xmmB); cc.evex().vmovapd(xmmA, xmmB); cc.evex().vmovapd(ymmA, ymmB); cc.evex().vmovapd(ymmA, ymmB); cc.evex().vmovapd(zmmA, zmmB); cc.evex().vmovapd(zmmA, zmmB); cc.evex().vmovaps(xmmA, xmmB); cc.evex().vmovaps(xmmA, xmmB); cc.evex().vmovaps(ymmA, ymmB); cc.evex().vmovaps(ymmA, ymmB); cc.evex().vmovaps(zmmA, zmmB); cc.evex().vmovaps(zmmA, zmmB); cc.evex().vmovd(gpd, xmmB); cc.evex().vmovd(xmmA, gpd); cc.evex().vmovddup(xmmA, xmmB); cc.evex().vmovddup(ymmA, ymmB); cc.evex().vmovddup(zmmA, zmmB); cc.evex().vmovdqa32(xmmA, xmmB); cc.evex().vmovdqa32(xmmA, xmmB); cc.evex().vmovdqa32(ymmA, ymmB); cc.evex().vmovdqa32(ymmA, ymmB); cc.evex().vmovdqa32(zmmA, zmmB); cc.evex().vmovdqa32(zmmA, zmmB); cc.evex().vmovdqa64(xmmA, xmmB); cc.evex().vmovdqa64(xmmA, xmmB); cc.evex().vmovdqa64(ymmA, ymmB); cc.evex().vmovdqa64(ymmA, ymmB); cc.evex().vmovdqa64(zmmA, zmmB); cc.evex().vmovdqa64(zmmA, zmmB); cc.evex().vmovdqu16(xmmA, xmmB); cc.evex().vmovdqu16(xmmA, xmmB); cc.evex().vmovdqu16(ymmA, ymmB); cc.evex().vmovdqu16(ymmA, ymmB); cc.evex().vmovdqu16(zmmA, zmmB); cc.evex().vmovdqu16(zmmA, zmmB); cc.evex().vmovdqu32(xmmA, xmmB); cc.evex().vmovdqu32(xmmA, xmmB); cc.evex().vmovdqu32(ymmA, ymmB); cc.evex().vmovdqu32(ymmA, ymmB); cc.evex().vmovdqu32(zmmA, zmmB); cc.evex().vmovdqu32(zmmA, zmmB); cc.evex().vmovdqu64(xmmA, xmmB); cc.evex().vmovdqu64(xmmA, xmmB); cc.evex().vmovdqu64(ymmA, ymmB); cc.evex().vmovdqu64(ymmA, ymmB); cc.evex().vmovdqu64(zmmA, zmmB); cc.evex().vmovdqu64(zmmA, zmmB); cc.evex().vmovdqu8(xmmA, xmmB); cc.evex().vmovdqu8(xmmA, xmmB); cc.evex().vmovdqu8(ymmA, ymmB); cc.evex().vmovdqu8(ymmA, ymmB); cc.evex().vmovdqu8(zmmA, zmmB); cc.evex().vmovdqu8(zmmA, zmmB); cc.evex().vmovhlps(xmmA, xmmB, xmmC); if (cc.is64Bit()) cc.evex().vmovq(gpq, xmmB); if (cc.is64Bit()) cc.evex().vmovq(xmmA, gpq); cc.evex().vmovq(xmmA, xmmB); cc.evex().vmovsd(xmmA, xmmB, xmmC); cc.evex().vmovshdup(xmmA, xmmB); cc.evex().vmovshdup(ymmA, ymmB); cc.evex().vmovshdup(zmmA, zmmB); cc.evex().vmovsldup(xmmA, xmmB); cc.evex().vmovsldup(ymmA, ymmB); cc.evex().vmovsldup(zmmA, zmmB); cc.evex().vmovss(xmmA, xmmB, xmmC); cc.evex().vmovupd(xmmA, xmmB); cc.evex().vmovupd(xmmA, xmmB); cc.evex().vmovupd(ymmA, ymmB); cc.evex().vmovupd(ymmA, ymmB); cc.evex().vmovupd(zmmA, zmmB); cc.evex().vmovupd(zmmA, zmmB); cc.evex().vmovups(xmmA, xmmB); cc.evex().vmovups(xmmA, xmmB); cc.evex().vmovups(ymmA, ymmB); cc.evex().vmovups(ymmA, ymmB); cc.evex().vmovups(zmmA, zmmB); cc.evex().vmovups(zmmA, zmmB); cc.evex().vmulpd(xmmA, xmmB, xmmC); cc.evex().vmulpd(ymmA, ymmB, ymmC); cc.evex().vmulpd(zmmA, zmmB, zmmC); cc.evex().vmulps(xmmA, xmmB, xmmC); cc.evex().vmulps(ymmA, ymmB, ymmC); cc.evex().vmulps(zmmA, zmmB, zmmC); cc.evex().vmulsd(xmmA, xmmB, xmmC); cc.evex().vmulss(xmmA, xmmB, xmmC); cc.evex().vorpd(xmmA, xmmB, xmmC); cc.evex().vorpd(ymmA, ymmB, ymmC); cc.evex().vorpd(zmmA, zmmB, zmmC); cc.evex().vorps(xmmA, xmmB, xmmC); cc.evex().vorps(ymmA, ymmB, ymmC); cc.evex().vorps(zmmA, zmmB, zmmC); cc.evex().vpabsb(xmmA, xmmB); cc.evex().vpabsb(ymmA, ymmB); cc.evex().vpabsb(zmmA, zmmB); cc.evex().vpabsd(xmmA, xmmB); cc.evex().vpabsd(ymmA, ymmB); cc.evex().vpabsd(zmmA, zmmB); cc.evex().vpabsq(xmmA, xmmB); cc.evex().vpabsq(ymmA, ymmB); cc.evex().vpabsq(zmmA, zmmB); cc.evex().vpabsw(xmmA, xmmB); cc.evex().vpabsw(ymmA, ymmB); cc.evex().vpabsw(zmmA, zmmB); cc.evex().vpackssdw(xmmA, xmmB, xmmC); cc.evex().vpackssdw(ymmA, ymmB, ymmC); cc.evex().vpackssdw(zmmA, zmmB, zmmC); cc.evex().vpacksswb(xmmA, xmmB, xmmC); cc.evex().vpacksswb(ymmA, ymmB, ymmC); cc.evex().vpacksswb(zmmA, zmmB, zmmC); cc.evex().vpackusdw(xmmA, xmmB, xmmC); cc.evex().vpackusdw(ymmA, ymmB, ymmC); cc.evex().vpackusdw(zmmA, zmmB, zmmC); cc.evex().vpackuswb(xmmA, xmmB, xmmC); cc.evex().vpackuswb(ymmA, ymmB, ymmC); cc.evex().vpackuswb(zmmA, zmmB, zmmC); cc.evex().vpaddb(xmmA, xmmB, xmmC); cc.evex().vpaddb(ymmA, ymmB, ymmC); cc.evex().vpaddb(zmmA, zmmB, zmmC); cc.evex().vpaddd(xmmA, xmmB, xmmC); cc.evex().vpaddd(ymmA, ymmB, ymmC); cc.evex().vpaddd(zmmA, zmmB, zmmC); cc.evex().vpaddq(xmmA, xmmB, xmmC); cc.evex().vpaddq(ymmA, ymmB, ymmC); cc.evex().vpaddq(zmmA, zmmB, zmmC); cc.evex().vpaddsb(xmmA, xmmB, xmmC); cc.evex().vpaddsb(ymmA, ymmB, ymmC); cc.evex().vpaddsb(zmmA, zmmB, zmmC); cc.evex().vpaddsw(xmmA, xmmB, xmmC); cc.evex().vpaddsw(ymmA, ymmB, ymmC); cc.evex().vpaddsw(zmmA, zmmB, zmmC); cc.evex().vpaddusb(xmmA, xmmB, xmmC); cc.evex().vpaddusb(ymmA, ymmB, ymmC); cc.evex().vpaddusb(zmmA, zmmB, zmmC); cc.evex().vpaddusw(xmmA, xmmB, xmmC); cc.evex().vpaddusw(ymmA, ymmB, ymmC); cc.evex().vpaddusw(zmmA, zmmB, zmmC); cc.evex().vpaddw(xmmA, xmmB, xmmC); cc.evex().vpaddw(ymmA, ymmB, ymmC); cc.evex().vpaddw(zmmA, zmmB, zmmC); cc.evex().vpalignr(xmmA, xmmB, xmmC, 0); cc.evex().vpalignr(ymmA, ymmB, ymmC, 0); cc.evex().vpalignr(zmmA, zmmB, zmmC, 0); cc.evex().vpandd(xmmA, xmmB, xmmC); cc.evex().vpandd(ymmA, ymmB, ymmC); cc.evex().vpandd(zmmA, zmmB, zmmC); cc.evex().vpandnd(xmmA, xmmB, xmmC); cc.evex().vpandnd(ymmA, ymmB, ymmC); cc.evex().vpandnd(zmmA, zmmB, zmmC); cc.evex().vpandnq(xmmA, xmmB, xmmC); cc.evex().vpandnq(ymmA, ymmB, ymmC); cc.evex().vpandnq(zmmA, zmmB, zmmC); cc.evex().vpandq(xmmA, xmmB, xmmC); cc.evex().vpandq(ymmA, ymmB, ymmC); cc.evex().vpandq(zmmA, zmmB, zmmC); cc.evex().vpavgb(xmmA, xmmB, xmmC); cc.evex().vpavgb(ymmA, ymmB, ymmC); cc.evex().vpavgb(zmmA, zmmB, zmmC); cc.evex().vpavgw(xmmA, xmmB, xmmC); cc.evex().vpavgw(ymmA, ymmB, ymmC); cc.evex().vpavgw(zmmA, zmmB, zmmC); cc.evex().vpblendmb(xmmA, xmmB, xmmC); cc.evex().vpblendmb(ymmA, ymmB, ymmC); cc.evex().vpblendmb(zmmA, zmmB, zmmC); cc.evex().vpblendmd(xmmA, xmmB, xmmC); cc.evex().vpblendmd(ymmA, ymmB, ymmC); cc.evex().vpblendmd(zmmA, zmmB, zmmC); cc.evex().vpblendmq(xmmA, xmmB, xmmC); cc.evex().vpblendmq(ymmA, ymmB, ymmC); cc.evex().vpblendmq(zmmA, zmmB, zmmC); cc.evex().vpblendmw(xmmA, xmmB, xmmC); cc.evex().vpblendmw(ymmA, ymmB, ymmC); cc.evex().vpblendmw(zmmA, zmmB, zmmC); cc.evex().vpbroadcastb(xmmA, gpd); cc.evex().vpbroadcastb(xmmA, xmmB); cc.evex().vpbroadcastb(ymmA, gpd); cc.evex().vpbroadcastb(ymmA, xmmB); cc.evex().vpbroadcastb(zmmA, gpd); cc.evex().vpbroadcastb(zmmA, xmmB); cc.evex().vpbroadcastd(xmmA, gpd); cc.evex().vpbroadcastd(xmmA, xmmB); cc.evex().vpbroadcastd(ymmA, gpd); cc.evex().vpbroadcastd(ymmA, xmmB); cc.evex().vpbroadcastd(zmmA, gpd); cc.evex().vpbroadcastd(zmmA, xmmB); cc.evex().vpbroadcastmb2q(xmmA, kB); cc.evex().vpbroadcastmb2q(ymmA, kB); cc.evex().vpbroadcastmb2q(zmmA, kB); cc.evex().vpbroadcastmw2d(xmmA, kB); cc.evex().vpbroadcastmw2d(ymmA, kB); cc.evex().vpbroadcastmw2d(zmmA, kB); if (cc.is64Bit()) cc.evex().vpbroadcastq(xmmA, gpq); cc.evex().vpbroadcastq(xmmA, xmmB); if (cc.is64Bit()) cc.evex().vpbroadcastq(ymmA, gpq); cc.evex().vpbroadcastq(ymmA, xmmB); if (cc.is64Bit()) cc.evex().vpbroadcastq(zmmA, gpq); cc.evex().vpbroadcastq(zmmA, xmmB); cc.evex().vpbroadcastw(xmmA, gpd); cc.evex().vpbroadcastw(xmmA, xmmB); cc.evex().vpbroadcastw(ymmA, gpd); cc.evex().vpbroadcastw(ymmA, xmmB); cc.evex().vpbroadcastw(zmmA, gpd); cc.evex().vpbroadcastw(zmmA, xmmB); cc.evex().vpcmpb(kA, xmmB, xmmC, 0); cc.evex().vpcmpb(kA, ymmB, ymmC, 0); cc.evex().vpcmpb(kA, zmmB, zmmC, 0); cc.evex().vpcmpd(kA, xmmB, xmmC, 0); cc.evex().vpcmpd(kA, ymmB, ymmC, 0); cc.evex().vpcmpd(kA, zmmB, zmmC, 0); cc.evex().vpcmpeqb(kA, xmmB, xmmC); cc.evex().vpcmpeqb(kA, ymmB, ymmC); cc.evex().vpcmpeqb(kA, zmmB, zmmC); cc.evex().vpcmpeqd(kA, xmmB, xmmC); cc.evex().vpcmpeqd(kA, ymmB, ymmC); cc.evex().vpcmpeqd(kA, zmmB, zmmC); cc.evex().vpcmpeqq(kA, xmmB, xmmC); cc.evex().vpcmpeqq(kA, ymmB, ymmC); cc.evex().vpcmpeqq(kA, zmmB, zmmC); cc.evex().vpcmpeqw(kA, xmmB, xmmC); cc.evex().vpcmpeqw(kA, ymmB, ymmC); cc.evex().vpcmpeqw(kA, zmmB, zmmC); cc.evex().vpcmpgtb(kA, xmmB, xmmC); cc.evex().vpcmpgtb(kA, ymmB, ymmC); cc.evex().vpcmpgtb(kA, zmmB, zmmC); cc.evex().vpcmpgtd(kA, xmmB, xmmC); cc.evex().vpcmpgtd(kA, ymmB, ymmC); cc.evex().vpcmpgtd(kA, zmmB, zmmC); cc.evex().vpcmpgtq(kA, xmmB, xmmC); cc.evex().vpcmpgtq(kA, ymmB, ymmC); cc.evex().vpcmpgtq(kA, zmmB, zmmC); cc.evex().vpcmpgtw(kA, xmmB, xmmC); cc.evex().vpcmpgtw(kA, ymmB, ymmC); cc.evex().vpcmpgtw(kA, zmmB, zmmC); cc.evex().vpcmpq(kA, xmmB, xmmC, 0); cc.evex().vpcmpq(kA, ymmB, ymmC, 0); cc.evex().vpcmpq(kA, zmmB, zmmC, 0); cc.evex().vpcmpub(kA, xmmB, xmmC, 0); cc.evex().vpcmpub(kA, ymmB, ymmC, 0); cc.evex().vpcmpub(kA, zmmB, zmmC, 0); cc.evex().vpcmpud(kA, xmmB, xmmC, 0); cc.evex().vpcmpud(kA, ymmB, ymmC, 0); cc.evex().vpcmpud(kA, zmmB, zmmC, 0); cc.evex().vpcmpuq(kA, xmmB, xmmC, 0); cc.evex().vpcmpuq(kA, ymmB, ymmC, 0); cc.evex().vpcmpuq(kA, zmmB, zmmC, 0); cc.evex().vpcmpuw(kA, xmmB, xmmC, 0); cc.evex().vpcmpuw(kA, ymmB, ymmC, 0); cc.evex().vpcmpuw(kA, zmmB, zmmC, 0); cc.evex().vpcmpw(kA, xmmB, xmmC, 0); cc.evex().vpcmpw(kA, ymmB, ymmC, 0); cc.evex().vpcmpw(kA, zmmB, zmmC, 0); cc.evex().vpcompressd(xmmA, xmmB); cc.evex().vpcompressd(ymmA, ymmB); cc.evex().vpcompressd(zmmA, zmmB); cc.evex().vpcompressq(xmmA, xmmB); cc.evex().vpcompressq(ymmA, ymmB); cc.evex().vpcompressq(zmmA, zmmB); cc.evex().vpconflictd(xmmA, xmmB); cc.evex().vpconflictd(ymmA, ymmB); cc.evex().vpconflictd(zmmA, zmmB); cc.evex().vpconflictq(xmmA, xmmB); cc.evex().vpconflictq(ymmA, ymmB); cc.evex().vpconflictq(zmmA, zmmB); cc.evex().vpermb(xmmA, xmmB, xmmC); cc.evex().vpermb(ymmA, ymmB, ymmC); cc.evex().vpermb(zmmA, zmmB, zmmC); cc.evex().vpermd(ymmA, ymmB, ymmC); cc.evex().vpermd(zmmA, zmmB, zmmC); cc.evex().vpermi2b(xmmA, xmmB, xmmC); cc.evex().vpermi2b(ymmA, ymmB, ymmC); cc.evex().vpermi2b(zmmA, zmmB, zmmC); cc.evex().vpermi2d(xmmA, xmmB, xmmC); cc.evex().vpermi2d(ymmA, ymmB, ymmC); cc.evex().vpermi2d(zmmA, zmmB, zmmC); cc.evex().vpermi2pd(xmmA, xmmB, xmmC); cc.evex().vpermi2pd(ymmA, ymmB, ymmC); cc.evex().vpermi2pd(zmmA, zmmB, zmmC); cc.evex().vpermi2ps(xmmA, xmmB, xmmC); cc.evex().vpermi2ps(ymmA, ymmB, ymmC); cc.evex().vpermi2ps(zmmA, zmmB, zmmC); cc.evex().vpermi2q(xmmA, xmmB, xmmC); cc.evex().vpermi2q(ymmA, ymmB, ymmC); cc.evex().vpermi2q(zmmA, zmmB, zmmC); cc.evex().vpermi2w(xmmA, xmmB, xmmC); cc.evex().vpermi2w(ymmA, ymmB, ymmC); cc.evex().vpermi2w(zmmA, zmmB, zmmC); cc.evex().vpermilpd(xmmA, xmmB, xmmC); cc.evex().vpermilpd(ymmA, ymmB, ymmC); cc.evex().vpermilpd(zmmA, zmmB, zmmC); cc.evex().vpermilpd(xmmA, xmmB, 0); cc.evex().vpermilpd(ymmA, ymmB, 0); cc.evex().vpermilpd(zmmA, zmmB, 0); cc.evex().vpermilps(xmmA, xmmB, xmmC); cc.evex().vpermilps(ymmA, ymmB, ymmC); cc.evex().vpermilps(zmmA, zmmB, zmmC); cc.evex().vpermilps(xmmA, xmmB, 0); cc.evex().vpermilps(ymmA, ymmB, 0); cc.evex().vpermilps(zmmA, zmmB, 0); cc.evex().vpermq(ymmA, ymmB, ymmC); cc.evex().vpermq(zmmA, zmmB, zmmC); cc.evex().vpermq(ymmA, ymmB, 0); cc.evex().vpermq(zmmA, zmmB, 0); cc.evex().vpermt2b(xmmA, xmmB, xmmC); cc.evex().vpermt2b(ymmA, ymmB, ymmC); cc.evex().vpermt2b(zmmA, zmmB, zmmC); cc.evex().vpermt2d(xmmA, xmmB, xmmC); cc.evex().vpermt2d(ymmA, ymmB, ymmC); cc.evex().vpermt2d(zmmA, zmmB, zmmC); cc.evex().vpermt2pd(xmmA, xmmB, xmmC); cc.evex().vpermt2pd(ymmA, ymmB, ymmC); cc.evex().vpermt2pd(zmmA, zmmB, zmmC); cc.evex().vpermt2ps(xmmA, xmmB, xmmC); cc.evex().vpermt2ps(ymmA, ymmB, ymmC); cc.evex().vpermt2ps(zmmA, zmmB, zmmC); cc.evex().vpermt2q(xmmA, xmmB, xmmC); cc.evex().vpermt2q(ymmA, ymmB, ymmC); cc.evex().vpermt2q(zmmA, zmmB, zmmC); cc.evex().vpermt2w(xmmA, xmmB, xmmC); cc.evex().vpermt2w(ymmA, ymmB, ymmC); cc.evex().vpermt2w(zmmA, zmmB, zmmC); cc.evex().vpermw(xmmA, xmmB, xmmC); cc.evex().vpermw(ymmA, ymmB, ymmC); cc.evex().vpermw(zmmA, zmmB, zmmC); cc.evex().vpexpandd(xmmA, xmmB); cc.evex().vpexpandd(ymmA, ymmB); cc.evex().vpexpandd(zmmA, zmmB); cc.evex().vpexpandq(xmmA, xmmB); cc.evex().vpexpandq(ymmA, ymmB); cc.evex().vpexpandq(zmmA, zmmB); cc.evex().vpextrb(gpd, xmmB, 0); cc.evex().vpextrd(gpd, xmmB, 0); if (cc.is64Bit()) cc.evex().vpextrq(gpq, xmmB, 0); cc.evex().vpextrw(gpd, xmmB, 0); cc.evex().vpinsrb(xmmA, xmmB, gpd, 0); cc.evex().vpinsrd(xmmA, xmmB, gpd, 0); if (cc.is64Bit()) cc.evex().vpinsrq(xmmA, xmmB, gpq, 0); cc.evex().vpinsrw(xmmA, xmmB, gpd, 0); cc.evex().vplzcntd(xmmA, xmmB); cc.evex().vplzcntd(ymmA, ymmB); cc.evex().vplzcntd(zmmA, zmmB); cc.evex().vplzcntq(xmmA, xmmB); cc.evex().vplzcntq(ymmA, ymmB); cc.evex().vplzcntq(zmmA, zmmB); cc.evex().vpmadd52huq(xmmA, xmmB, xmmC); cc.evex().vpmadd52huq(ymmA, ymmB, ymmC); cc.evex().vpmadd52huq(zmmA, zmmB, zmmC); cc.evex().vpmadd52luq(xmmA, xmmB, xmmC); cc.evex().vpmadd52luq(ymmA, ymmB, ymmC); cc.evex().vpmadd52luq(zmmA, zmmB, zmmC); cc.evex().vpmaddubsw(xmmA, xmmB, xmmC); cc.evex().vpmaddubsw(ymmA, ymmB, ymmC); cc.evex().vpmaddubsw(zmmA, zmmB, zmmC); cc.evex().vpmaddwd(xmmA, xmmB, xmmC); cc.evex().vpmaddwd(ymmA, ymmB, ymmC); cc.evex().vpmaddwd(zmmA, zmmB, zmmC); cc.evex().vpmaxsb(xmmA, xmmB, xmmC); cc.evex().vpmaxsb(ymmA, ymmB, ymmC); cc.evex().vpmaxsb(zmmA, zmmB, zmmC); cc.evex().vpmaxsd(xmmA, xmmB, xmmC); cc.evex().vpmaxsd(ymmA, ymmB, ymmC); cc.evex().vpmaxsd(zmmA, zmmB, zmmC); cc.evex().vpmaxsq(xmmA, xmmB, xmmC); cc.evex().vpmaxsq(ymmA, ymmB, ymmC); cc.evex().vpmaxsq(zmmA, zmmB, zmmC); cc.evex().vpmaxsw(xmmA, xmmB, xmmC); cc.evex().vpmaxsw(ymmA, ymmB, ymmC); cc.evex().vpmaxsw(zmmA, zmmB, zmmC); cc.evex().vpmaxub(xmmA, xmmB, xmmC); cc.evex().vpmaxub(ymmA, ymmB, ymmC); cc.evex().vpmaxub(zmmA, zmmB, zmmC); cc.evex().vpmaxud(xmmA, xmmB, xmmC); cc.evex().vpmaxud(ymmA, ymmB, ymmC); cc.evex().vpmaxud(zmmA, zmmB, zmmC); cc.evex().vpmaxuq(xmmA, xmmB, xmmC); cc.evex().vpmaxuq(ymmA, ymmB, ymmC); cc.evex().vpmaxuq(zmmA, zmmB, zmmC); cc.evex().vpmaxuw(xmmA, xmmB, xmmC); cc.evex().vpmaxuw(ymmA, ymmB, ymmC); cc.evex().vpmaxuw(zmmA, zmmB, zmmC); cc.evex().vpminsb(xmmA, xmmB, xmmC); cc.evex().vpminsb(ymmA, ymmB, ymmC); cc.evex().vpminsb(zmmA, zmmB, zmmC); cc.evex().vpminsd(xmmA, xmmB, xmmC); cc.evex().vpminsd(ymmA, ymmB, ymmC); cc.evex().vpminsd(zmmA, zmmB, zmmC); cc.evex().vpminsq(xmmA, xmmB, xmmC); cc.evex().vpminsq(ymmA, ymmB, ymmC); cc.evex().vpminsq(zmmA, zmmB, zmmC); cc.evex().vpminsw(xmmA, xmmB, xmmC); cc.evex().vpminsw(ymmA, ymmB, ymmC); cc.evex().vpminsw(zmmA, zmmB, zmmC); cc.evex().vpminub(xmmA, xmmB, xmmC); cc.evex().vpminub(ymmA, ymmB, ymmC); cc.evex().vpminub(zmmA, zmmB, zmmC); cc.evex().vpminud(xmmA, xmmB, xmmC); cc.evex().vpminud(ymmA, ymmB, ymmC); cc.evex().vpminud(zmmA, zmmB, zmmC); cc.evex().vpminuq(xmmA, xmmB, xmmC); cc.evex().vpminuq(ymmA, ymmB, ymmC); cc.evex().vpminuq(zmmA, zmmB, zmmC); cc.evex().vpminuw(xmmA, xmmB, xmmC); cc.evex().vpminuw(ymmA, ymmB, ymmC); cc.evex().vpminuw(zmmA, zmmB, zmmC); cc.evex().vpmovb2m(kA, xmmB); cc.evex().vpmovb2m(kA, ymmB); cc.evex().vpmovb2m(kA, zmmB); cc.evex().vpmovd2m(kA, xmmB); cc.evex().vpmovd2m(kA, ymmB); cc.evex().vpmovd2m(kA, zmmB); cc.evex().vpmovdb(xmmA, xmmB); cc.evex().vpmovdb(xmmA, ymmB); cc.evex().vpmovdb(xmmA, zmmB); cc.evex().vpmovdw(xmmA, xmmB); cc.evex().vpmovdw(xmmA, ymmB); cc.evex().vpmovdw(ymmA, zmmB); cc.evex().vpmovm2b(xmmA, kB); cc.evex().vpmovm2b(ymmA, kB); cc.evex().vpmovm2b(zmmA, kB); cc.evex().vpmovm2d(xmmA, kB); cc.evex().vpmovm2d(ymmA, kB); cc.evex().vpmovm2d(zmmA, kB); cc.evex().vpmovm2q(xmmA, kB); cc.evex().vpmovm2q(ymmA, kB); cc.evex().vpmovm2q(zmmA, kB); cc.evex().vpmovm2w(xmmA, kB); cc.evex().vpmovm2w(ymmA, kB); cc.evex().vpmovm2w(zmmA, kB); cc.evex().vpmovq2m(kA, xmmB); cc.evex().vpmovq2m(kA, ymmB); cc.evex().vpmovq2m(kA, zmmB); cc.evex().vpmovqb(xmmA, xmmB); cc.evex().vpmovqb(xmmA, ymmB); cc.evex().vpmovqb(xmmA, zmmB); cc.evex().vpmovqd(xmmA, xmmB); cc.evex().vpmovqd(xmmA, ymmB); cc.evex().vpmovqd(ymmA, zmmB); cc.evex().vpmovqw(xmmA, xmmB); cc.evex().vpmovqw(xmmA, ymmB); cc.evex().vpmovqw(xmmA, zmmB); cc.evex().vpmovsdb(xmmA, xmmB); cc.evex().vpmovsdb(xmmA, ymmB); cc.evex().vpmovsdb(xmmA, zmmB); cc.evex().vpmovsdw(xmmA, xmmB); cc.evex().vpmovsdw(xmmA, ymmB); cc.evex().vpmovsdw(ymmA, zmmB); cc.evex().vpmovsqb(xmmA, xmmB); cc.evex().vpmovsqb(xmmA, ymmB); cc.evex().vpmovsqb(xmmA, zmmB); cc.evex().vpmovsqd(xmmA, xmmB); cc.evex().vpmovsqd(xmmA, ymmB); cc.evex().vpmovsqd(ymmA, zmmB); cc.evex().vpmovsqw(xmmA, xmmB); cc.evex().vpmovsqw(xmmA, ymmB); cc.evex().vpmovsqw(xmmA, zmmB); cc.evex().vpmovswb(xmmA, xmmB); cc.evex().vpmovswb(xmmA, ymmB); cc.evex().vpmovswb(ymmA, zmmB); cc.evex().vpmovsxbd(xmmA, xmmB); cc.evex().vpmovsxbd(ymmA, xmmB); cc.evex().vpmovsxbd(zmmA, xmmB); cc.evex().vpmovsxbq(xmmA, xmmB); cc.evex().vpmovsxbq(ymmA, xmmB); cc.evex().vpmovsxbq(zmmA, xmmB); cc.evex().vpmovsxbw(xmmA, xmmB); cc.evex().vpmovsxbw(ymmA, xmmB); cc.evex().vpmovsxbw(zmmA, ymmB); cc.evex().vpmovsxdq(xmmA, xmmB); cc.evex().vpmovsxdq(ymmA, xmmB); cc.evex().vpmovsxdq(zmmA, ymmB); cc.evex().vpmovsxwd(xmmA, xmmB); cc.evex().vpmovsxwd(ymmA, xmmB); cc.evex().vpmovsxwd(zmmA, ymmB); cc.evex().vpmovsxwq(xmmA, xmmB); cc.evex().vpmovsxwq(ymmA, xmmB); cc.evex().vpmovsxwq(zmmA, xmmB); cc.evex().vpmovusdb(xmmA, xmmB); cc.evex().vpmovusdb(xmmA, ymmB); cc.evex().vpmovusdb(xmmA, zmmB); cc.evex().vpmovusdw(xmmA, xmmB); cc.evex().vpmovusdw(xmmA, ymmB); cc.evex().vpmovusdw(ymmA, zmmB); cc.evex().vpmovusqb(xmmA, xmmB); cc.evex().vpmovusqb(xmmA, ymmB); cc.evex().vpmovusqb(xmmA, zmmB); cc.evex().vpmovusqd(xmmA, xmmB); cc.evex().vpmovusqd(xmmA, ymmB); cc.evex().vpmovusqd(ymmA, zmmB); cc.evex().vpmovusqw(xmmA, xmmB); cc.evex().vpmovusqw(xmmA, ymmB); cc.evex().vpmovusqw(xmmA, zmmB); cc.evex().vpmovuswb(xmmA, xmmB); cc.evex().vpmovuswb(xmmA, ymmB); cc.evex().vpmovuswb(ymmA, zmmB); cc.evex().vpmovw2m(kA, xmmB); cc.evex().vpmovw2m(kA, ymmB); cc.evex().vpmovw2m(kA, zmmB); cc.evex().vpmovwb(xmmA, xmmB); cc.evex().vpmovwb(xmmA, ymmB); cc.evex().vpmovwb(ymmA, zmmB); cc.evex().vpmovzxbd(xmmA, xmmB); cc.evex().vpmovzxbd(ymmA, xmmB); cc.evex().vpmovzxbd(zmmA, xmmB); cc.evex().vpmovzxbq(xmmA, xmmB); cc.evex().vpmovzxbq(ymmA, xmmB); cc.evex().vpmovzxbq(zmmA, xmmB); cc.evex().vpmovzxbw(xmmA, xmmB); cc.evex().vpmovzxbw(ymmA, xmmB); cc.evex().vpmovzxbw(zmmA, ymmB); cc.evex().vpmovzxdq(xmmA, xmmB); cc.evex().vpmovzxdq(ymmA, xmmB); cc.evex().vpmovzxdq(zmmA, ymmB); cc.evex().vpmovzxwd(xmmA, xmmB); cc.evex().vpmovzxwd(ymmA, xmmB); cc.evex().vpmovzxwd(zmmA, ymmB); cc.evex().vpmovzxwq(xmmA, xmmB); cc.evex().vpmovzxwq(ymmA, xmmB); cc.evex().vpmovzxwq(zmmA, xmmB); cc.evex().vpmuldq(xmmA, xmmB, xmmC); cc.evex().vpmuldq(ymmA, ymmB, ymmC); cc.evex().vpmuldq(zmmA, zmmB, zmmC); cc.evex().vpmulhrsw(xmmA, xmmB, xmmC); cc.evex().vpmulhrsw(ymmA, ymmB, ymmC); cc.evex().vpmulhrsw(zmmA, zmmB, zmmC); cc.evex().vpmulhuw(xmmA, xmmB, xmmC); cc.evex().vpmulhuw(ymmA, ymmB, ymmC); cc.evex().vpmulhuw(zmmA, zmmB, zmmC); cc.evex().vpmulhw(xmmA, xmmB, xmmC); cc.evex().vpmulhw(ymmA, ymmB, ymmC); cc.evex().vpmulhw(zmmA, zmmB, zmmC); cc.evex().vpmulld(xmmA, xmmB, xmmC); cc.evex().vpmulld(ymmA, ymmB, ymmC); cc.evex().vpmulld(zmmA, zmmB, zmmC); cc.evex().vpmullq(xmmA, xmmB, xmmC); cc.evex().vpmullq(ymmA, ymmB, ymmC); cc.evex().vpmullq(zmmA, zmmB, zmmC); cc.evex().vpmullw(xmmA, xmmB, xmmC); cc.evex().vpmullw(ymmA, ymmB, ymmC); cc.evex().vpmullw(zmmA, zmmB, zmmC); cc.evex().vpmultishiftqb(xmmA, xmmB, xmmC); cc.evex().vpmultishiftqb(ymmA, ymmB, ymmC); cc.evex().vpmultishiftqb(zmmA, zmmB, zmmC); cc.evex().vpmuludq(xmmA, xmmB, xmmC); cc.evex().vpmuludq(ymmA, ymmB, ymmC); cc.evex().vpmuludq(zmmA, zmmB, zmmC); cc.evex().vpopcntd(zmmA, zmmB); cc.evex().vpopcntq(zmmA, zmmB); cc.evex().vpord(xmmA, xmmB, xmmC); cc.evex().vpord(ymmA, ymmB, ymmC); cc.evex().vpord(zmmA, zmmB, zmmC); cc.evex().vporq(xmmA, xmmB, xmmC); cc.evex().vporq(ymmA, ymmB, ymmC); cc.evex().vporq(zmmA, zmmB, zmmC); cc.evex().vprold(xmmA, xmmB, 0); cc.evex().vprold(ymmA, ymmB, 0); cc.evex().vprold(zmmA, zmmB, 0); cc.evex().vprolq(xmmA, xmmB, 0); cc.evex().vprolq(ymmA, ymmB, 0); cc.evex().vprolq(zmmA, zmmB, 0); cc.evex().vprolvd(xmmA, xmmB, xmmC); cc.evex().vprolvd(ymmA, ymmB, ymmC); cc.evex().vprolvd(zmmA, zmmB, zmmC); cc.evex().vprolvq(xmmA, xmmB, xmmC); cc.evex().vprolvq(ymmA, ymmB, ymmC); cc.evex().vprolvq(zmmA, zmmB, zmmC); cc.evex().vprord(xmmA, xmmB, 0); cc.evex().vprord(ymmA, ymmB, 0); cc.evex().vprord(zmmA, zmmB, 0); cc.evex().vprorq(xmmA, xmmB, 0); cc.evex().vprorq(ymmA, ymmB, 0); cc.evex().vprorq(zmmA, zmmB, 0); cc.evex().vprorvd(xmmA, xmmB, xmmC); cc.evex().vprorvd(ymmA, ymmB, ymmC); cc.evex().vprorvd(zmmA, zmmB, zmmC); cc.evex().vprorvq(xmmA, xmmB, xmmC); cc.evex().vprorvq(ymmA, ymmB, ymmC); cc.evex().vprorvq(zmmA, zmmB, zmmC); cc.evex().vpsadbw(xmmA, xmmB, xmmC); cc.evex().vpsadbw(ymmA, ymmB, ymmC); cc.evex().vpsadbw(zmmA, zmmB, zmmC); cc.evex().vpshufb(xmmA, xmmB, xmmC); cc.evex().vpshufb(ymmA, ymmB, ymmC); cc.evex().vpshufb(zmmA, zmmB, zmmC); cc.evex().vpshufd(xmmA, xmmB, 0); cc.evex().vpshufd(ymmA, ymmB, 0); cc.evex().vpshufd(zmmA, zmmB, 0); cc.evex().vpshufhw(xmmA, xmmB, 0); cc.evex().vpshufhw(ymmA, ymmB, 0); cc.evex().vpshufhw(zmmA, zmmB, 0); cc.evex().vpshuflw(xmmA, xmmB, 0); cc.evex().vpshuflw(ymmA, ymmB, 0); cc.evex().vpshuflw(zmmA, zmmB, 0); cc.evex().vpslld(xmmA, xmmB, xmmC); cc.evex().vpslld(xmmA, xmmB, 0); cc.evex().vpslld(ymmA, ymmB, xmmC); cc.evex().vpslld(ymmA, ymmB, 0); cc.evex().vpslld(zmmA, zmmB, xmmC); cc.evex().vpslld(zmmA, zmmB, 0); cc.evex().vpslldq(xmmA, xmmB, 0); cc.evex().vpslldq(ymmA, ymmB, 0); cc.evex().vpslldq(zmmA, zmmB, 0); cc.evex().vpsllq(xmmA, xmmB, xmmC); cc.evex().vpsllq(xmmA, xmmB, 0); cc.evex().vpsllq(ymmA, ymmB, xmmC); cc.evex().vpsllq(ymmA, ymmB, 0); cc.evex().vpsllq(zmmA, zmmB, xmmC); cc.evex().vpsllq(zmmA, zmmB, 0); cc.evex().vpsllvd(xmmA, xmmB, xmmC); cc.evex().vpsllvd(ymmA, ymmB, ymmC); cc.evex().vpsllvd(zmmA, zmmB, zmmC); cc.evex().vpsllvq(xmmA, xmmB, xmmC); cc.evex().vpsllvq(ymmA, ymmB, ymmC); cc.evex().vpsllvq(zmmA, zmmB, zmmC); cc.evex().vpsllvw(xmmA, xmmB, xmmC); cc.evex().vpsllvw(ymmA, ymmB, ymmC); cc.evex().vpsllvw(zmmA, zmmB, zmmC); cc.evex().vpsllw(xmmA, xmmB, xmmC); cc.evex().vpsllw(xmmA, xmmB, 0); cc.evex().vpsllw(ymmA, ymmB, xmmC); cc.evex().vpsllw(ymmA, ymmB, 0); cc.evex().vpsllw(zmmA, zmmB, xmmC); cc.evex().vpsllw(zmmA, zmmB, 0); cc.evex().vpsrad(xmmA, xmmB, xmmC); cc.evex().vpsrad(xmmA, xmmB, 0); cc.evex().vpsrad(ymmA, ymmB, xmmC); cc.evex().vpsrad(ymmA, ymmB, 0); cc.evex().vpsrad(zmmA, zmmB, xmmC); cc.evex().vpsrad(zmmA, zmmB, 0); cc.evex().vpsraq(xmmA, xmmB, xmmC); cc.evex().vpsraq(xmmA, xmmB, 0); cc.evex().vpsraq(ymmA, ymmB, xmmC); cc.evex().vpsraq(ymmA, ymmB, 0); cc.evex().vpsraq(zmmA, zmmB, xmmC); cc.evex().vpsraq(zmmA, zmmB, 0); cc.evex().vpsravd(xmmA, xmmB, xmmC); cc.evex().vpsravd(ymmA, ymmB, ymmC); cc.evex().vpsravd(zmmA, zmmB, zmmC); cc.evex().vpsravq(xmmA, xmmB, xmmC); cc.evex().vpsravq(ymmA, ymmB, ymmC); cc.evex().vpsravq(zmmA, zmmB, zmmC); cc.evex().vpsravw(xmmA, xmmB, xmmC); cc.evex().vpsravw(ymmA, ymmB, ymmC); cc.evex().vpsravw(zmmA, zmmB, zmmC); cc.evex().vpsraw(xmmA, xmmB, xmmC); cc.evex().vpsraw(xmmA, xmmB, 0); cc.evex().vpsraw(ymmA, ymmB, xmmC); cc.evex().vpsraw(ymmA, ymmB, 0); cc.evex().vpsraw(zmmA, zmmB, xmmC); cc.evex().vpsraw(zmmA, zmmB, 0); cc.evex().vpsrld(xmmA, xmmB, xmmC); cc.evex().vpsrld(xmmA, xmmB, 0); cc.evex().vpsrld(ymmA, ymmB, xmmC); cc.evex().vpsrld(ymmA, ymmB, 0); cc.evex().vpsrld(zmmA, zmmB, xmmC); cc.evex().vpsrld(zmmA, zmmB, 0); cc.evex().vpsrldq(xmmA, xmmB, 0); cc.evex().vpsrldq(ymmA, ymmB, 0); cc.evex().vpsrldq(zmmA, zmmB, 0); cc.evex().vpsrlq(xmmA, xmmB, xmmC); cc.evex().vpsrlq(xmmA, xmmB, 0); cc.evex().vpsrlq(ymmA, ymmB, xmmC); cc.evex().vpsrlq(ymmA, ymmB, 0); cc.evex().vpsrlq(zmmA, zmmB, xmmC); cc.evex().vpsrlq(zmmA, zmmB, 0); cc.evex().vpsrlvd(xmmA, xmmB, xmmC); cc.evex().vpsrlvd(ymmA, ymmB, ymmC); cc.evex().vpsrlvd(zmmA, zmmB, zmmC); cc.evex().vpsrlvq(xmmA, xmmB, xmmC); cc.evex().vpsrlvq(ymmA, ymmB, ymmC); cc.evex().vpsrlvq(zmmA, zmmB, zmmC); cc.evex().vpsrlvw(xmmA, xmmB, xmmC); cc.evex().vpsrlvw(ymmA, ymmB, ymmC); cc.evex().vpsrlvw(zmmA, zmmB, zmmC); cc.evex().vpsrlw(xmmA, xmmB, xmmC); cc.evex().vpsrlw(xmmA, xmmB, 0); cc.evex().vpsrlw(ymmA, ymmB, xmmC); cc.evex().vpsrlw(ymmA, ymmB, 0); cc.evex().vpsrlw(zmmA, zmmB, xmmC); cc.evex().vpsrlw(zmmA, zmmB, 0); cc.evex().vpsubb(xmmA, xmmB, xmmC); cc.evex().vpsubb(ymmA, ymmB, ymmC); cc.evex().vpsubb(zmmA, zmmB, zmmC); cc.evex().vpsubd(xmmA, xmmB, xmmC); cc.evex().vpsubd(ymmA, ymmB, ymmC); cc.evex().vpsubd(zmmA, zmmB, zmmC); cc.evex().vpsubq(xmmA, xmmB, xmmC); cc.evex().vpsubq(ymmA, ymmB, ymmC); cc.evex().vpsubq(zmmA, zmmB, zmmC); cc.evex().vpsubsb(xmmA, xmmB, xmmC); cc.evex().vpsubsb(ymmA, ymmB, ymmC); cc.evex().vpsubsb(zmmA, zmmB, zmmC); cc.evex().vpsubsw(xmmA, xmmB, xmmC); cc.evex().vpsubsw(ymmA, ymmB, ymmC); cc.evex().vpsubsw(zmmA, zmmB, zmmC); cc.evex().vpsubusb(xmmA, xmmB, xmmC); cc.evex().vpsubusb(ymmA, ymmB, ymmC); cc.evex().vpsubusb(zmmA, zmmB, zmmC); cc.evex().vpsubusw(xmmA, xmmB, xmmC); cc.evex().vpsubusw(ymmA, ymmB, ymmC); cc.evex().vpsubusw(zmmA, zmmB, zmmC); cc.evex().vpsubw(xmmA, xmmB, xmmC); cc.evex().vpsubw(ymmA, ymmB, ymmC); cc.evex().vpsubw(zmmA, zmmB, zmmC); cc.evex().vpternlogd(xmmA, xmmB, xmmC, 0); cc.evex().vpternlogd(ymmA, ymmB, ymmC, 0); cc.evex().vpternlogd(zmmA, zmmB, zmmC, 0); cc.evex().vpternlogq(xmmA, xmmB, xmmC, 0); cc.evex().vpternlogq(ymmA, ymmB, ymmC, 0); cc.evex().vpternlogq(zmmA, zmmB, zmmC, 0); cc.evex().vptestmb(kA, xmmB, xmmC); cc.evex().vptestmb(kA, ymmB, ymmC); cc.evex().vptestmb(kA, zmmB, zmmC); cc.evex().vptestmd(kA, xmmB, xmmC); cc.evex().vptestmd(kA, ymmB, ymmC); cc.evex().vptestmd(kA, zmmB, zmmC); cc.evex().vptestmq(kA, xmmB, xmmC); cc.evex().vptestmq(kA, ymmB, ymmC); cc.evex().vptestmq(kA, zmmB, zmmC); cc.evex().vptestmw(kA, xmmB, xmmC); cc.evex().vptestmw(kA, ymmB, ymmC); cc.evex().vptestmw(kA, zmmB, zmmC); cc.evex().vptestnmb(kA, xmmB, xmmC); cc.evex().vptestnmb(kA, ymmB, ymmC); cc.evex().vptestnmb(kA, zmmB, zmmC); cc.evex().vptestnmd(kA, xmmB, xmmC); cc.evex().vptestnmd(kA, ymmB, ymmC); cc.evex().vptestnmd(kA, zmmB, zmmC); cc.evex().vptestnmq(kA, xmmB, xmmC); cc.evex().vptestnmq(kA, ymmB, ymmC); cc.evex().vptestnmq(kA, zmmB, zmmC); cc.evex().vptestnmw(kA, xmmB, xmmC); cc.evex().vptestnmw(kA, ymmB, ymmC); cc.evex().vptestnmw(kA, zmmB, zmmC); cc.evex().vpunpckhbw(xmmA, xmmB, xmmC); cc.evex().vpunpckhbw(ymmA, ymmB, ymmC); cc.evex().vpunpckhbw(zmmA, zmmB, zmmC); cc.evex().vpunpckhdq(xmmA, xmmB, xmmC); cc.evex().vpunpckhdq(ymmA, ymmB, ymmC); cc.evex().vpunpckhdq(zmmA, zmmB, zmmC); cc.evex().vpunpckhqdq(xmmA, xmmB, xmmC); cc.evex().vpunpckhqdq(ymmA, ymmB, ymmC); cc.evex().vpunpckhqdq(zmmA, zmmB, zmmC); cc.evex().vpunpckhwd(xmmA, xmmB, xmmC); cc.evex().vpunpckhwd(ymmA, ymmB, ymmC); cc.evex().vpunpckhwd(zmmA, zmmB, zmmC); cc.evex().vpunpcklbw(xmmA, xmmB, xmmC); cc.evex().vpunpcklbw(ymmA, ymmB, ymmC); cc.evex().vpunpcklbw(zmmA, zmmB, zmmC); cc.evex().vpunpckldq(xmmA, xmmB, xmmC); cc.evex().vpunpckldq(ymmA, ymmB, ymmC); cc.evex().vpunpckldq(zmmA, zmmB, zmmC); cc.evex().vpunpcklqdq(xmmA, xmmB, xmmC); cc.evex().vpunpcklqdq(ymmA, ymmB, ymmC); cc.evex().vpunpcklqdq(zmmA, zmmB, zmmC); cc.evex().vpunpcklwd(xmmA, xmmB, xmmC); cc.evex().vpunpcklwd(ymmA, ymmB, ymmC); cc.evex().vpunpcklwd(zmmA, zmmB, zmmC); cc.evex().vpxord(xmmA, xmmB, xmmC); cc.evex().vpxord(ymmA, ymmB, ymmC); cc.evex().vpxord(zmmA, zmmB, zmmC); cc.evex().vpxorq(xmmA, xmmB, xmmC); cc.evex().vpxorq(ymmA, ymmB, ymmC); cc.evex().vpxorq(zmmA, zmmB, zmmC); cc.evex().vrangepd(xmmA, xmmB, xmmC, 0); cc.evex().vrangepd(ymmA, ymmB, ymmC, 0); cc.evex().vrangepd(zmmA, zmmB, zmmC, 0); cc.evex().vrangeps(xmmA, xmmB, xmmC, 0); cc.evex().vrangeps(ymmA, ymmB, ymmC, 0); cc.evex().vrangeps(zmmA, zmmB, zmmC, 0); cc.evex().vrangesd(xmmA, xmmB, xmmC, 0); cc.evex().vrangess(xmmA, xmmB, xmmC, 0); cc.evex().vrcp14pd(xmmA, xmmB); cc.evex().vrcp14pd(ymmA, ymmB); cc.evex().vrcp14pd(zmmA, zmmB); cc.evex().vrcp14ps(xmmA, xmmB); cc.evex().vrcp14ps(ymmA, ymmB); cc.evex().vrcp14ps(zmmA, zmmB); cc.evex().vrcp14sd(xmmA, xmmB, xmmC); cc.evex().vrcp14ss(xmmA, xmmB, xmmC); cc.evex().vrcp28pd(zmmA, zmmB); cc.evex().vrcp28ps(zmmA, zmmB); cc.evex().vrcp28sd(xmmA, xmmB, xmmC); cc.evex().vrcp28ss(xmmA, xmmB, xmmC); cc.evex().vreducepd(xmmA, xmmB, 0); cc.evex().vreducepd(ymmA, ymmB, 0); cc.evex().vreducepd(zmmA, zmmB, 0); cc.evex().vreduceps(xmmA, xmmB, 0); cc.evex().vreduceps(ymmA, ymmB, 0); cc.evex().vreduceps(zmmA, zmmB, 0); cc.evex().vreducesd(xmmA, xmmB, xmmC, 0); cc.evex().vreducess(xmmA, xmmB, xmmC, 0); cc.evex().vrndscalepd(xmmA, xmmB, 0); cc.evex().vrndscalepd(ymmA, ymmB, 0); cc.evex().vrndscalepd(zmmA, zmmB, 0); cc.evex().vrndscaleps(xmmA, xmmB, 0); cc.evex().vrndscaleps(ymmA, ymmB, 0); cc.evex().vrndscaleps(zmmA, zmmB, 0); cc.evex().vrndscalesd(xmmA, xmmB, xmmC, 0); cc.evex().vrndscaless(xmmA, xmmB, xmmC, 0); cc.evex().vrsqrt14pd(xmmA, xmmB); cc.evex().vrsqrt14pd(ymmA, ymmB); cc.evex().vrsqrt14pd(zmmA, zmmB); cc.evex().vrsqrt14ps(xmmA, xmmB); cc.evex().vrsqrt14ps(ymmA, ymmB); cc.evex().vrsqrt14ps(zmmA, zmmB); cc.evex().vrsqrt14sd(xmmA, xmmB, xmmC); cc.evex().vrsqrt14ss(xmmA, xmmB, xmmC); cc.evex().vrsqrt28pd(zmmA, zmmB); cc.evex().vrsqrt28ps(zmmA, zmmB); cc.evex().vrsqrt28sd(xmmA, xmmB, xmmC); cc.evex().vrsqrt28ss(xmmA, xmmB, xmmC); cc.evex().vscalefpd(xmmA, xmmB, xmmC); cc.evex().vscalefpd(ymmA, ymmB, ymmC); cc.evex().vscalefpd(zmmA, zmmB, zmmC); cc.evex().vscalefps(xmmA, xmmB, xmmC); cc.evex().vscalefps(ymmA, ymmB, ymmC); cc.evex().vscalefps(zmmA, zmmB, zmmC); cc.evex().vscalefsd(xmmA, xmmB, xmmC); cc.evex().vscalefss(xmmA, xmmB, xmmC); cc.evex().vshuff32x4(ymmA, ymmB, ymmC, 0); cc.evex().vshuff32x4(zmmA, zmmB, zmmC, 0); cc.evex().vshuff64x2(ymmA, ymmB, ymmC, 0); cc.evex().vshuff64x2(zmmA, zmmB, zmmC, 0); cc.evex().vshufi32x4(ymmA, ymmB, ymmC, 0); cc.evex().vshufi32x4(zmmA, zmmB, zmmC, 0); cc.evex().vshufi64x2(ymmA, ymmB, ymmC, 0); cc.evex().vshufi64x2(zmmA, zmmB, zmmC, 0); cc.evex().vshufpd(xmmA, xmmB, xmmC, 0); cc.evex().vshufpd(ymmA, ymmB, ymmC, 0); cc.evex().vshufpd(zmmA, zmmB, zmmC, 0); cc.evex().vshufps(xmmA, xmmB, xmmC, 0); cc.evex().vshufps(ymmA, ymmB, ymmC, 0); cc.evex().vshufps(zmmA, zmmB, zmmC, 0); cc.evex().vsqrtpd(xmmA, xmmB); cc.evex().vsqrtpd(ymmA, ymmB); cc.evex().vsqrtpd(zmmA, zmmB); cc.evex().vsqrtps(xmmA, xmmB); cc.evex().vsqrtps(ymmA, ymmB); cc.evex().vsqrtps(zmmA, zmmB); cc.evex().vsqrtsd(xmmA, xmmB, xmmC); cc.evex().vsqrtss(xmmA, xmmB, xmmC); cc.evex().vsubpd(xmmA, xmmB, xmmC); cc.evex().vsubpd(ymmA, ymmB, ymmC); cc.evex().vsubpd(zmmA, zmmB, zmmC); cc.evex().vsubps(xmmA, xmmB, xmmC); cc.evex().vsubps(ymmA, ymmB, ymmC); cc.evex().vsubps(zmmA, zmmB, zmmC); cc.evex().vsubsd(xmmA, xmmB, xmmC); cc.evex().vsubss(xmmA, xmmB, xmmC); cc.evex().vucomisd(xmmA, xmmB); cc.evex().vucomiss(xmmA, xmmB); cc.evex().vunpckhpd(xmmA, xmmB, xmmC); cc.evex().vunpckhpd(ymmA, ymmB, ymmC); cc.evex().vunpckhpd(zmmA, zmmB, zmmC); cc.evex().vunpckhps(xmmA, xmmB, xmmC); cc.evex().vunpckhps(ymmA, ymmB, ymmC); cc.evex().vunpckhps(zmmA, zmmB, zmmC); cc.evex().vunpcklpd(xmmA, xmmB, xmmC); cc.evex().vunpcklpd(ymmA, ymmB, ymmC); cc.evex().vunpcklpd(zmmA, zmmB, zmmC); cc.evex().vunpcklps(xmmA, xmmB, xmmC); cc.evex().vunpcklps(ymmA, ymmB, ymmC); cc.evex().vunpcklps(zmmA, zmmB, zmmC); cc.evex().vxorpd(xmmA, xmmB, xmmC); cc.evex().vxorpd(ymmA, ymmB, ymmC); cc.evex().vxorpd(zmmA, zmmB, zmmC); cc.evex().vxorps(xmmA, xmmB, xmmC); cc.evex().vxorps(ymmA, ymmB, ymmC); cc.evex().vxorps(zmmA, zmmB, zmmC); } template static void generateAvx512SequenceInternalRegMem( Emitter& cc, const x86::Gp& gp, const x86::KReg& kA, const x86::KReg& kB, const x86::KReg& kC, const x86::Vec& vecA, const x86::Vec& vecB, const x86::Vec& vecC, const x86::Vec& vecD) { DebugUtils::unused(kC); x86::Gp gpd = gp.r32(); x86::Gp gpq = gp.r64(); x86::Gp gpz = cc.is32Bit() ? gpd : gpq; x86::Xmm xmmA = vecA.xmm(); x86::Xmm xmmB = vecB.xmm(); x86::Xmm xmmC = vecC.xmm(); x86::Xmm xmmD = vecD.xmm(); x86::Ymm ymmA = vecA.ymm(); x86::Ymm ymmB = vecB.ymm(); x86::Ymm ymmD = vecD.ymm(); x86::Zmm zmmA = vecA.zmm(); x86::Zmm zmmB = vecB.zmm(); x86::Zmm zmmD = vecD.zmm(); x86::Mem m = x86::ptr(gpz); x86::Mem m32 = x86::dword_ptr(gpz); x86::Mem m64 = x86::qword_ptr(gpz); x86::Mem m128 = x86::xmmword_ptr(gpz); x86::Mem m256 = x86::ymmword_ptr(gpz); x86::Mem m512 = x86::zmmword_ptr(gpz); x86::Mem vx_ptr = x86::ptr(gpz, xmmD); x86::Mem vy_ptr = x86::ptr(gpz, ymmD); x86::Mem vz_ptr = x86::ptr(gpz, zmmD); cc.xor_(gpd, gpd); cc.vxorps(xmmA, xmmA, xmmA); cc.vxorps(xmmB, xmmB, xmmB); cc.vxorps(xmmC, xmmC, xmmC); cc.vxorps(xmmD, xmmD, xmmD); cc.kmovb(kA, m); cc.kmovb(m, kB); cc.kmovd(kA, m); cc.kmovd(m, kB); cc.kmovq(kA, m); cc.kmovq(m, kB); cc.kmovw(kA, m); cc.kmovw(m, kB); cc.evex().vaddpd(xmmA, xmmB, m); cc.evex().vaddpd(ymmA, ymmB, m); cc.evex().vaddpd(zmmA, zmmB, m); cc.evex().vaddps(xmmA, xmmB, m); cc.evex().vaddps(ymmA, ymmB, m); cc.evex().vaddps(zmmA, zmmB, m); cc.evex().vaddsd(xmmA, xmmB, m); cc.evex().vaddss(xmmA, xmmB, m); cc.evex().valignd(xmmA, xmmB, m, 0); cc.evex().valignd(ymmA, ymmB, m, 0); cc.evex().valignd(zmmA, zmmB, m, 0); cc.evex().valignq(xmmA, xmmB, m, 0); cc.evex().valignq(ymmA, ymmB, m, 0); cc.evex().valignq(zmmA, zmmB, m, 0); cc.evex().vandnpd(xmmA, xmmB, m); cc.evex().vandnpd(ymmA, ymmB, m); cc.evex().vandnpd(zmmA, zmmB, m); cc.evex().vandnps(xmmA, xmmB, m); cc.evex().vandnps(ymmA, ymmB, m); cc.evex().vandnps(zmmA, zmmB, m); cc.evex().vandpd(xmmA, xmmB, m); cc.evex().vandpd(ymmA, ymmB, m); cc.evex().vandpd(zmmA, zmmB, m); cc.evex().vandps(xmmA, xmmB, m); cc.evex().vandps(ymmA, ymmB, m); cc.evex().vandps(zmmA, zmmB, m); cc.evex().vblendmpd(xmmA, xmmB, m); cc.evex().vblendmpd(ymmA, ymmB, m); cc.evex().vblendmpd(zmmA, zmmB, m); cc.evex().vblendmps(xmmA, xmmB, m); cc.evex().vblendmps(ymmA, ymmB, m); cc.evex().vblendmps(zmmA, zmmB, m); cc.evex().vbroadcastf32x2(ymmA, m); cc.evex().vbroadcastf32x2(zmmA, m); cc.evex().vbroadcastf32x4(ymmA, m); cc.evex().vbroadcastf32x4(zmmA, m); cc.evex().vbroadcastf32x8(zmmA, m); cc.evex().vbroadcastf64x2(ymmA, m); cc.evex().vbroadcastf64x2(zmmA, m); cc.evex().vbroadcastf64x4(zmmA, m); cc.evex().vbroadcasti32x2(xmmA, m); cc.evex().vbroadcasti32x2(ymmA, m); cc.evex().vbroadcasti32x2(zmmA, m); cc.evex().vbroadcasti32x4(ymmA, m); cc.evex().vbroadcasti32x4(zmmA, m); cc.evex().vbroadcasti32x8(zmmA, m); cc.evex().vbroadcasti64x2(ymmA, m); cc.evex().vbroadcasti64x2(zmmA, m); cc.evex().vbroadcasti64x4(zmmA, m); cc.evex().vbroadcastsd(ymmA, m); cc.evex().vbroadcastsd(zmmA, m); cc.evex().vbroadcastss(xmmA, m); cc.evex().vbroadcastss(ymmA, m); cc.evex().vbroadcastss(zmmA, m); cc.evex().vcmppd(kA, xmmB, m, 0); cc.evex().vcmppd(kA, ymmB, m, 0); cc.evex().vcmppd(kA, zmmB, m, 0); cc.evex().vcmpps(kA, xmmB, m, 0); cc.evex().vcmpps(kA, ymmB, m, 0); cc.evex().vcmpps(kA, zmmB, m, 0); cc.evex().vcmpsd(kA, xmmB, m, 0); cc.evex().vcmpss(kA, xmmB, m, 0); cc.evex().vcomisd(xmmA, m); cc.evex().vcomiss(xmmA, m); cc.evex().vcompresspd(m, xmmB); cc.evex().vcompresspd(m, ymmB); cc.evex().vcompresspd(m, zmmB); cc.evex().vcompressps(m, xmmB); cc.evex().vcompressps(m, ymmB); cc.evex().vcompressps(m, zmmB); cc.evex().vcvtdq2pd(xmmA, m); cc.evex().vcvtdq2pd(ymmA, m); cc.evex().vcvtdq2pd(zmmA, m); cc.evex().vcvtdq2ps(xmmA, m); cc.evex().vcvtdq2ps(ymmA, m); cc.evex().vcvtdq2ps(zmmA, m); cc.evex().vcvtpd2dq(xmmA, m128); cc.evex().vcvtpd2dq(xmmA, m256); cc.evex().vcvtpd2dq(ymmA, m512); cc.evex().vcvtpd2qq(xmmA, m); cc.evex().vcvtpd2qq(ymmA, m); cc.evex().vcvtpd2qq(zmmA, m); cc.evex().vcvtpd2udq(xmmA, m128); cc.evex().vcvtpd2udq(xmmA, m256); cc.evex().vcvtpd2udq(ymmA, m512); cc.evex().vcvtpd2uqq(xmmA, m); cc.evex().vcvtpd2uqq(ymmA, m); cc.evex().vcvtpd2uqq(zmmA, m); cc.evex().vcvtph2ps(xmmA, m); cc.evex().vcvtph2ps(ymmA, m); cc.evex().vcvtph2ps(zmmA, m); cc.evex().vcvtps2dq(xmmA, m); cc.evex().vcvtps2dq(ymmA, m); cc.evex().vcvtps2dq(zmmA, m); cc.evex().vcvtps2pd(xmmA, m); cc.evex().vcvtps2pd(ymmA, m); cc.evex().vcvtps2pd(zmmA, m); cc.evex().vcvtps2ph(m, xmmB, 0); cc.evex().vcvtps2ph(m, ymmB, 0); cc.evex().vcvtps2ph(m, zmmB, 0); cc.evex().vcvtps2qq(xmmA, m); cc.evex().vcvtps2qq(ymmA, m); cc.evex().vcvtps2qq(zmmA, m); cc.evex().vcvtps2udq(xmmA, m); cc.evex().vcvtps2udq(ymmA, m); cc.evex().vcvtps2udq(zmmA, m); cc.evex().vcvtps2uqq(xmmA, m); cc.evex().vcvtps2uqq(ymmA, m); cc.evex().vcvtps2uqq(zmmA, m); cc.evex().vcvtqq2pd(xmmA, m); cc.evex().vcvtqq2pd(ymmA, m); cc.evex().vcvtqq2pd(zmmA, m); cc.evex().vcvtqq2ps(xmmA, m128); cc.evex().vcvtqq2ps(xmmA, m256); cc.evex().vcvtqq2ps(ymmA, m512); cc.evex().vcvtsd2si(gpd, m); cc.evex().vcvtsd2si(gpz, m); cc.evex().vcvtsd2ss(xmmA, xmmB, m); cc.evex().vcvtsd2usi(gpd, m); cc.evex().vcvtsd2usi(gpz, m); cc.evex().vcvtsi2sd(xmmA, xmmB, m32); if (cc.is64Bit()) cc.evex().vcvtsi2sd(xmmA, xmmB, m64); cc.evex().vcvtsi2ss(xmmA, xmmB, m32); if (cc.is64Bit()) cc.evex().vcvtsi2ss(xmmA, xmmB, m64); cc.evex().vcvtss2sd(xmmA, xmmB, m); cc.evex().vcvtss2si(gpd, m); cc.evex().vcvtss2si(gpz, m); cc.evex().vcvtss2usi(gpd, m); cc.evex().vcvtss2usi(gpz, m); cc.evex().vcvttpd2dq(xmmA, m128); cc.evex().vcvttpd2dq(xmmA, m256); cc.evex().vcvttpd2dq(ymmA, m512); cc.evex().vcvttpd2qq(xmmA, m); cc.evex().vcvttpd2qq(ymmA, m); cc.evex().vcvttpd2qq(zmmA, m); cc.evex().vcvttpd2udq(xmmA, m128); cc.evex().vcvttpd2udq(xmmA, m256); cc.evex().vcvttpd2udq(ymmA, m512); cc.evex().vcvttpd2uqq(xmmA, m); cc.evex().vcvttpd2uqq(ymmA, m); cc.evex().vcvttpd2uqq(zmmA, m); cc.evex().vcvttps2dq(xmmA, m); cc.evex().vcvttps2dq(ymmA, m); cc.evex().vcvttps2dq(zmmA, m); cc.evex().vcvttps2qq(xmmA, m); cc.evex().vcvttps2qq(ymmA, m); cc.evex().vcvttps2qq(zmmA, m); cc.evex().vcvttps2udq(xmmA, m); cc.evex().vcvttps2udq(ymmA, m); cc.evex().vcvttps2udq(zmmA, m); cc.evex().vcvttps2uqq(xmmA, m); cc.evex().vcvttps2uqq(ymmA, m); cc.evex().vcvttps2uqq(zmmA, m); cc.evex().vcvttsd2si(gpd, m); cc.evex().vcvttsd2si(gpz, m); cc.evex().vcvttsd2usi(gpd, m); cc.evex().vcvttsd2usi(gpz, m); cc.evex().vcvttss2si(gpd, m); cc.evex().vcvttss2si(gpz, m); cc.evex().vcvttss2usi(gpd, m); cc.evex().vcvttss2usi(gpz, m); cc.evex().vcvtudq2pd(xmmA, m); cc.evex().vcvtudq2pd(ymmA, m); cc.evex().vcvtudq2pd(zmmA, m); cc.evex().vcvtudq2ps(xmmA, m); cc.evex().vcvtudq2ps(ymmA, m); cc.evex().vcvtudq2ps(zmmA, m); cc.evex().vcvtuqq2pd(xmmA, m); cc.evex().vcvtuqq2pd(ymmA, m); cc.evex().vcvtuqq2pd(zmmA, m); cc.evex().vcvtuqq2ps(xmmA, m128); cc.evex().vcvtuqq2ps(xmmA, m256); cc.evex().vcvtuqq2ps(ymmA, m512); cc.evex().vcvtusi2sd(xmmA, xmmB, m32); if (cc.is64Bit()) cc.evex().vcvtusi2sd(xmmA, xmmB, m64); cc.evex().vcvtusi2ss(xmmA, xmmB, m32); if (cc.is64Bit()) cc.evex().vcvtusi2ss(xmmA, xmmB, m64); cc.evex().vdbpsadbw(xmmA, xmmB, m, 0); cc.evex().vdbpsadbw(ymmA, ymmB, m, 0); cc.evex().vdbpsadbw(zmmA, zmmB, m, 0); cc.evex().vdivpd(xmmA, xmmB, m); cc.evex().vdivpd(ymmA, ymmB, m); cc.evex().vdivpd(zmmA, zmmB, m); cc.evex().vdivps(xmmA, xmmB, m); cc.evex().vdivps(ymmA, ymmB, m); cc.evex().vdivps(zmmA, zmmB, m); cc.evex().vdivsd(xmmA, xmmB, m); cc.evex().vdivss(xmmA, xmmB, m); cc.evex().vexp2pd(zmmA, m); cc.evex().vexp2ps(zmmA, m); cc.evex().vexpandpd(xmmA, m); cc.evex().vexpandpd(ymmA, m); cc.evex().vexpandpd(zmmA, m); cc.evex().vexpandps(xmmA, m); cc.evex().vexpandps(ymmA, m); cc.evex().vexpandps(zmmA, m); cc.evex().vextractf32x4(m, ymmB, 0); cc.evex().vextractf32x4(m, zmmB, 0); cc.evex().vextractf32x8(m, zmmB, 0); cc.evex().vextractf64x2(m, ymmB, 0); cc.evex().vextractf64x2(m, zmmB, 0); cc.evex().vextractf64x4(m, zmmB, 0); cc.evex().vextracti32x4(m, ymmB, 0); cc.evex().vextracti32x4(m, zmmB, 0); cc.evex().vextracti32x8(m, zmmB, 0); cc.evex().vextracti64x2(m, ymmB, 0); cc.evex().vextracti64x2(m, zmmB, 0); cc.evex().vextracti64x4(m, zmmB, 0); cc.evex().vextractps(m, xmmB, 0); cc.evex().vfixupimmpd(xmmA, xmmB, m, 0); cc.evex().vfixupimmpd(ymmA, ymmB, m, 0); cc.evex().vfixupimmpd(zmmA, zmmB, m, 0); cc.evex().vfixupimmps(xmmA, xmmB, m, 0); cc.evex().vfixupimmps(ymmA, ymmB, m, 0); cc.evex().vfixupimmps(zmmA, zmmB, m, 0); cc.evex().vfixupimmsd(xmmA, xmmB, m, 0); cc.evex().vfixupimmss(xmmA, xmmB, m, 0); cc.evex().vfmadd132pd(xmmA, xmmB, m); cc.evex().vfmadd132pd(ymmA, ymmB, m); cc.evex().vfmadd132pd(zmmA, zmmB, m); cc.evex().vfmadd132ps(xmmA, xmmB, m); cc.evex().vfmadd132ps(ymmA, ymmB, m); cc.evex().vfmadd132ps(zmmA, zmmB, m); cc.evex().vfmadd132sd(xmmA, xmmB, m); cc.evex().vfmadd132ss(xmmA, xmmB, m); cc.evex().vfmadd213pd(xmmA, xmmB, m); cc.evex().vfmadd213pd(ymmA, ymmB, m); cc.evex().vfmadd213pd(zmmA, zmmB, m); cc.evex().vfmadd213ps(xmmA, xmmB, m); cc.evex().vfmadd213ps(ymmA, ymmB, m); cc.evex().vfmadd213ps(zmmA, zmmB, m); cc.evex().vfmadd213sd(xmmA, xmmB, m); cc.evex().vfmadd213ss(xmmA, xmmB, m); cc.evex().vfmadd231pd(xmmA, xmmB, m); cc.evex().vfmadd231pd(ymmA, ymmB, m); cc.evex().vfmadd231pd(zmmA, zmmB, m); cc.evex().vfmadd231ps(xmmA, xmmB, m); cc.evex().vfmadd231ps(ymmA, ymmB, m); cc.evex().vfmadd231ps(zmmA, zmmB, m); cc.evex().vfmadd231sd(xmmA, xmmB, m); cc.evex().vfmadd231ss(xmmA, xmmB, m); cc.evex().vfmaddsub132pd(xmmA, xmmB, m); cc.evex().vfmaddsub132pd(ymmA, ymmB, m); cc.evex().vfmaddsub132pd(zmmA, zmmB, m); cc.evex().vfmaddsub132ps(xmmA, xmmB, m); cc.evex().vfmaddsub132ps(ymmA, ymmB, m); cc.evex().vfmaddsub132ps(zmmA, zmmB, m); cc.evex().vfmaddsub213pd(xmmA, xmmB, m); cc.evex().vfmaddsub213pd(ymmA, ymmB, m); cc.evex().vfmaddsub213pd(zmmA, zmmB, m); cc.evex().vfmaddsub213ps(xmmA, xmmB, m); cc.evex().vfmaddsub213ps(ymmA, ymmB, m); cc.evex().vfmaddsub213ps(zmmA, zmmB, m); cc.evex().vfmaddsub231pd(xmmA, xmmB, m); cc.evex().vfmaddsub231pd(ymmA, ymmB, m); cc.evex().vfmaddsub231pd(zmmA, zmmB, m); cc.evex().vfmaddsub231ps(xmmA, xmmB, m); cc.evex().vfmaddsub231ps(ymmA, ymmB, m); cc.evex().vfmaddsub231ps(zmmA, zmmB, m); cc.evex().vfmsub132pd(xmmA, xmmB, m); cc.evex().vfmsub132pd(ymmA, ymmB, m); cc.evex().vfmsub132pd(zmmA, zmmB, m); cc.evex().vfmsub132ps(xmmA, xmmB, m); cc.evex().vfmsub132ps(ymmA, ymmB, m); cc.evex().vfmsub132ps(zmmA, zmmB, m); cc.evex().vfmsub132sd(xmmA, xmmB, m); cc.evex().vfmsub132ss(xmmA, xmmB, m); cc.evex().vfmsub213pd(xmmA, xmmB, m); cc.evex().vfmsub213pd(ymmA, ymmB, m); cc.evex().vfmsub213pd(zmmA, zmmB, m); cc.evex().vfmsub213ps(xmmA, xmmB, m); cc.evex().vfmsub213ps(ymmA, ymmB, m); cc.evex().vfmsub213ps(zmmA, zmmB, m); cc.evex().vfmsub213sd(xmmA, xmmB, m); cc.evex().vfmsub213ss(xmmA, xmmB, m); cc.evex().vfmsub231pd(xmmA, xmmB, m); cc.evex().vfmsub231pd(ymmA, ymmB, m); cc.evex().vfmsub231pd(zmmA, zmmB, m); cc.evex().vfmsub231ps(xmmA, xmmB, m); cc.evex().vfmsub231ps(ymmA, ymmB, m); cc.evex().vfmsub231ps(zmmA, zmmB, m); cc.evex().vfmsub231sd(xmmA, xmmB, m); cc.evex().vfmsub231ss(xmmA, xmmB, m); cc.evex().vfmsubadd132pd(xmmA, xmmB, m); cc.evex().vfmsubadd132pd(ymmA, ymmB, m); cc.evex().vfmsubadd132pd(zmmA, zmmB, m); cc.evex().vfmsubadd132ps(xmmA, xmmB, m); cc.evex().vfmsubadd132ps(ymmA, ymmB, m); cc.evex().vfmsubadd132ps(zmmA, zmmB, m); cc.evex().vfmsubadd213pd(xmmA, xmmB, m); cc.evex().vfmsubadd213pd(ymmA, ymmB, m); cc.evex().vfmsubadd213pd(zmmA, zmmB, m); cc.evex().vfmsubadd213ps(xmmA, xmmB, m); cc.evex().vfmsubadd213ps(ymmA, ymmB, m); cc.evex().vfmsubadd213ps(zmmA, zmmB, m); cc.evex().vfmsubadd231pd(xmmA, xmmB, m); cc.evex().vfmsubadd231pd(ymmA, ymmB, m); cc.evex().vfmsubadd231pd(zmmA, zmmB, m); cc.evex().vfmsubadd231ps(xmmA, xmmB, m); cc.evex().vfmsubadd231ps(ymmA, ymmB, m); cc.evex().vfmsubadd231ps(zmmA, zmmB, m); cc.evex().vfnmadd132pd(xmmA, xmmB, m); cc.evex().vfnmadd132pd(ymmA, ymmB, m); cc.evex().vfnmadd132pd(zmmA, zmmB, m); cc.evex().vfnmadd132ps(xmmA, xmmB, m); cc.evex().vfnmadd132ps(ymmA, ymmB, m); cc.evex().vfnmadd132ps(zmmA, zmmB, m); cc.evex().vfnmadd132sd(xmmA, xmmB, m); cc.evex().vfnmadd132ss(xmmA, xmmB, m); cc.evex().vfnmadd213pd(xmmA, xmmB, m); cc.evex().vfnmadd213pd(ymmA, ymmB, m); cc.evex().vfnmadd213pd(zmmA, zmmB, m); cc.evex().vfnmadd213ps(xmmA, xmmB, m); cc.evex().vfnmadd213ps(ymmA, ymmB, m); cc.evex().vfnmadd213ps(zmmA, zmmB, m); cc.evex().vfnmadd213sd(xmmA, xmmB, m); cc.evex().vfnmadd213ss(xmmA, xmmB, m); cc.evex().vfnmadd231pd(xmmA, xmmB, m); cc.evex().vfnmadd231pd(ymmA, ymmB, m); cc.evex().vfnmadd231pd(zmmA, zmmB, m); cc.evex().vfnmadd231ps(xmmA, xmmB, m); cc.evex().vfnmadd231ps(ymmA, ymmB, m); cc.evex().vfnmadd231ps(zmmA, zmmB, m); cc.evex().vfnmadd231sd(xmmA, xmmB, m); cc.evex().vfnmadd231ss(xmmA, xmmB, m); cc.evex().vfnmsub132pd(xmmA, xmmB, m); cc.evex().vfnmsub132pd(ymmA, ymmB, m); cc.evex().vfnmsub132pd(zmmA, zmmB, m); cc.evex().vfnmsub132ps(xmmA, xmmB, m); cc.evex().vfnmsub132ps(ymmA, ymmB, m); cc.evex().vfnmsub132ps(zmmA, zmmB, m); cc.evex().vfnmsub132sd(xmmA, xmmB, m); cc.evex().vfnmsub132ss(xmmA, xmmB, m); cc.evex().vfnmsub213pd(xmmA, xmmB, m); cc.evex().vfnmsub213pd(ymmA, ymmB, m); cc.evex().vfnmsub213pd(zmmA, zmmB, m); cc.evex().vfnmsub213ps(xmmA, xmmB, m); cc.evex().vfnmsub213ps(ymmA, ymmB, m); cc.evex().vfnmsub213ps(zmmA, zmmB, m); cc.evex().vfnmsub213sd(xmmA, xmmB, m); cc.evex().vfnmsub213ss(xmmA, xmmB, m); cc.evex().vfnmsub231pd(xmmA, xmmB, m); cc.evex().vfnmsub231pd(ymmA, ymmB, m); cc.evex().vfnmsub231pd(zmmA, zmmB, m); cc.evex().vfnmsub231ps(xmmA, xmmB, m); cc.evex().vfnmsub231ps(ymmA, ymmB, m); cc.evex().vfnmsub231ps(zmmA, zmmB, m); cc.evex().vfnmsub231sd(xmmA, xmmB, m); cc.evex().vfnmsub231ss(xmmA, xmmB, m); cc.evex().vfpclasspd(kA, m128, 0); cc.evex().vfpclasspd(kA, m256, 0); cc.evex().vfpclasspd(kA, m512, 0); cc.evex().vfpclassps(kA, m128, 0); cc.evex().vfpclassps(kA, m256, 0); cc.evex().vfpclassps(kA, m512, 0); cc.evex().vfpclasssd(kA, m, 0); cc.evex().vfpclassss(kA, m, 0); cc.evex().k(kA).vgatherdpd(xmmA, vx_ptr); cc.evex().k(kA).vgatherdpd(ymmA, vx_ptr); cc.evex().k(kA).vgatherdpd(zmmA, vy_ptr); cc.evex().k(kA).vgatherdps(xmmA, vx_ptr); cc.evex().k(kA).vgatherdps(ymmA, vy_ptr); cc.evex().k(kA).vgatherdps(zmmA, vz_ptr); cc.evex().k(kA).vgatherpf0dpd(vy_ptr); cc.evex().k(kA).vgatherpf0dps(vz_ptr); cc.evex().k(kA).vgatherpf0qpd(vz_ptr); cc.evex().k(kA).vgatherpf0qps(vz_ptr); cc.evex().k(kA).vgatherpf1dpd(vy_ptr); cc.evex().k(kA).vgatherpf1dps(vz_ptr); cc.evex().k(kA).vgatherpf1qpd(vz_ptr); cc.evex().k(kA).vgatherpf1qps(vz_ptr); cc.evex().k(kA).vgatherqpd(xmmA, vx_ptr); cc.evex().k(kA).vgatherqpd(ymmA, vy_ptr); cc.evex().k(kA).vgatherqpd(zmmA, vz_ptr); cc.evex().k(kA).vgatherqps(xmmA, vx_ptr); cc.evex().k(kA).vgatherqps(xmmA, vy_ptr); cc.evex().k(kA).vgatherqps(ymmA, vz_ptr); cc.evex().vgetexppd(xmmA, m); cc.evex().vgetexppd(ymmA, m); cc.evex().vgetexppd(zmmA, m); cc.evex().vgetexpps(xmmA, m); cc.evex().vgetexpps(ymmA, m); cc.evex().vgetexpps(zmmA, m); cc.evex().vgetexpsd(xmmA, xmmB, m); cc.evex().vgetexpss(xmmA, xmmB, m); cc.evex().vgetmantpd(xmmA, m, 0); cc.evex().vgetmantpd(ymmA, m, 0); cc.evex().vgetmantpd(zmmA, m, 0); cc.evex().vgetmantps(xmmA, m, 0); cc.evex().vgetmantps(ymmA, m, 0); cc.evex().vgetmantps(zmmA, m, 0); cc.evex().vgetmantsd(xmmA, xmmB, m, 0); cc.evex().vgetmantss(xmmA, xmmB, m, 0); cc.evex().vinsertf32x4(ymmA, ymmB, m, 0); cc.evex().vinsertf32x4(zmmA, zmmB, m, 0); cc.evex().vinsertf32x8(zmmA, zmmB, m, 0); cc.evex().vinsertf64x2(ymmA, ymmB, m, 0); cc.evex().vinsertf64x2(zmmA, zmmB, m, 0); cc.evex().vinsertf64x4(zmmA, zmmB, m, 0); cc.evex().vinserti32x4(ymmA, ymmB, m, 0); cc.evex().vinserti32x4(zmmA, zmmB, m, 0); cc.evex().vinserti32x8(zmmA, zmmB, m, 0); cc.evex().vinserti64x2(ymmA, ymmB, m, 0); cc.evex().vinserti64x2(zmmA, zmmB, m, 0); cc.evex().vinserti64x4(zmmA, zmmB, m, 0); cc.evex().vinsertps(xmmA, xmmB, m, 0); cc.evex().vmaxpd(xmmA, xmmB, m); cc.evex().vmaxpd(ymmA, ymmB, m); cc.evex().vmaxpd(zmmA, zmmB, m); cc.evex().vmaxps(xmmA, xmmB, m); cc.evex().vmaxps(ymmA, ymmB, m); cc.evex().vmaxps(zmmA, zmmB, m); cc.evex().vmaxsd(xmmA, xmmB, m); cc.evex().vmaxss(xmmA, xmmB, m); cc.evex().vminpd(xmmA, xmmB, m); cc.evex().vminpd(ymmA, ymmB, m); cc.evex().vminpd(zmmA, zmmB, m); cc.evex().vminps(xmmA, xmmB, m); cc.evex().vminps(ymmA, ymmB, m); cc.evex().vminps(zmmA, zmmB, m); cc.evex().vminsd(xmmA, xmmB, m); cc.evex().vminss(xmmA, xmmB, m); cc.evex().vmovapd(xmmA, m); cc.evex().vmovapd(m, xmmB); cc.evex().vmovapd(ymmA, m); cc.evex().vmovapd(m, ymmB); cc.evex().vmovapd(zmmA, m); cc.evex().vmovapd(m, zmmB); cc.evex().vmovaps(xmmA, m); cc.evex().vmovaps(m, xmmB); cc.evex().vmovaps(ymmA, m); cc.evex().vmovaps(m, ymmB); cc.evex().vmovaps(zmmA, m); cc.evex().vmovaps(m, zmmB); cc.evex().vmovd(m, xmmB); cc.evex().vmovd(xmmA, m); cc.evex().vmovddup(xmmA, m); cc.evex().vmovddup(ymmA, m); cc.evex().vmovddup(zmmA, m); cc.evex().vmovdqa32(xmmA, m); cc.evex().vmovdqa32(m, xmmB); cc.evex().vmovdqa32(ymmA, m); cc.evex().vmovdqa32(m, ymmB); cc.evex().vmovdqa32(zmmA, m); cc.evex().vmovdqa32(m, zmmB); cc.evex().vmovdqa64(xmmA, m); cc.evex().vmovdqa64(m, xmmB); cc.evex().vmovdqa64(ymmA, m); cc.evex().vmovdqa64(m, ymmB); cc.evex().vmovdqa64(zmmA, m); cc.evex().vmovdqa64(m, zmmB); cc.evex().vmovdqu16(xmmA, m); cc.evex().vmovdqu16(m, xmmB); cc.evex().vmovdqu16(ymmA, m); cc.evex().vmovdqu16(m, ymmB); cc.evex().vmovdqu16(zmmA, m); cc.evex().vmovdqu16(m, zmmB); cc.evex().vmovdqu32(xmmA, m); cc.evex().vmovdqu32(m, xmmB); cc.evex().vmovdqu32(ymmA, m); cc.evex().vmovdqu32(m, ymmB); cc.evex().vmovdqu32(zmmA, m); cc.evex().vmovdqu32(m, zmmB); cc.evex().vmovdqu64(xmmA, m); cc.evex().vmovdqu64(m, xmmB); cc.evex().vmovdqu64(ymmA, m); cc.evex().vmovdqu64(m, ymmB); cc.evex().vmovdqu64(zmmA, m); cc.evex().vmovdqu64(m, zmmB); cc.evex().vmovdqu8(xmmA, m); cc.evex().vmovdqu8(m, xmmB); cc.evex().vmovdqu8(ymmA, m); cc.evex().vmovdqu8(m, ymmB); cc.evex().vmovdqu8(zmmA, m); cc.evex().vmovdqu8(m, zmmB); cc.evex().vmovhpd(m, xmmB); cc.evex().vmovhpd(xmmA, xmmB, m); cc.evex().vmovhps(m, xmmB); cc.evex().vmovhps(xmmA, xmmB, m); cc.evex().vmovlpd(m, xmmB); cc.evex().vmovlpd(xmmA, xmmB, m); cc.evex().vmovlps(m, xmmB); cc.evex().vmovlps(xmmA, xmmB, m); cc.evex().vmovntdq(m, xmmB); cc.evex().vmovntdq(m, ymmB); cc.evex().vmovntdq(m, zmmB); cc.evex().vmovntdqa(xmmA, m); cc.evex().vmovntdqa(ymmA, m); cc.evex().vmovntdqa(zmmA, m); cc.evex().vmovntpd(m, xmmB); cc.evex().vmovntpd(m, ymmB); cc.evex().vmovntpd(m, zmmB); cc.evex().vmovntps(m, xmmB); cc.evex().vmovntps(m, ymmB); cc.evex().vmovntps(m, zmmB); cc.evex().vmovq(m, xmmB); cc.evex().vmovq(xmmA, m); cc.evex().vmovq(xmmA, m); cc.evex().vmovq(m, xmmB); cc.evex().vmovsd(m, xmmB); cc.evex().vmovsd(xmmA, m); cc.evex().vmovshdup(xmmA, m); cc.evex().vmovshdup(ymmA, m); cc.evex().vmovshdup(zmmA, m); cc.evex().vmovsldup(xmmA, m); cc.evex().vmovsldup(ymmA, m); cc.evex().vmovsldup(zmmA, m); cc.evex().vmovss(m, xmmB); cc.evex().vmovss(xmmA, m); cc.evex().vmovupd(xmmA, m); cc.evex().vmovupd(m, xmmB); cc.evex().vmovupd(ymmA, m); cc.evex().vmovupd(m, ymmB); cc.evex().vmovupd(zmmA, m); cc.evex().vmovupd(m, zmmB); cc.evex().vmovups(xmmA, m); cc.evex().vmovups(m, xmmB); cc.evex().vmovups(ymmA, m); cc.evex().vmovups(m, ymmB); cc.evex().vmovups(zmmA, m); cc.evex().vmovups(m, zmmB); cc.evex().vmulpd(xmmA, xmmB, m); cc.evex().vmulpd(ymmA, ymmB, m); cc.evex().vmulpd(zmmA, zmmB, m); cc.evex().vmulps(xmmA, xmmB, m); cc.evex().vmulps(ymmA, ymmB, m); cc.evex().vmulps(zmmA, zmmB, m); cc.evex().vmulsd(xmmA, xmmB, m); cc.evex().vmulss(xmmA, xmmB, m); cc.evex().vorpd(xmmA, xmmB, m); cc.evex().vorpd(ymmA, ymmB, m); cc.evex().vorpd(zmmA, zmmB, m); cc.evex().vorps(xmmA, xmmB, m); cc.evex().vorps(ymmA, ymmB, m); cc.evex().vorps(zmmA, zmmB, m); cc.evex().vpabsb(xmmA, m); cc.evex().vpabsb(ymmA, m); cc.evex().vpabsb(zmmA, m); cc.evex().vpabsd(xmmA, m); cc.evex().vpabsd(ymmA, m); cc.evex().vpabsd(zmmA, m); cc.evex().vpabsq(xmmA, m); cc.evex().vpabsq(ymmA, m); cc.evex().vpabsq(zmmA, m); cc.evex().vpabsw(xmmA, m); cc.evex().vpabsw(ymmA, m); cc.evex().vpabsw(zmmA, m); cc.evex().vpackssdw(xmmA, xmmB, m); cc.evex().vpackssdw(ymmA, ymmB, m); cc.evex().vpackssdw(zmmA, zmmB, m); cc.evex().vpacksswb(xmmA, xmmB, m); cc.evex().vpacksswb(ymmA, ymmB, m); cc.evex().vpacksswb(zmmA, zmmB, m); cc.evex().vpackusdw(xmmA, xmmB, m); cc.evex().vpackusdw(ymmA, ymmB, m); cc.evex().vpackusdw(zmmA, zmmB, m); cc.evex().vpackuswb(xmmA, xmmB, m); cc.evex().vpackuswb(ymmA, ymmB, m); cc.evex().vpackuswb(zmmA, zmmB, m); cc.evex().vpaddb(xmmA, xmmB, m); cc.evex().vpaddb(ymmA, ymmB, m); cc.evex().vpaddb(zmmA, zmmB, m); cc.evex().vpaddd(xmmA, xmmB, m); cc.evex().vpaddd(ymmA, ymmB, m); cc.evex().vpaddd(zmmA, zmmB, m); cc.evex().vpaddq(xmmA, xmmB, m); cc.evex().vpaddq(ymmA, ymmB, m); cc.evex().vpaddq(zmmA, zmmB, m); cc.evex().vpaddsb(xmmA, xmmB, m); cc.evex().vpaddsb(ymmA, ymmB, m); cc.evex().vpaddsb(zmmA, zmmB, m); cc.evex().vpaddsw(xmmA, xmmB, m); cc.evex().vpaddsw(ymmA, ymmB, m); cc.evex().vpaddsw(zmmA, zmmB, m); cc.evex().vpaddusb(xmmA, xmmB, m); cc.evex().vpaddusb(ymmA, ymmB, m); cc.evex().vpaddusb(zmmA, zmmB, m); cc.evex().vpaddusw(xmmA, xmmB, m); cc.evex().vpaddusw(ymmA, ymmB, m); cc.evex().vpaddusw(zmmA, zmmB, m); cc.evex().vpaddw(xmmA, xmmB, m); cc.evex().vpaddw(ymmA, ymmB, m); cc.evex().vpaddw(zmmA, zmmB, m); cc.evex().vpalignr(xmmA, xmmB, m, 0); cc.evex().vpalignr(ymmA, ymmB, m, 0); cc.evex().vpalignr(zmmA, zmmB, m, 0); cc.evex().vpandd(xmmA, xmmB, m); cc.evex().vpandd(ymmA, ymmB, m); cc.evex().vpandd(zmmA, zmmB, m); cc.evex().vpandnd(xmmA, xmmB, m); cc.evex().vpandnd(ymmA, ymmB, m); cc.evex().vpandnd(zmmA, zmmB, m); cc.evex().vpandnq(xmmA, xmmB, m); cc.evex().vpandnq(ymmA, ymmB, m); cc.evex().vpandnq(zmmA, zmmB, m); cc.evex().vpandq(xmmA, xmmB, m); cc.evex().vpandq(ymmA, ymmB, m); cc.evex().vpandq(zmmA, zmmB, m); cc.evex().vpavgb(xmmA, xmmB, m); cc.evex().vpavgb(ymmA, ymmB, m); cc.evex().vpavgb(zmmA, zmmB, m); cc.evex().vpavgw(xmmA, xmmB, m); cc.evex().vpavgw(ymmA, ymmB, m); cc.evex().vpavgw(zmmA, zmmB, m); cc.evex().vpblendmb(xmmA, xmmB, m); cc.evex().vpblendmb(ymmA, ymmB, m); cc.evex().vpblendmb(zmmA, zmmB, m); cc.evex().vpblendmd(xmmA, xmmB, m); cc.evex().vpblendmd(ymmA, ymmB, m); cc.evex().vpblendmd(zmmA, zmmB, m); cc.evex().vpblendmq(xmmA, xmmB, m); cc.evex().vpblendmq(ymmA, ymmB, m); cc.evex().vpblendmq(zmmA, zmmB, m); cc.evex().vpblendmw(xmmA, xmmB, m); cc.evex().vpblendmw(ymmA, ymmB, m); cc.evex().vpblendmw(zmmA, zmmB, m); cc.evex().vpbroadcastb(xmmA, m); cc.evex().vpbroadcastb(ymmA, m); cc.evex().vpbroadcastb(zmmA, m); cc.evex().vpbroadcastd(xmmA, m); cc.evex().vpbroadcastd(ymmA, m); cc.evex().vpbroadcastd(zmmA, m); cc.evex().vpbroadcastq(xmmA, m); cc.evex().vpbroadcastq(ymmA, m); cc.evex().vpbroadcastq(zmmA, m); cc.evex().vpbroadcastw(xmmA, m); cc.evex().vpbroadcastw(ymmA, m); cc.evex().vpbroadcastw(zmmA, m); cc.evex().vpcmpb(kA, xmmB, m, 0); cc.evex().vpcmpb(kA, ymmB, m, 0); cc.evex().vpcmpb(kA, zmmB, m, 0); cc.evex().vpcmpd(kA, xmmB, m, 0); cc.evex().vpcmpd(kA, ymmB, m, 0); cc.evex().vpcmpd(kA, zmmB, m, 0); cc.evex().vpcmpeqb(kA, xmmB, m); cc.evex().vpcmpeqb(kA, ymmB, m); cc.evex().vpcmpeqb(kA, zmmB, m); cc.evex().vpcmpeqd(kA, xmmB, m); cc.evex().vpcmpeqd(kA, ymmB, m); cc.evex().vpcmpeqd(kA, zmmB, m); cc.evex().vpcmpeqq(kA, xmmB, m); cc.evex().vpcmpeqq(kA, ymmB, m); cc.evex().vpcmpeqq(kA, zmmB, m); cc.evex().vpcmpeqw(kA, xmmB, m); cc.evex().vpcmpeqw(kA, ymmB, m); cc.evex().vpcmpeqw(kA, zmmB, m); cc.evex().vpcmpgtb(kA, xmmB, m); cc.evex().vpcmpgtb(kA, ymmB, m); cc.evex().vpcmpgtb(kA, zmmB, m); cc.evex().vpcmpgtd(kA, xmmB, m); cc.evex().vpcmpgtd(kA, ymmB, m); cc.evex().vpcmpgtd(kA, zmmB, m); cc.evex().vpcmpgtq(kA, xmmB, m); cc.evex().vpcmpgtq(kA, ymmB, m); cc.evex().vpcmpgtq(kA, zmmB, m); cc.evex().vpcmpgtw(kA, xmmB, m); cc.evex().vpcmpgtw(kA, ymmB, m); cc.evex().vpcmpgtw(kA, zmmB, m); cc.evex().vpcmpq(kA, xmmB, m, 0); cc.evex().vpcmpq(kA, ymmB, m, 0); cc.evex().vpcmpq(kA, zmmB, m, 0); cc.evex().vpcmpub(kA, xmmB, m, 0); cc.evex().vpcmpub(kA, ymmB, m, 0); cc.evex().vpcmpub(kA, zmmB, m, 0); cc.evex().vpcmpud(kA, xmmB, m, 0); cc.evex().vpcmpud(kA, ymmB, m, 0); cc.evex().vpcmpud(kA, zmmB, m, 0); cc.evex().vpcmpuq(kA, xmmB, m, 0); cc.evex().vpcmpuq(kA, ymmB, m, 0); cc.evex().vpcmpuq(kA, zmmB, m, 0); cc.evex().vpcmpuw(kA, xmmB, m, 0); cc.evex().vpcmpuw(kA, ymmB, m, 0); cc.evex().vpcmpuw(kA, zmmB, m, 0); cc.evex().vpcmpw(kA, xmmB, m, 0); cc.evex().vpcmpw(kA, ymmB, m, 0); cc.evex().vpcmpw(kA, zmmB, m, 0); cc.evex().vpcompressd(m, xmmB); cc.evex().vpcompressd(m, ymmB); cc.evex().vpcompressd(m, zmmB); cc.evex().vpcompressq(m, xmmB); cc.evex().vpcompressq(m, ymmB); cc.evex().vpcompressq(m, zmmB); cc.evex().vpconflictd(xmmA, m); cc.evex().vpconflictd(ymmA, m); cc.evex().vpconflictd(zmmA, m); cc.evex().vpconflictq(xmmA, m); cc.evex().vpconflictq(ymmA, m); cc.evex().vpconflictq(zmmA, m); cc.evex().vpermb(xmmA, xmmB, m); cc.evex().vpermb(ymmA, ymmB, m); cc.evex().vpermb(zmmA, zmmB, m); cc.evex().vpermd(ymmA, ymmB, m); cc.evex().vpermd(zmmA, zmmB, m); cc.evex().vpermi2b(xmmA, xmmB, m); cc.evex().vpermi2b(ymmA, ymmB, m); cc.evex().vpermi2b(zmmA, zmmB, m); cc.evex().vpermi2d(xmmA, xmmB, m); cc.evex().vpermi2d(ymmA, ymmB, m); cc.evex().vpermi2d(zmmA, zmmB, m); cc.evex().vpermi2pd(xmmA, xmmB, m); cc.evex().vpermi2pd(ymmA, ymmB, m); cc.evex().vpermi2pd(zmmA, zmmB, m); cc.evex().vpermi2ps(xmmA, xmmB, m); cc.evex().vpermi2ps(ymmA, ymmB, m); cc.evex().vpermi2ps(zmmA, zmmB, m); cc.evex().vpermi2q(xmmA, xmmB, m); cc.evex().vpermi2q(ymmA, ymmB, m); cc.evex().vpermi2q(zmmA, zmmB, m); cc.evex().vpermi2w(xmmA, xmmB, m); cc.evex().vpermi2w(ymmA, ymmB, m); cc.evex().vpermi2w(zmmA, zmmB, m); cc.evex().vpermilpd(xmmA, xmmB, m); cc.evex().vpermilpd(ymmA, ymmB, m); cc.evex().vpermilpd(zmmA, zmmB, m); cc.evex().vpermilpd(xmmA, m, 0); cc.evex().vpermilpd(ymmA, m, 0); cc.evex().vpermilpd(zmmA, m, 0); cc.evex().vpermilps(xmmA, xmmB, m); cc.evex().vpermilps(ymmA, ymmB, m); cc.evex().vpermilps(zmmA, zmmB, m); cc.evex().vpermilps(xmmA, m, 0); cc.evex().vpermilps(ymmA, m, 0); cc.evex().vpermilps(zmmA, m, 0); cc.evex().vpermq(ymmA, ymmB, m); cc.evex().vpermq(zmmA, zmmB, m); cc.evex().vpermq(ymmA, m, 0); cc.evex().vpermq(zmmA, m, 0); cc.evex().vpermt2b(xmmA, xmmB, m); cc.evex().vpermt2b(ymmA, ymmB, m); cc.evex().vpermt2b(zmmA, zmmB, m); cc.evex().vpermt2d(xmmA, xmmB, m); cc.evex().vpermt2d(ymmA, ymmB, m); cc.evex().vpermt2d(zmmA, zmmB, m); cc.evex().vpermt2pd(xmmA, xmmB, m); cc.evex().vpermt2pd(ymmA, ymmB, m); cc.evex().vpermt2pd(zmmA, zmmB, m); cc.evex().vpermt2ps(xmmA, xmmB, m); cc.evex().vpermt2ps(ymmA, ymmB, m); cc.evex().vpermt2ps(zmmA, zmmB, m); cc.evex().vpermt2q(xmmA, xmmB, m); cc.evex().vpermt2q(ymmA, ymmB, m); cc.evex().vpermt2q(zmmA, zmmB, m); cc.evex().vpermt2w(xmmA, xmmB, m); cc.evex().vpermt2w(ymmA, ymmB, m); cc.evex().vpermt2w(zmmA, zmmB, m); cc.evex().vpermw(xmmA, xmmB, m); cc.evex().vpermw(ymmA, ymmB, m); cc.evex().vpermw(zmmA, zmmB, m); cc.evex().vpexpandd(xmmA, m); cc.evex().vpexpandd(ymmA, m); cc.evex().vpexpandd(zmmA, m); cc.evex().vpexpandq(xmmA, m); cc.evex().vpexpandq(ymmA, m); cc.evex().vpexpandq(zmmA, m); cc.evex().vpextrb(m, xmmB, 0); cc.evex().vpextrd(m, xmmB, 0); if (cc.is64Bit()) cc.evex().vpextrq(m, xmmB, 0); cc.evex().vpextrw(m, xmmB, 0); cc.evex().k(kA).vpgatherdd(xmmA, vx_ptr); cc.evex().k(kA).vpgatherdd(ymmA, vy_ptr); cc.evex().k(kA).vpgatherdd(zmmA, vz_ptr); cc.evex().k(kA).vpgatherdq(xmmA, vx_ptr); cc.evex().k(kA).vpgatherdq(ymmA, vx_ptr); cc.evex().k(kA).vpgatherdq(zmmA, vy_ptr); cc.evex().k(kA).vpgatherqd(xmmA, vx_ptr); cc.evex().k(kA).vpgatherqd(xmmA, vy_ptr); cc.evex().k(kA).vpgatherqd(ymmA, vz_ptr); cc.evex().k(kA).vpgatherqq(xmmA, vx_ptr); cc.evex().k(kA).vpgatherqq(ymmA, vy_ptr); cc.evex().k(kA).vpgatherqq(zmmA, vz_ptr); cc.evex().vpinsrb(xmmA, xmmB, m, 0); cc.evex().vpinsrd(xmmA, xmmB, m, 0); if (cc.is64Bit()) cc.evex().vpinsrq(xmmA, xmmB, m, 0); cc.evex().vpinsrw(xmmA, xmmB, m, 0); cc.evex().vplzcntd(xmmA, m); cc.evex().vplzcntd(ymmA, m); cc.evex().vplzcntd(zmmA, m); cc.evex().vplzcntq(xmmA, m); cc.evex().vplzcntq(ymmA, m); cc.evex().vplzcntq(zmmA, m); cc.evex().vpmadd52huq(xmmA, xmmB, m); cc.evex().vpmadd52huq(ymmA, ymmB, m); cc.evex().vpmadd52huq(zmmA, zmmB, m); cc.evex().vpmadd52luq(xmmA, xmmB, m); cc.evex().vpmadd52luq(ymmA, ymmB, m); cc.evex().vpmadd52luq(zmmA, zmmB, m); cc.evex().vpmaddubsw(xmmA, xmmB, m); cc.evex().vpmaddubsw(ymmA, ymmB, m); cc.evex().vpmaddubsw(zmmA, zmmB, m); cc.evex().vpmaddwd(xmmA, xmmB, m); cc.evex().vpmaddwd(ymmA, ymmB, m); cc.evex().vpmaddwd(zmmA, zmmB, m); cc.evex().vpmaxsb(xmmA, xmmB, m); cc.evex().vpmaxsb(ymmA, ymmB, m); cc.evex().vpmaxsb(zmmA, zmmB, m); cc.evex().vpmaxsd(xmmA, xmmB, m); cc.evex().vpmaxsd(ymmA, ymmB, m); cc.evex().vpmaxsd(zmmA, zmmB, m); cc.evex().vpmaxsq(xmmA, xmmB, m); cc.evex().vpmaxsq(ymmA, ymmB, m); cc.evex().vpmaxsq(zmmA, zmmB, m); cc.evex().vpmaxsw(xmmA, xmmB, m); cc.evex().vpmaxsw(ymmA, ymmB, m); cc.evex().vpmaxsw(zmmA, zmmB, m); cc.evex().vpmaxub(xmmA, xmmB, m); cc.evex().vpmaxub(ymmA, ymmB, m); cc.evex().vpmaxub(zmmA, zmmB, m); cc.evex().vpmaxud(xmmA, xmmB, m); cc.evex().vpmaxud(ymmA, ymmB, m); cc.evex().vpmaxud(zmmA, zmmB, m); cc.evex().vpmaxuq(xmmA, xmmB, m); cc.evex().vpmaxuq(ymmA, ymmB, m); cc.evex().vpmaxuq(zmmA, zmmB, m); cc.evex().vpmaxuw(xmmA, xmmB, m); cc.evex().vpmaxuw(ymmA, ymmB, m); cc.evex().vpmaxuw(zmmA, zmmB, m); cc.evex().vpminsb(xmmA, xmmB, m); cc.evex().vpminsb(ymmA, ymmB, m); cc.evex().vpminsb(zmmA, zmmB, m); cc.evex().vpminsd(xmmA, xmmB, m); cc.evex().vpminsd(ymmA, ymmB, m); cc.evex().vpminsd(zmmA, zmmB, m); cc.evex().vpminsq(xmmA, xmmB, m); cc.evex().vpminsq(ymmA, ymmB, m); cc.evex().vpminsq(zmmA, zmmB, m); cc.evex().vpminsw(xmmA, xmmB, m); cc.evex().vpminsw(ymmA, ymmB, m); cc.evex().vpminsw(zmmA, zmmB, m); cc.evex().vpminub(xmmA, xmmB, m); cc.evex().vpminub(ymmA, ymmB, m); cc.evex().vpminub(zmmA, zmmB, m); cc.evex().vpminud(xmmA, xmmB, m); cc.evex().vpminud(ymmA, ymmB, m); cc.evex().vpminud(zmmA, zmmB, m); cc.evex().vpminuq(xmmA, xmmB, m); cc.evex().vpminuq(ymmA, ymmB, m); cc.evex().vpminuq(zmmA, zmmB, m); cc.evex().vpminuw(xmmA, xmmB, m); cc.evex().vpminuw(ymmA, ymmB, m); cc.evex().vpminuw(zmmA, zmmB, m); cc.evex().vpmovdb(m, xmmB); cc.evex().vpmovdb(m, ymmB); cc.evex().vpmovdb(m, zmmB); cc.evex().vpmovdw(m, xmmB); cc.evex().vpmovdw(m, ymmB); cc.evex().vpmovdw(m, zmmB); cc.evex().vpmovqb(m, xmmB); cc.evex().vpmovqb(m, ymmB); cc.evex().vpmovqb(m, zmmB); cc.evex().vpmovqd(m, xmmB); cc.evex().vpmovqd(m, ymmB); cc.evex().vpmovqd(m, zmmB); cc.evex().vpmovqw(m, xmmB); cc.evex().vpmovqw(m, ymmB); cc.evex().vpmovqw(m, zmmB); cc.evex().vpmovsdb(m, xmmB); cc.evex().vpmovsdb(m, ymmB); cc.evex().vpmovsdb(m, zmmB); cc.evex().vpmovsdw(m, xmmB); cc.evex().vpmovsdw(m, ymmB); cc.evex().vpmovsdw(m, zmmB); cc.evex().vpmovsqb(m, xmmB); cc.evex().vpmovsqb(m, ymmB); cc.evex().vpmovsqb(m, zmmB); cc.evex().vpmovsqd(m, xmmB); cc.evex().vpmovsqd(m, ymmB); cc.evex().vpmovsqd(m, zmmB); cc.evex().vpmovsqw(m, xmmB); cc.evex().vpmovsqw(m, ymmB); cc.evex().vpmovsqw(m, zmmB); cc.evex().vpmovswb(m, xmmB); cc.evex().vpmovswb(m, ymmB); cc.evex().vpmovswb(m, zmmB); cc.evex().vpmovsxbd(xmmA, m); cc.evex().vpmovsxbd(ymmA, m); cc.evex().vpmovsxbd(zmmA, m); cc.evex().vpmovsxbq(xmmA, m); cc.evex().vpmovsxbq(ymmA, m); cc.evex().vpmovsxbq(zmmA, m); cc.evex().vpmovsxbw(xmmA, m); cc.evex().vpmovsxbw(ymmA, m); cc.evex().vpmovsxbw(zmmA, m); cc.evex().vpmovsxdq(xmmA, m); cc.evex().vpmovsxdq(ymmA, m); cc.evex().vpmovsxdq(zmmA, m); cc.evex().vpmovsxwd(xmmA, m); cc.evex().vpmovsxwd(ymmA, m); cc.evex().vpmovsxwd(zmmA, m); cc.evex().vpmovsxwq(xmmA, m); cc.evex().vpmovsxwq(ymmA, m); cc.evex().vpmovsxwq(zmmA, m); cc.evex().vpmovusdb(m, xmmB); cc.evex().vpmovusdb(m, ymmB); cc.evex().vpmovusdb(m, zmmB); cc.evex().vpmovusdw(m, xmmB); cc.evex().vpmovusdw(m, ymmB); cc.evex().vpmovusdw(m, zmmB); cc.evex().vpmovusqb(m, xmmB); cc.evex().vpmovusqb(m, ymmB); cc.evex().vpmovusqb(m, zmmB); cc.evex().vpmovusqd(m, xmmB); cc.evex().vpmovusqd(m, ymmB); cc.evex().vpmovusqd(m, zmmB); cc.evex().vpmovusqw(m, xmmB); cc.evex().vpmovusqw(m, ymmB); cc.evex().vpmovusqw(m, zmmB); cc.evex().vpmovuswb(m, xmmB); cc.evex().vpmovuswb(m, ymmB); cc.evex().vpmovuswb(m, zmmB); cc.evex().vpmovwb(m, xmmB); cc.evex().vpmovwb(m, ymmB); cc.evex().vpmovwb(m, zmmB); cc.evex().vpmovzxbd(xmmA, m); cc.evex().vpmovzxbd(ymmA, m); cc.evex().vpmovzxbd(zmmA, m); cc.evex().vpmovzxbq(xmmA, m); cc.evex().vpmovzxbq(ymmA, m); cc.evex().vpmovzxbq(zmmA, m); cc.evex().vpmovzxbw(xmmA, m); cc.evex().vpmovzxbw(ymmA, m); cc.evex().vpmovzxbw(zmmA, m); cc.evex().vpmovzxdq(xmmA, m); cc.evex().vpmovzxdq(ymmA, m); cc.evex().vpmovzxdq(zmmA, m); cc.evex().vpmovzxwd(xmmA, m); cc.evex().vpmovzxwd(ymmA, m); cc.evex().vpmovzxwd(zmmA, m); cc.evex().vpmovzxwq(xmmA, m); cc.evex().vpmovzxwq(ymmA, m); cc.evex().vpmovzxwq(zmmA, m); cc.evex().vpmuldq(xmmA, xmmB, m); cc.evex().vpmuldq(ymmA, ymmB, m); cc.evex().vpmuldq(zmmA, zmmB, m); cc.evex().vpmulhrsw(xmmA, xmmB, m); cc.evex().vpmulhrsw(ymmA, ymmB, m); cc.evex().vpmulhrsw(zmmA, zmmB, m); cc.evex().vpmulhuw(xmmA, xmmB, m); cc.evex().vpmulhuw(ymmA, ymmB, m); cc.evex().vpmulhuw(zmmA, zmmB, m); cc.evex().vpmulhw(xmmA, xmmB, m); cc.evex().vpmulhw(ymmA, ymmB, m); cc.evex().vpmulhw(zmmA, zmmB, m); cc.evex().vpmulld(xmmA, xmmB, m); cc.evex().vpmulld(ymmA, ymmB, m); cc.evex().vpmulld(zmmA, zmmB, m); cc.evex().vpmullq(xmmA, xmmB, m); cc.evex().vpmullq(ymmA, ymmB, m); cc.evex().vpmullq(zmmA, zmmB, m); cc.evex().vpmullw(xmmA, xmmB, m); cc.evex().vpmullw(ymmA, ymmB, m); cc.evex().vpmullw(zmmA, zmmB, m); cc.evex().vpmultishiftqb(xmmA, xmmB, m); cc.evex().vpmultishiftqb(ymmA, ymmB, m); cc.evex().vpmultishiftqb(zmmA, zmmB, m); cc.evex().vpmuludq(xmmA, xmmB, m); cc.evex().vpmuludq(ymmA, ymmB, m); cc.evex().vpmuludq(zmmA, zmmB, m); cc.evex().vpopcntd(zmmA, m); cc.evex().vpopcntq(zmmA, m); cc.evex().vpord(xmmA, xmmB, m); cc.evex().vpord(ymmA, ymmB, m); cc.evex().vpord(zmmA, zmmB, m); cc.evex().vporq(xmmA, xmmB, m); cc.evex().vporq(ymmA, ymmB, m); cc.evex().vporq(zmmA, zmmB, m); cc.evex().vprold(xmmA, m, 0); cc.evex().vprold(ymmA, m, 0); cc.evex().vprold(zmmA, m, 0); cc.evex().vprolq(xmmA, m, 0); cc.evex().vprolq(ymmA, m, 0); cc.evex().vprolq(zmmA, m, 0); cc.evex().vprolvd(xmmA, xmmB, m); cc.evex().vprolvd(ymmA, ymmB, m); cc.evex().vprolvd(zmmA, zmmB, m); cc.evex().vprolvq(xmmA, xmmB, m); cc.evex().vprolvq(ymmA, ymmB, m); cc.evex().vprolvq(zmmA, zmmB, m); cc.evex().vprord(xmmA, m, 0); cc.evex().vprord(ymmA, m, 0); cc.evex().vprord(zmmA, m, 0); cc.evex().vprorq(xmmA, m, 0); cc.evex().vprorq(ymmA, m, 0); cc.evex().vprorq(zmmA, m, 0); cc.evex().vprorvd(xmmA, xmmB, m); cc.evex().vprorvd(ymmA, ymmB, m); cc.evex().vprorvd(zmmA, zmmB, m); cc.evex().vprorvq(xmmA, xmmB, m); cc.evex().vprorvq(ymmA, ymmB, m); cc.evex().vprorvq(zmmA, zmmB, m); cc.evex().vpsadbw(xmmA, xmmB, m); cc.evex().vpsadbw(ymmA, ymmB, m); cc.evex().vpsadbw(zmmA, zmmB, m); cc.evex().k(kA).vpscatterdd(vx_ptr, xmmB); cc.evex().k(kA).vpscatterdd(vy_ptr, ymmB); cc.evex().k(kA).vpscatterdd(vz_ptr, zmmB); cc.evex().k(kA).vpscatterdq(vx_ptr, xmmB); cc.evex().k(kA).vpscatterdq(vx_ptr, ymmB); cc.evex().k(kA).vpscatterdq(vy_ptr, zmmB); cc.evex().k(kA).vpscatterqd(vx_ptr, xmmB); cc.evex().k(kA).vpscatterqd(vy_ptr, xmmB); cc.evex().k(kA).vpscatterqd(vz_ptr, ymmB); cc.evex().k(kA).vpscatterqq(vx_ptr, xmmB); cc.evex().k(kA).vpscatterqq(vy_ptr, ymmB); cc.evex().k(kA).vpscatterqq(vz_ptr, zmmB); cc.evex().vpshufb(xmmA, xmmB, m); cc.evex().vpshufb(ymmA, ymmB, m); cc.evex().vpshufb(zmmA, zmmB, m); cc.evex().vpshufd(xmmA, m, 0); cc.evex().vpshufd(ymmA, m, 0); cc.evex().vpshufd(zmmA, m, 0); cc.evex().vpshufhw(xmmA, m, 0); cc.evex().vpshufhw(ymmA, m, 0); cc.evex().vpshufhw(zmmA, m, 0); cc.evex().vpshuflw(xmmA, m, 0); cc.evex().vpshuflw(ymmA, m, 0); cc.evex().vpshuflw(zmmA, m, 0); cc.evex().vpslld(xmmA, xmmB, m); cc.evex().vpslld(xmmA, m, 0); cc.evex().vpslld(ymmA, ymmB, m); cc.evex().vpslld(ymmA, m, 0); cc.evex().vpslld(zmmA, zmmB, m); cc.evex().vpslld(zmmA, m, 0); cc.evex().vpslldq(xmmA, m, 0); cc.evex().vpslldq(ymmA, m, 0); cc.evex().vpslldq(zmmA, m, 0); cc.evex().vpsllq(xmmA, xmmB, m); cc.evex().vpsllq(xmmA, m, 0); cc.evex().vpsllq(ymmA, ymmB, m); cc.evex().vpsllq(ymmA, m, 0); cc.evex().vpsllq(zmmA, zmmB, m); cc.evex().vpsllq(zmmA, m, 0); cc.evex().vpsllvd(xmmA, xmmB, m); cc.evex().vpsllvd(ymmA, ymmB, m); cc.evex().vpsllvd(zmmA, zmmB, m); cc.evex().vpsllvq(xmmA, xmmB, m); cc.evex().vpsllvq(ymmA, ymmB, m); cc.evex().vpsllvq(zmmA, zmmB, m); cc.evex().vpsllvw(xmmA, xmmB, m); cc.evex().vpsllvw(ymmA, ymmB, m); cc.evex().vpsllvw(zmmA, zmmB, m); cc.evex().vpsllw(xmmA, xmmB, m); cc.evex().vpsllw(xmmA, m, 0); cc.evex().vpsllw(ymmA, ymmB, m); cc.evex().vpsllw(ymmA, m, 0); cc.evex().vpsllw(zmmA, zmmB, m); cc.evex().vpsllw(zmmA, m, 0); cc.evex().vpsrad(xmmA, xmmB, m); cc.evex().vpsrad(xmmA, m, 0); cc.evex().vpsrad(ymmA, ymmB, m); cc.evex().vpsrad(ymmA, m, 0); cc.evex().vpsrad(zmmA, zmmB, m); cc.evex().vpsrad(zmmA, m, 0); cc.evex().vpsraq(xmmA, xmmB, m); cc.evex().vpsraq(xmmA, m, 0); cc.evex().vpsraq(ymmA, ymmB, m); cc.evex().vpsraq(ymmA, m, 0); cc.evex().vpsraq(zmmA, zmmB, m); cc.evex().vpsraq(zmmA, m, 0); cc.evex().vpsravd(xmmA, xmmB, m); cc.evex().vpsravd(ymmA, ymmB, m); cc.evex().vpsravd(zmmA, zmmB, m); cc.evex().vpsravq(xmmA, xmmB, m); cc.evex().vpsravq(ymmA, ymmB, m); cc.evex().vpsravq(zmmA, zmmB, m); cc.evex().vpsravw(xmmA, xmmB, m); cc.evex().vpsravw(ymmA, ymmB, m); cc.evex().vpsravw(zmmA, zmmB, m); cc.evex().vpsraw(xmmA, xmmB, m); cc.evex().vpsraw(xmmA, m, 0); cc.evex().vpsraw(ymmA, ymmB, m); cc.evex().vpsraw(ymmA, m, 0); cc.evex().vpsraw(zmmA, zmmB, m); cc.evex().vpsraw(zmmA, m, 0); cc.evex().vpsrld(xmmA, xmmB, m); cc.evex().vpsrld(xmmA, m, 0); cc.evex().vpsrld(ymmA, ymmB, m); cc.evex().vpsrld(ymmA, m, 0); cc.evex().vpsrld(zmmA, zmmB, m); cc.evex().vpsrld(zmmA, m, 0); cc.evex().vpsrldq(xmmA, m, 0); cc.evex().vpsrldq(ymmA, m, 0); cc.evex().vpsrldq(zmmA, m, 0); cc.evex().vpsrlq(xmmA, xmmB, m); cc.evex().vpsrlq(xmmA, m, 0); cc.evex().vpsrlq(ymmA, ymmB, m); cc.evex().vpsrlq(ymmA, m, 0); cc.evex().vpsrlq(zmmA, zmmB, m); cc.evex().vpsrlq(zmmA, m, 0); cc.evex().vpsrlvd(xmmA, xmmB, m); cc.evex().vpsrlvd(ymmA, ymmB, m); cc.evex().vpsrlvd(zmmA, zmmB, m); cc.evex().vpsrlvq(xmmA, xmmB, m); cc.evex().vpsrlvq(ymmA, ymmB, m); cc.evex().vpsrlvq(zmmA, zmmB, m); cc.evex().vpsrlvw(xmmA, xmmB, m); cc.evex().vpsrlvw(ymmA, ymmB, m); cc.evex().vpsrlvw(zmmA, zmmB, m); cc.evex().vpsrlw(xmmA, xmmB, m); cc.evex().vpsrlw(xmmA, m, 0); cc.evex().vpsrlw(ymmA, ymmB, m); cc.evex().vpsrlw(ymmA, m, 0); cc.evex().vpsrlw(zmmA, zmmB, m); cc.evex().vpsrlw(zmmA, m, 0); cc.evex().vpsubb(xmmA, xmmB, m); cc.evex().vpsubb(ymmA, ymmB, m); cc.evex().vpsubb(zmmA, zmmB, m); cc.evex().vpsubd(xmmA, xmmB, m); cc.evex().vpsubd(ymmA, ymmB, m); cc.evex().vpsubd(zmmA, zmmB, m); cc.evex().vpsubq(xmmA, xmmB, m); cc.evex().vpsubq(ymmA, ymmB, m); cc.evex().vpsubq(zmmA, zmmB, m); cc.evex().vpsubsb(xmmA, xmmB, m); cc.evex().vpsubsb(ymmA, ymmB, m); cc.evex().vpsubsb(zmmA, zmmB, m); cc.evex().vpsubsw(xmmA, xmmB, m); cc.evex().vpsubsw(ymmA, ymmB, m); cc.evex().vpsubsw(zmmA, zmmB, m); cc.evex().vpsubusb(xmmA, xmmB, m); cc.evex().vpsubusb(ymmA, ymmB, m); cc.evex().vpsubusb(zmmA, zmmB, m); cc.evex().vpsubusw(xmmA, xmmB, m); cc.evex().vpsubusw(ymmA, ymmB, m); cc.evex().vpsubusw(zmmA, zmmB, m); cc.evex().vpsubw(xmmA, xmmB, m); cc.evex().vpsubw(ymmA, ymmB, m); cc.evex().vpsubw(zmmA, zmmB, m); cc.evex().vpternlogd(xmmA, xmmB, m, 0); cc.evex().vpternlogd(ymmA, ymmB, m, 0); cc.evex().vpternlogd(zmmA, zmmB, m, 0); cc.evex().vpternlogq(xmmA, xmmB, m, 0); cc.evex().vpternlogq(ymmA, ymmB, m, 0); cc.evex().vpternlogq(zmmA, zmmB, m, 0); cc.evex().vptestmb(kA, xmmB, m); cc.evex().vptestmb(kA, ymmB, m); cc.evex().vptestmb(kA, zmmB, m); cc.evex().vptestmd(kA, xmmB, m); cc.evex().vptestmd(kA, ymmB, m); cc.evex().vptestmd(kA, zmmB, m); cc.evex().vptestmq(kA, xmmB, m); cc.evex().vptestmq(kA, ymmB, m); cc.evex().vptestmq(kA, zmmB, m); cc.evex().vptestmw(kA, xmmB, m); cc.evex().vptestmw(kA, ymmB, m); cc.evex().vptestmw(kA, zmmB, m); cc.evex().vptestnmb(kA, xmmB, m); cc.evex().vptestnmb(kA, ymmB, m); cc.evex().vptestnmb(kA, zmmB, m); cc.evex().vptestnmd(kA, xmmB, m); cc.evex().vptestnmd(kA, ymmB, m); cc.evex().vptestnmd(kA, zmmB, m); cc.evex().vptestnmq(kA, xmmB, m); cc.evex().vptestnmq(kA, ymmB, m); cc.evex().vptestnmq(kA, zmmB, m); cc.evex().vptestnmw(kA, xmmB, m); cc.evex().vptestnmw(kA, ymmB, m); cc.evex().vptestnmw(kA, zmmB, m); cc.evex().vpunpckhbw(xmmA, xmmB, m); cc.evex().vpunpckhbw(ymmA, ymmB, m); cc.evex().vpunpckhbw(zmmA, zmmB, m); cc.evex().vpunpckhdq(xmmA, xmmB, m); cc.evex().vpunpckhdq(ymmA, ymmB, m); cc.evex().vpunpckhdq(zmmA, zmmB, m); cc.evex().vpunpckhqdq(xmmA, xmmB, m); cc.evex().vpunpckhqdq(ymmA, ymmB, m); cc.evex().vpunpckhqdq(zmmA, zmmB, m); cc.evex().vpunpckhwd(xmmA, xmmB, m); cc.evex().vpunpckhwd(ymmA, ymmB, m); cc.evex().vpunpckhwd(zmmA, zmmB, m); cc.evex().vpunpcklbw(xmmA, xmmB, m); cc.evex().vpunpcklbw(ymmA, ymmB, m); cc.evex().vpunpcklbw(zmmA, zmmB, m); cc.evex().vpunpckldq(xmmA, xmmB, m); cc.evex().vpunpckldq(ymmA, ymmB, m); cc.evex().vpunpckldq(zmmA, zmmB, m); cc.evex().vpunpcklqdq(xmmA, xmmB, m); cc.evex().vpunpcklqdq(ymmA, ymmB, m); cc.evex().vpunpcklqdq(zmmA, zmmB, m); cc.evex().vpunpcklwd(xmmA, xmmB, m); cc.evex().vpunpcklwd(ymmA, ymmB, m); cc.evex().vpunpcklwd(zmmA, zmmB, m); cc.evex().vpxord(xmmA, xmmB, m); cc.evex().vpxord(ymmA, ymmB, m); cc.evex().vpxord(zmmA, zmmB, m); cc.evex().vpxorq(xmmA, xmmB, m); cc.evex().vpxorq(ymmA, ymmB, m); cc.evex().vpxorq(zmmA, zmmB, m); cc.evex().vrangepd(xmmA, xmmB, m, 0); cc.evex().vrangepd(ymmA, ymmB, m, 0); cc.evex().vrangepd(zmmA, zmmB, m, 0); cc.evex().vrangeps(xmmA, xmmB, m, 0); cc.evex().vrangeps(ymmA, ymmB, m, 0); cc.evex().vrangeps(zmmA, zmmB, m, 0); cc.evex().vrangesd(xmmA, xmmB, m, 0); cc.evex().vrangess(xmmA, xmmB, m, 0); cc.evex().vrcp14pd(xmmA, m); cc.evex().vrcp14pd(ymmA, m); cc.evex().vrcp14pd(zmmA, m); cc.evex().vrcp14ps(xmmA, m); cc.evex().vrcp14ps(ymmA, m); cc.evex().vrcp14ps(zmmA, m); cc.evex().vrcp14sd(xmmA, xmmB, m); cc.evex().vrcp14ss(xmmA, xmmB, m); cc.evex().vrcp28pd(zmmA, m); cc.evex().vrcp28ps(zmmA, m); cc.evex().vrcp28sd(xmmA, xmmB, m); cc.evex().vrcp28ss(xmmA, xmmB, m); cc.evex().vreducepd(xmmA, m, 0); cc.evex().vreducepd(ymmA, m, 0); cc.evex().vreducepd(zmmA, m, 0); cc.evex().vreduceps(xmmA, m, 0); cc.evex().vreduceps(ymmA, m, 0); cc.evex().vreduceps(zmmA, m, 0); cc.evex().vreducesd(xmmA, xmmB, m, 0); cc.evex().vreducess(xmmA, xmmB, m, 0); cc.evex().vrndscalepd(xmmA, m, 0); cc.evex().vrndscalepd(ymmA, m, 0); cc.evex().vrndscalepd(zmmA, m, 0); cc.evex().vrndscaleps(xmmA, m, 0); cc.evex().vrndscaleps(ymmA, m, 0); cc.evex().vrndscaleps(zmmA, m, 0); cc.evex().vrndscalesd(xmmA, xmmB, m, 0); cc.evex().vrndscaless(xmmA, xmmB, m, 0); cc.evex().vrsqrt14pd(xmmA, m); cc.evex().vrsqrt14pd(ymmA, m); cc.evex().vrsqrt14pd(zmmA, m); cc.evex().vrsqrt14ps(xmmA, m); cc.evex().vrsqrt14ps(ymmA, m); cc.evex().vrsqrt14ps(zmmA, m); cc.evex().vrsqrt14sd(xmmA, xmmB, m); cc.evex().vrsqrt14ss(xmmA, xmmB, m); cc.evex().vrsqrt28pd(zmmA, m); cc.evex().vrsqrt28ps(zmmA, m); cc.evex().vrsqrt28sd(xmmA, xmmB, m); cc.evex().vrsqrt28ss(xmmA, xmmB, m); cc.evex().vscalefpd(xmmA, xmmB, m); cc.evex().vscalefpd(ymmA, ymmB, m); cc.evex().vscalefpd(zmmA, zmmB, m); cc.evex().vscalefps(xmmA, xmmB, m); cc.evex().vscalefps(ymmA, ymmB, m); cc.evex().vscalefps(zmmA, zmmB, m); cc.evex().vscalefsd(xmmA, xmmB, m); cc.evex().vscalefss(xmmA, xmmB, m); cc.evex().k(kA).vscatterdpd(vx_ptr, xmmB); cc.evex().k(kA).vscatterdpd(vx_ptr, ymmB); cc.evex().k(kA).vscatterdpd(vy_ptr, zmmB); cc.evex().k(kA).vscatterdps(vx_ptr, xmmB); cc.evex().k(kA).vscatterdps(vy_ptr, ymmB); cc.evex().k(kA).vscatterdps(vz_ptr, zmmB); cc.evex().k(kA).vscatterpf0dpd(vy_ptr); cc.evex().k(kA).vscatterpf0dps(vz_ptr); cc.evex().k(kA).vscatterpf0qpd(vz_ptr); cc.evex().k(kA).vscatterpf0qps(vz_ptr); cc.evex().k(kA).vscatterpf1dpd(vy_ptr); cc.evex().k(kA).vscatterpf1dps(vz_ptr); cc.evex().k(kA).vscatterpf1qpd(vz_ptr); cc.evex().k(kA).vscatterpf1qps(vz_ptr); cc.evex().k(kA).vscatterqpd(vx_ptr, xmmB); cc.evex().k(kA).vscatterqpd(vy_ptr, ymmB); cc.evex().k(kA).vscatterqpd(vz_ptr, zmmB); cc.evex().k(kA).vscatterqps(vx_ptr, xmmB); cc.evex().k(kA).vscatterqps(vy_ptr, xmmB); cc.evex().k(kA).vscatterqps(vz_ptr, ymmB); cc.evex().vshuff32x4(ymmA, ymmB, m, 0); cc.evex().vshuff32x4(zmmA, zmmB, m, 0); cc.evex().vshuff64x2(ymmA, ymmB, m, 0); cc.evex().vshuff64x2(zmmA, zmmB, m, 0); cc.evex().vshufi32x4(ymmA, ymmB, m, 0); cc.evex().vshufi32x4(zmmA, zmmB, m, 0); cc.evex().vshufi64x2(ymmA, ymmB, m, 0); cc.evex().vshufi64x2(zmmA, zmmB, m, 0); cc.evex().vshufpd(xmmA, xmmB, m, 0); cc.evex().vshufpd(ymmA, ymmB, m, 0); cc.evex().vshufpd(zmmA, zmmB, m, 0); cc.evex().vshufps(xmmA, xmmB, m, 0); cc.evex().vshufps(ymmA, ymmB, m, 0); cc.evex().vshufps(zmmA, zmmB, m, 0); cc.evex().vsqrtpd(xmmA, m); cc.evex().vsqrtpd(ymmA, m); cc.evex().vsqrtpd(zmmA, m); cc.evex().vsqrtps(xmmA, m); cc.evex().vsqrtps(ymmA, m); cc.evex().vsqrtps(zmmA, m); cc.evex().vsqrtsd(xmmA, xmmB, m); cc.evex().vsqrtss(xmmA, xmmB, m); cc.evex().vsubpd(xmmA, xmmB, m); cc.evex().vsubpd(ymmA, ymmB, m); cc.evex().vsubpd(zmmA, zmmB, m); cc.evex().vsubps(xmmA, xmmB, m); cc.evex().vsubps(ymmA, ymmB, m); cc.evex().vsubps(zmmA, zmmB, m); cc.evex().vsubsd(xmmA, xmmB, m); cc.evex().vsubss(xmmA, xmmB, m); cc.evex().vucomisd(xmmA, m); cc.evex().vucomiss(xmmA, m); cc.evex().vunpckhpd(xmmA, xmmB, m); cc.evex().vunpckhpd(ymmA, ymmB, m); cc.evex().vunpckhpd(zmmA, zmmB, m); cc.evex().vunpckhps(xmmA, xmmB, m); cc.evex().vunpckhps(ymmA, ymmB, m); cc.evex().vunpckhps(zmmA, zmmB, m); cc.evex().vunpcklpd(xmmA, xmmB, m); cc.evex().vunpcklpd(ymmA, ymmB, m); cc.evex().vunpcklpd(zmmA, zmmB, m); cc.evex().vunpcklps(xmmA, xmmB, m); cc.evex().vunpcklps(ymmA, ymmB, m); cc.evex().vunpcklps(zmmA, zmmB, m); cc.evex().vxorpd(xmmA, xmmB, m); cc.evex().vxorpd(ymmA, ymmB, m); cc.evex().vxorpd(zmmA, zmmB, m); cc.evex().vxorps(xmmA, xmmB, m); cc.evex().vxorps(ymmA, ymmB, m); cc.evex().vxorps(zmmA, zmmB, m); } // Generates a long sequence of AVX512 instructions. template static void generateAvx512SequenceInternal( Emitter& cc, InstForm form, const x86::Gp& gp, const x86::KReg& kA, const x86::KReg& kB, const x86::KReg& kC, const x86::Vec& vecA, const x86::Vec& vecB, const x86::Vec& vecC, const x86::Vec& vecD) { if (form == InstForm::kReg) generateAvx512SequenceInternalRegOnly(cc, gp, kA, kB, kC, vecA, vecB, vecC, vecD); else generateAvx512SequenceInternalRegMem(cc, gp, kA, kB, kC, vecA, vecB, vecC, vecD); } static void generateAvx512Sequence(BaseEmitter& emitter, InstForm form, bool emitPrologEpilog) { using namespace asmjit::x86; if (emitter.isAssembler()) { Assembler& cc = *emitter.as(); if (emitPrologEpilog) { FuncDetail func; func.init(FuncSignature::build(), cc.environment()); FuncFrame frame; frame.init(func); frame.addDirtyRegs(eax, k1, k2, k3, zmm0, zmm1, zmm2, zmm3); frame.finalize(); cc.emitProlog(frame); generateAvx512SequenceInternal(cc, form, eax, k1, k2, k3, zmm0, zmm1, zmm2, zmm3); cc.emitEpilog(frame); } else { generateAvx512SequenceInternal(cc, form, eax, k1, k2, k3, zmm0, zmm1, zmm2, zmm3); } } #ifndef ASMJIT_NO_BUILDER else if (emitter.isBuilder()) { Builder& cc = *emitter.as(); if (emitPrologEpilog) { FuncDetail func; func.init(FuncSignature::build(), cc.environment()); FuncFrame frame; frame.init(func); frame.addDirtyRegs(eax, k1, k2, k3, zmm0, zmm1, zmm2, zmm3); frame.finalize(); cc.emitProlog(frame); generateAvx512SequenceInternal(cc, form, eax, k1, k2, k3, zmm0, zmm1, zmm2, zmm3); cc.emitEpilog(frame); } else { generateAvx512SequenceInternal(cc, form, eax, k1, k2, k3, zmm0, zmm1, zmm2, zmm3); } } #endif #ifndef ASMJIT_NO_COMPILER else if (emitter.isCompiler()) { Compiler& cc = *emitter.as(); Gp gp = cc.newGpz("gp"); Zmm vecA = cc.newZmm("vecA"); Zmm vecB = cc.newZmm("vecB"); Zmm vecC = cc.newZmm("vecC"); Zmm vecD = cc.newZmm("vecD"); KReg kA = cc.newKq("kA"); KReg kB = cc.newKq("kB"); KReg kC = cc.newKq("kC"); cc.addFunc(FuncSignature::build()); generateAvx512SequenceInternal(cc, form, gp, kA, kB, kC, vecA, vecB, vecC, vecD); cc.endFunc(); } #endif } template static void benchmarkX86Function(Arch arch, uint32_t numIterations, const char* description, const EmitterFn& emitterFn) noexcept { CodeHolder code; printf("%s:\n", description); uint32_t instCount = 0; #ifndef ASMJIT_NO_BUILDER instCount = asmjit_perf_utils::calculateInstructionCount(code, arch, [&](x86::Builder& cc) { emitterFn(cc, false); }); #endif asmjit_perf_utils::bench(code, arch, numIterations, "[raw]", instCount, [&](x86::Assembler& cc) { emitterFn(cc, false); }); asmjit_perf_utils::bench(code, arch, numIterations, "[validated]", instCount, [&](x86::Assembler& cc) { cc.addDiagnosticOptions(DiagnosticOptions::kValidateAssembler); emitterFn(cc, false); }); asmjit_perf_utils::bench(code, arch, numIterations, "[prolog/epilog]", instCount, [&](x86::Assembler& cc) { emitterFn(cc, true); }); #ifndef ASMJIT_NO_BUILDER asmjit_perf_utils::bench(code, arch, numIterations, "[no-asm]", instCount, [&](x86::Builder& cc) { emitterFn(cc, false); }); asmjit_perf_utils::bench(code, arch, numIterations, "[finalized]", instCount, [&](x86::Builder& cc) { emitterFn(cc, false); cc.finalize(); }); asmjit_perf_utils::bench(code, arch, numIterations, "[prolog/epilog]", instCount, [&](x86::Builder& cc) { emitterFn(cc, true); cc.finalize(); }); #endif #ifndef ASMJIT_NO_COMPILER asmjit_perf_utils::bench(code, arch, numIterations, "[no-asm]", instCount, [&](x86::Compiler& cc) { emitterFn(cc, true); }); asmjit_perf_utils::bench(code, arch, numIterations, "[finalized]", instCount, [&](x86::Compiler& cc) { emitterFn(cc, true); cc.finalize(); }); #endif printf("\n"); } void benchmarkX86Emitters(uint32_t numIterations, bool testX86, bool testX64) { uint32_t i = 0; uint32_t n = 0; Arch archs[2] {}; if (testX86) archs[n++] = Arch::kX86; if (testX64) archs[n++] = Arch::kX64; for (i = 0; i < n; i++) { static const char description[] = "GpSequence (Sequence of GP instructions - reg-only)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { generateGpSequence(emitter, InstForm::kReg, emitPrologEpilog); }); } for (i = 0; i < n; i++) { static const char description[] = "GpSequence (Sequence of GP instructions - reg/mem)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { generateGpSequence(emitter, InstForm::kMem, emitPrologEpilog); }); } for (i = 0; i < n; i++) { static const char description[] = "SseSequence (sequence of SSE+ instructions - reg-only)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { generateSseSequence(emitter, InstForm::kReg, emitPrologEpilog); }); } for (i = 0; i < n; i++) { static const char description[] = "SseSequence (sequence of SSE+ instructions - reg/mem)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { generateSseSequence(emitter, InstForm::kMem, emitPrologEpilog); }); } for (i = 0; i < n; i++) { static const char description[] = "AvxSequence (sequence of AVX+ instructions - reg-only)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { generateAvxSequence(emitter, InstForm::kReg, emitPrologEpilog); }); } for (i = 0; i < n; i++) { static const char description[] = "AvxSequence (sequence of AVX+ instructions - reg/mem)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { generateAvxSequence(emitter, InstForm::kMem, emitPrologEpilog); }); } for (i = 0; i < n; i++) { static const char description[] = "Avx512Sequence (sequence of AVX512+ instructions - reg-only)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { generateAvx512Sequence(emitter, InstForm::kReg, emitPrologEpilog); }); } for (i = 0; i < n; i++) { static const char description[] = "Avx512Sequence (sequence of AVX512+ instructions - reg/mem)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { generateAvx512Sequence(emitter, InstForm::kMem, emitPrologEpilog); }); } for (i = 0; i < n; i++) { static const char description[] = "SseAlphaBlend (alpha-blend function with labels and jumps)"; benchmarkX86Function(archs[i], numIterations, description, [](BaseEmitter& emitter, bool emitPrologEpilog) { asmtest::generateSseAlphaBlend(emitter, emitPrologEpilog); }); } } #endif // !ASMJIT_NO_X86