Lines Matching defs:ib

13990       /* 66 0F 71 /2 ib = PSRLW by immediate */
13997 /* 66 0F 71 /4 ib = PSRAW by immediate */
14004 /* 66 0F 71 /6 ib = PSLLW by immediate */
14014 /* 66 0F 72 /2 ib = PSRLD by immediate */
14021 /* 66 0F 72 /4 ib = PSRAD by immediate */
14028 /* 66 0F 72 /6 ib = PSLLD by immediate */
14038 /* 66 0F 73 /3 ib = PSRLDQ by immediate */
14052 /* 66 0F 73 /7 ib = PSLLDQ by immediate */
14067 /* 66 0F 73 /2 ib = PSRLQ by immediate */
14074 /* 66 0F 73 /6 ib = PSLLQ by immediate */
14481 /* 0F C6 /r ib = SHUFPS -- shuffle packed F32s */
14506 /* 66 0F C6 /r ib = SHUFPD -- shuffle packed F64s */
16796 UChar ib = getUChar(delta);
16797 rIS4 = (ib >> 4) & 0xF;
16805 UChar ib = getUChar(delta);
16806 rIS4 = (ib >> 4) & 0xF;
16835 UChar ib = getUChar(delta);
16836 rIS4 = (ib >> 4) & 0xF;
16844 UChar ib = getUChar(delta);
16845 rIS4 = (ib >> 4) & 0xF;
19263 /* 66 0F 3A 08 /r ib = ROUNDPS imm8, xmm2/m128, xmm1 */
19333 /* 66 0F 3A 09 /r ib = ROUNDPD imm8, xmm2/m128, xmm1 */
19388 /* 66 0F 3A 0A /r ib = ROUNDSS imm8, xmm2/m32, xmm1
19389 66 0F 3A 0B /r ib = ROUNDSD imm8, xmm2/m64, xmm1
19441 /* 66 0F 3A 0C /r ib = BLENDPS xmm1, xmm2/m128, imm8
19478 /* 66 0F 3A 0D /r ib = BLENDPD xmm1, xmm2/m128, imm8
19514 /* 66 0F 3A 0E /r ib = PBLENDW xmm1, xmm2/m128, imm8
19551 /* 66 0F 3A 14 /r ib = PEXTRB r/m16, xmm, imm8
19561 /* 66 0F 3A 15 /r ib = PEXTRW r/m16, xmm, imm8
19571 /* 66 no-REX.W 0F 3A 16 /r ib = PEXTRD reg/mem32, xmm2, imm8
19580 /* 66 REX.W 0F 3A 16 /r ib = PEXTRQ reg/mem64, xmm2, imm8
19592 /* 66 0F 3A 17 /r ib = EXTRACTPS reg/mem32, xmm2, imm8 Extract
19604 /* 66 0F 3A 20 /r ib = PINSRB xmm1, r32/m8, imm8
19635 /* 66 0F 3A 21 /r ib = INSERTPS imm8, xmm2/m32, xmm1
19674 /* 66 no-REX.W 0F 3A 22 /r ib = PINSRD xmm1, r/m32, imm8
19705 /* 66 REX.W 0F 3A 22 /r ib = PINSRQ xmm1, r/m64, imm8
19739 /* 66 0F 3A 40 /r ib = DPPS xmm1, xmm2/m128, imm8
19772 /* 66 0F 3A 41 /r ib = DPPD xmm1, xmm2/m128, imm8
19805 /* 66 0F 3A 42 /r ib = MPSADBW xmm1, xmm2/m128, imm8
19840 /* 66 0F 3A 44 /r ib = PCLMULQDQ xmm1, xmm2/m128, imm8
19881 /* 66 0F 3A 63 /r ib = PCMPISTRI imm8, xmm2/m128, xmm1
19882 66 0F 3A 62 /r ib = PCMPISTRM imm8, xmm2/m128, xmm1
19883 66 0F 3A 61 /r ib = PCMPESTRI imm8, xmm2/m128, xmm1
19884 66 0F 3A 60 /r ib = PCMPESTRM imm8, xmm2/m128, xmm1
19897 /* 66 0F 3A DF /r ib = AESKEYGENASSIST imm8, xmm2/m128, xmm1 */
25941 /* VPSHUFD imm8, xmm2/m128, xmm1 = VEX.128.66.0F.WIG 70 /r ib */
25946 /* VPSHUFD imm8, ymm2/m256, ymm1 = VEX.256.66.0F.WIG 70 /r ib */
25951 /* VPSHUFLW imm8, xmm2/m128, xmm1 = VEX.128.F2.0F.WIG 70 /r ib */
25957 /* VPSHUFLW imm8, ymm2/m256, ymm1 = VEX.256.F2.0F.WIG 70 /r ib */
25962 /* VPSHUFHW imm8, xmm2/m128, xmm1 = VEX.128.F3.0F.WIG 70 /r ib */
25968 /* VPSHUFHW imm8, ymm2/m256, ymm1 = VEX.256.F3.0F.WIG 70 /r ib */
25976 /* VPSRLW imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 71 /2 ib */
25977 /* VPSRAW imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 71 /4 ib */
25978 /* VPSLLW imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 71 /6 ib */
26002 /* VPSRLW imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 71 /2 ib */
26003 /* VPSRAW imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 71 /4 ib */
26004 /* VPSLLW imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 71 /6 ib */
26031 /* VPSRLD imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 72 /2 ib */
26032 /* VPSRAD imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 72 /4 ib */
26033 /* VPSLLD imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 72 /6 ib */
26057 /* VPSRLD imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 72 /2 ib */
26058 /* VPSRAD imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 72 /4 ib */
26059 /* VPSLLD imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 72 /6 ib */
26086 /* VPSRLDQ imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 73 /3 ib */
26087 /* VPSLLDQ imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 73 /7 ib */
26088 /* VPSRLQ imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 73 /2 ib */
26089 /* VPSLLQ imm8, xmm2, xmm1 = VEX.NDD.128.66.0F.WIG 73 /6 ib */
26127 /* VPSRLDQ imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 73 /3 ib */
26128 /* VPSLLDQ imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 73 /7 ib */
26129 /* VPSRLQ imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 73 /2 ib */
26130 /* VPSLLQ imm8, ymm2, ymm1 = VEX.NDD.256.66.0F.WIG 73 /6 ib */
26525 /* = VEX.NDS.LIG.F2.0F.WIG C2 /r ib */
26535 /* = VEX.NDS.LIG.F3.0F.WIG C2 /r ib */
26545 /* = VEX.NDS.128.66.0F.WIG C2 /r ib */
26555 /* = VEX.NDS.256.66.0F.WIG C2 /r ib */
26564 /* = VEX.NDS.128.0F.WIG C2 /r ib */
26574 /* = VEX.NDS.256.0F.WIG C2 /r ib */
26585 /* VPINSRW r32/m16, xmm2, xmm1 = VEX.NDS.128.66.0F.WIG C4 /r ib */
26619 /* VPEXTRW imm8, xmm1, reg32 = VEX.128.66.0F.W0 C5 /r ib */
26632 /* = VEX.NDS.128.0F.WIG C6 /r ib */
26662 /* = VEX.NDS.256.0F.WIG C6 /r ib */
26692 /* = VEX.NDS.128.66.0F.WIG C6 /r ib */
26722 /* = VEX.NDS.256.66.0F.WIG C6 /r ib */
30411 /* VPERMQ imm8, ymm2/m256, ymm1 = VEX.256.66.0F3A.W1 00 /r ib */
30412 /* VPERMPD imm8, ymm2/m256, ymm1 = VEX.256.66.0F3A.W1 01 /r ib */
30451 /* VPBLENDD imm8, xmm3/m128, xmm2, xmm1 = VEX.NDS.128.66.0F3A.W0 02 /r ib */
30491 /* VPBLENDD imm8, ymm3/m256, ymm2, ymm1 = VEX.NDS.256.66.0F3A.W0 02 /r ib */
30535 /* VPERMILPS imm8, ymm2/m256, ymm1 = VEX.256.66.0F3A.WIG 04 /r ib */
30565 /* VPERMILPS imm8, xmm2/m128, xmm1 = VEX.128.66.0F3A.WIG 04 /r ib */
30593 /* VPERMILPD imm8, xmm2/m128, xmm1 = VEX.128.66.0F3A.WIG 05 /r ib */
30626 /* VPERMILPD imm8, ymm2/m256, ymm1 = VEX.256.66.0F3A.WIG 05 /r ib */
30663 /* VPERM2F128 imm8, ymm3/m256, ymm2, ymm1 = VEX.NDS.66.0F3A.W0 06 /r ib */
30710 /* VROUNDPS = VEX.NDS.128.66.0F3A.WIG 08 ib */
30758 /* VROUNDPS = VEX.NDS.256.66.0F3A.WIG 08 ib */
30815 /* VROUNDPD = VEX.NDS.128.66.0F3A.WIG 09 ib */
30859 /* VROUNDPD = VEX.NDS.256.66.0F3A.WIG 09 ib */
30909 /* VROUNDSS = VEX.NDS.128.66.0F3A.WIG 0A ib */
30911 /* VROUNDSD = VEX.NDS.128.66.0F3A.WIG 0B ib */
30966 /* VBLENDPS = VEX.NDS.256.66.0F3A.WIG 0C /r ib */
30997 /* VBLENDPS = VEX.NDS.128.66.0F3A.WIG 0C /r ib */
31031 /* VBLENDPD = VEX.NDS.256.66.0F3A.WIG 0D /r ib */
31062 /* VBLENDPD = VEX.NDS.128.66.0F3A.WIG 0D /r ib */
31096 /* VPBLENDW = VEX.NDS.128.66.0F3A.WIG 0E /r ib */
31127 /* VPBLENDW = VEX.NDS.256.66.0F3A.WIG 0E /r ib */
31166 /* VPALIGNR = VEX.NDS.128.66.0F3A.WIG 0F /r ib */
31199 /* VPALIGNR = VEX.NDS.256.66.0F3A.WIG 0F /r ib */
31240 /* VPEXTRB imm8, xmm2, reg/m8 = VEX.128.66.0F3A.W0 14 /r ib */
31250 /* VPEXTRW = VEX.128.66.0F3A.W0 15 /r ib */
31260 /* VPEXTRD = VEX.128.66.0F3A.W0 16 /r ib */
31266 /* VPEXTRQ = VEX.128.66.0F3A.W1 16 /r ib */
31275 /* VEXTRACTPS imm8, xmm1, r32/m32 = VEX.128.66.0F3A.WIG 17 /r ib */
31285 /* VINSERTF128 = VEX.NDS.256.66.0F3A.W0 18 /r ib */
31289 UInt ib = 0;
31297 ib = getUChar(delta);
31299 ib, nameXMMReg(rE), nameYMMReg(rV), nameYMMReg(rG));
31304 ib = getUChar(delta);
31306 ib, dis_buf, nameYMMReg(rV), nameYMMReg(rG));
31311 putYMMRegLane128(rG, ib & 1, mkexpr(t128));
31320 /* VEXTRACTF128 = VEX.256.66.0F3A.W0 19 /r ib */
31324 UInt ib = 0;
31330 ib = getUChar(delta);
31331 assign(t128, getYMMRegLane128(rS, ib & 1));
31334 ib, nameXMMReg(rS), nameYMMReg(rD));
31338 ib = getUChar(delta);
31339 assign(t128, getYMMRegLane128(rS, ib & 1));
31342 ib, nameYMMReg(rS), dis_buf);
31351 /* VPINSRB r32/m8, xmm2, xmm1 = VEX.NDS.128.66.0F3A.W0 20 /r ib */
31387 = VEX.NDS.128.66.0F3A.WIG 21 /r ib */
31426 /* VPINSRD r32/m32, xmm2, xmm1 = VEX.NDS.128.66.0F3A.W0 22 /r ib */
31458 /* VPINSRQ r64/m64, xmm2, xmm1 = VEX.NDS.128.66.0F3A.W1 22 /r ib */
31495 /* VINSERTI128 = VEX.NDS.256.66.0F3A.W0 38 /r ib */
31499 UInt ib = 0;
31507 ib = getUChar(delta);
31509 ib, nameXMMReg(rE), nameYMMReg(rV), nameYMMReg(rG));
31514 ib = getUChar(delta);
31516 ib, dis_buf, nameYMMReg(rV), nameYMMReg(rG));
31521 putYMMRegLane128(rG, ib & 1, mkexpr(t128));
31530 /* VEXTRACTI128 = VEX.256.66.0F3A.W0 39 /r ib */
31534 UInt ib = 0;
31540 ib = getUChar(delta);
31541 assign(t128, getYMMRegLane128(rS, ib & 1));
31544 ib, nameXMMReg(rS), nameYMMReg(rD));
31548 ib = getUChar(delta);
31549 assign(t128, getYMMRegLane128(rS, ib & 1));
31552 ib, nameYMMReg(rS), dis_buf);
31561 /* VDPPS imm8, xmm3/m128,xmm2,xmm1 = VEX.NDS.128.66.0F3A.WIG 40 /r ib */
31591 /* VDPPS imm8, ymm3/m128,ymm2,ymm1 = VEX.NDS.256.66.0F3A.WIG 40 /r ib */
31629 /* VDPPD imm8, xmm3/m128,xmm2,xmm1 = VEX.NDS.128.66.0F3A.WIG 41 /r ib */
31663 /* VMPSADBW = VEX.NDS.128.66.0F3A.WIG 42 /r ib */
31698 /* VMPSADBW = VEX.NDS.256.66.0F3A.WIG 42 /r ib */
31741 /* VPCLMULQDQ = VEX.NDS.128.66.0F3A.WIG 44 /r ib */
31742 /* 66 0F 3A 44 /r ib = PCLMULQDQ xmm1, xmm2/m128, imm8
31780 /* VPERM2I128 imm8, ymm3/m256, ymm2, ymm1 = VEX.NDS.66.0F3A.W0 46 /r ib */
31892 /* VEX.128.66.0F3A.WIG 63 /r ib = VPCMPISTRI imm8, xmm2/m128, xmm1
31893 VEX.128.66.0F3A.WIG 62 /r ib = VPCMPISTRM imm8, xmm2/m128, xmm1
31894 VEX.128.66.0F3A.WIG 61 /r ib = VPCMPESTRI imm8, xmm2/m128, xmm1
31895 VEX.128.66.0F3A.WIG 60 /r ib = VPCMPESTRM imm8, xmm2/m128, xmm1