Lines Matching refs:D3

86 my ($A0,$A1,$A2,$A3,$B0,$B1,$B2,$B3,$C0,$C1,$C2,$C3,$D0,$D1,$D2,$D3)=map("%xmm$_",(0..15));
87 my ($T0,$T1,$T2,$T3)=($A3,$B3,$C3,$D3);
225 $code.="movdqa $ctr0_store, $D3
226 paddd .sse_inc(%rip), $D3
227 movdqa $D3, $D2
236 movdqa $D3, $ctr3_store\n" if ($n eq 4);
244 paddd $ctr3_store, $D3\n" if ($n eq 4);
264 movdqu 3*16 + $offset($inp), $D3
268 pxor $D, $D3
272 movdqu $D3, 3*16 + $offset($oup)\n";
301 pxor $A3, $D3
305 pshufb $C0, $D3
310 paddd $D3, $C3
340 palignr \$$s3, $D3, $D3
529 &xor_stream_using_temp($A3, $B3, $C3, $D3, "0*16", $D0);
646 &chacha_qr($A3,$B3,$C3,$D3,$C1,"store_left_load");
653 &chacha_qr($A3,$B3,$C3,$D3,$C1,"store_right_load"); $code.="
666 &xor_stream_using_temp($A3, $B3, $C3, $D3, "0*16", $D0);
883 movdqa $D0, $D3
895 movdqa $D3, $ctr3_store
959 &xor_stream_using_temp($A3,$B3,$C3,$D3,0*16,$D2); $code.="
1214 ($A0,$A1,$A2,$A3,$B0,$B1,$B2,$B3,$C0,$C1,$C2,$C3,$D0,$D1,$D2,$D3)=map("%ymm$_",(0..15));
1216 ($T0,$T1,$T2,$T3)=($A3,$B3,$C3,$D3);
1308 vpaddd $ctr0_store, $D0, $D3
1309 vpaddd $D3, $D0, $D2
1312 vmovdqa $D3, $ctr3_store
1325 vpaddd $ctr3_store, $D3, $D3
1423 vpxor $A3, $D3, $D3
1427 vpshufb $C0, $D3, $D3
1432 vpaddd $D3, $C3, $C3
1458 vpalignr \$$s3, $D3, $D3, $D3
1579 &xor_stream_avx2($A3, $B3, $C3, $D3, 0*32, $A0); $code.="
1765 &xor_stream_avx2($A3, $B3, $C3, $D3, 0*32, $A0); $code.="
1961 vmovdqa $D0, $D3
1962 vpaddd .avx2_inc(%rip), $D3, $D2
1968 vmovdqa $D3, $ctr3_store
1976 vperm2i128 \$0x13, $C3, $D3, $C3
1977 vperm2i128 \$0x02, $A3, $B3, $D3
1979 vpand .clamp(%rip), $D3, $D3
1980 vmovdqa $D3, $r_store
2068 &xor_stream_avx2($A3, $B3, $C3, $D3, 0*32, $A0); $code.="
2213 &xor_stream_avx2($A3, $B3, $C3, $D3, 0*32, $A0); $code.="