/external/openssl/crypto/modes/asm/ |
H A D | ghash-x86.pl | 253 &mov ($inp,&wparam(0)); # load Xi 256 &mov ($Zhh,&DWP(0,$inp)); # load Xi[16] 263 &mov (&DWP(0,"esp"),$Zhh); # copy Xi[16] on stack 286 &mov ($Zll,&wparam(0)); # load Xi 293 &mov ($Zhh,&DWP(0,$Zll)); # load Xi[16] 324 &mov ($inp,&wparam(0)); # load Xi 414 &mov ($inp,&wparam(0)); # load Xi 426 &mov ($inp,&wparam(0)); # load Xi 438 &mov ($Zhh,&wparam(0)); # load Xi 452 &mov ($Zll,&DWP(12,$Zhh)); # load Xi[1 [all...] |
H A D | ghash-x86_64.pl | 63 $Xi="%rdi"; 173 movzb 15($Xi),$Zlo 176 &loop ($Xi); 178 mov $Zlo,8($Xi) 179 mov $Zhi,($Xi) 242 &mov ($Zlo,"8($Xi)"); 243 &mov ($Zhi,"0($Xi)"); 253 &mov ("($Xi)",$Zhi); 254 &mov ("8($Xi)","%rdx"); 289 &mov ($dat,"$j($Xi)") i [all...] |
H A D | ghash-s390x.pl | 58 $Xi="%r2"; # argument block 96 la %r1,0($Xi) # H lies right after Xi in gcm128_context 108 aghi $Xi,-1 113 lg $Zlo,8+1($Xi) # Xi 134 la %r1,0($Xi) # H lies right after Xi in gcm128_context 147 aghi $Xi,-1 152 lg $Zlo,8+1($Xi) # X [all...] |
H A D | ghash-armv4.pl | 63 $Xi="r0"; # argument block 91 str $_,[$Xi,#$i] 93 str $_,[$Xi,#$i] 96 strb $_,[$Xi,#$i+3] 98 strb $Tlh,[$Xi,#$i+2] 100 strb $Thl,[$Xi,#$i+1] 101 strb $Thh,[$Xi,#$i] 144 ldrb $nhi,[$Xi,#15] 162 ldrb $nhi,[$Xi,#14] 195 ldrplb $Tll,[$Xi, [all...] |
H A D | ghash-sparcv9.pl | 64 $Xi="%i0"; # input argument block 86 ldub [$Xi+15],$xi0 87 ldub [$Xi+14],$xi1 136 ldub [$Xi+$cnt],$xi1 188 stx $Zlo,[$Xi+8] 190 stx $Zhi,[$Xi] 208 stx $Zlo,[$Xi+8] 210 stx $Zhi,[$Xi] 226 ldub [$Xi+15],$nlo 238 ldub [$Xi [all...] |
H A D | ghash-parisc.pl | 53 $Xi="%r26"; # argument block 125 ldb 15($Xi),$nlo 137 ldb 14($Xi),$nlo 166 ldbx $cnt($Xi),$nlo 206 std $Zll,8($Xi) 207 std $Zhh,0($Xi) 216 ldb 15($Xi),$nlo 229 ldb 14($Xi),$nlo 260 ldbx $cnt($Xi),$nlo 315 stw $Zll,12($Xi) [all...] |
H A D | ghash-alpha.pl | 34 $Xi="a0"; # $16, input argument block 266 ldq $Xlo,8($Xi) 267 ldq $Xhi,0($Xi) 314 stq $Xlo,8($Xi) 315 stq $Xhi,0($Xi) 341 ldq $Xhi,0($Xi) 342 ldq $Xlo,8($Xi) 427 stq $Xlo,8($Xi) 428 stq $Xhi,0($Xi)
|
H A D | ghash-ia64.pl | 80 { .mmi; (p16) ld1 xi[0]=[Xi],-1 96 Xi=r24; Htbl=r25; 109 $ADDP Xi=15,in0 // &Xi[15] 120 { .mib; ld1 xi[2]=[Xi],-1 // Xi[15] 123 { .mmi; ld1 xi[1]=[Xi],-1 // Xi[14] 141 { .mmi; add Hlo=9,Xi;; // ;; is here to prevent 142 add Hhi=1,Xi };; // pipelin [all...] |
H A D | ghash-armv4.s | 317 vld1.64 d29,[r0,:64]!@ load Xi 340 vld1.64 d21,[r0,:64]! @ load Xi 360 veor q14,q10 @ inp^=Xi 365 vmull.p8 q9,d1,d4 @ H.lo�Xi[i] 366 vmull.p8 q8,d0,d4 @ H.hi�Xi[i] 401 vst1.64 d21,[r0,:64]! @ write out Xi
|
/external/openssl/crypto/sha/asm/ |
H A D | sha1-sparcv9.pl | 32 $Xi="%g4"; 54 my $xi=($i&1)?@X[($i/2)%8]:$Xi; 72 " srlx @X[(($i+1)/2)%8],32,$Xi\n"; 91 sllx @X[($j+6)%8],32,$Xi ! Xupdate($i) 96 or $tmp1,$Xi,$Xi 98 xor $Xi,@X[$j%8],@X[$j%8] 99 srlx @X[$j%8],31,$Xi 101 and $Xi,$rot1m,$Xi [all...] |
H A D | sha1-586.pl | 415 my $Xi=4; # 4xSIMD Xupdate round, start pre-seeded 416 my @X=map("xmm$_",(4..7,0..3)); # pre-seeded for $Xi=4 525 sub Xupdate_ssse3_16_31() # recall that $Xi starts wtih 4 539 &movdqa (&QWP(64+16*(($Xi-4)%3),"esp"),@X[-4&7]);# save X[] to backtrace buffer 558 &movdqa (&QWP(0+16*(($Xi-1)&3),"esp"),@X[3]); # X[]+K xfer to IALU 587 &movdqa (@X[2],&QWP(64+16*(($Xi-6)%3),"esp")) if ($Xi>5); # restore X[] from backtrace buffer 595 &movdqa (@X[4],&QWP(112-16+16*(($Xi)/5),"esp")); # K_XX_XX 600 &movdqa (@X[1],@X[-2&7]) if ($Xi<7); 606 $Xi [all...] |
H A D | sha1-armv4-large.pl | 71 $Xi="r14"; 77 ldr $t0,[$Xi,#15*4] 78 ldr $t1,[$Xi,#13*4] 79 ldr $t2,[$Xi,#7*4] 81 ldr $t3,[$Xi,#2*4] 88 str $t0,[$Xi,#-4]! 121 str $t0,[$Xi,#-4]! 167 mov $Xi,sp 178 teq $Xi,sp 198 teq $Xi,s [all...] |
H A D | sha1-thumb.pl | 40 $Xi="r12"; 133 mov $Xi,sp 170 mov $t0,$Xi 175 mov $Xi,$t1 181 cmp $Xi,$t0 187 mov $Xi,$t1 193 cmp $Xi,$t0 201 mov $Xi,$t1 207 cmp $Xi,$t0 211 mov $Xi,s [all...] |
H A D | sha1-x86_64.pl | 294 my $Xi=4; 369 sub Xupdate_ssse3_16_31() # recall that $Xi starts wtih 4 402 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 438 &movdqa (@Tx[2],eval(16*(($Xi)/5))."($K_XX_XX)"); # K_XX_XX 446 $Xi++; push(@X,shift(@X)); # "rotate" X[] 456 &movdqa (@Tx[0],@X[-1&7]) if ($Xi==8); 467 if ($Xi%5) { 470 &movdqa (@Tx[2],eval(16*($Xi/5))."($K_XX_XX)"); 483 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 503 &movdqa (@Tx[1],@X[0]) if ($Xi<1 [all...] |
H A D | sha1-sparcv9a.pl | 58 $Xi="%o7"; 152 ld [$Xfer+`4*($i%16)`],$Xi 160 add $Xi,$e,$e 171 ld [$Xfer+`4*($i%16)`],$Xi 180 add $Xi,$e,$e 204 ld [$Xfer+`4*($i%16)`],$Xi 217 add $Xi,$e,$e 222 ld [$Xfer+`4*($i%16)`],$Xi 237 add $Xi,$e,$e 242 ld [$Xfer+`4*($i%16)`],$Xi [all...] |
/external/openssl/crypto/modes/ |
H A D | gcm128.c | 145 static void gcm_gmult_8bit(u64 Xi[2], const u128 Htable[256]) argument 148 const u8 *xi = (const u8 *)Xi+15; 221 if ((u8 *)Xi==xi) break; 236 Xi[0] = BSWAP8(Z.hi); 237 Xi[1] = BSWAP8(Z.lo); 239 u8 *p = (u8 *)Xi; 248 Xi[0] = Z.hi; 249 Xi[1] = Z.lo; 252 #define GCM_MUL(ctx,Xi) gcm_gmult_8bit(ctx->Xi 335 gcm_gmult_4bit(u64 Xi[2], const u128 Htable[16]) argument 406 gcm_ghash_4bit(u64 Xi[2],const u128 Htable[16], const u8 *inp,size_t len) argument 585 gcm_gmult_1bit(u64 Xi[2],const u64 H[2]) argument [all...] |
H A D | modes_lcl.h | 105 Xi,H; member in struct:gcm128_context 106 /* Relative position of Xi, H and pre-computed Htable is used 112 void (*gmult)(u64 Xi[2],const u128 Htable[16]); 113 void (*ghash)(u64 Xi[2],const u128 Htable[16],const u8 *inp,size_t len);
|
/external/openssl/crypto/aes/asm/ |
H A D | aesni-sha1-x86_64.pl | 109 my $Xi=4; 251 sub Xupdate_ssse3_16_31() # recall that $Xi starts wtih 4 284 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 320 &movdqa (@Tx[2],eval(16*(($Xi)/5))."($K_XX_XX)"); # K_XX_XX 328 $Xi++; push(@X,shift(@X)); # "rotate" X[] 338 &movdqa (@Tx[0],@X[-1&7]) if ($Xi==8); 349 if ($Xi%5) { 352 &movdqa (@Tx[2],eval(16*($Xi/5))."($K_XX_XX)"); 365 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 385 &movdqa (@Tx[1],@X[0]) if ($Xi<1 [all...] |