Lines Matching refs:tmp

180     __m128i tmp = _mm_srai_epi32(bl, 31);
181 tmp = _mm_and_si128(tmp, mask);
182 tmp = _mm_shuffle_epi32(tmp, _MM_SHUFFLE(2, 1, 0, 3));
184 return _mm_xor_si128(bl, tmp);
269 block tmp = vshrq_n_s8(b, 7);
270 tmp = vandq_s8(tmp, mask);
271 tmp = vextq_s8(tmp, tmp, 1); /* Rotate high byte to end */
273 return veorq_s8(tmp, b);
441 kp[idx] = _mm_castps_si128(_mm_shuffle_ps(_mm_castsi128_ps(tmp), _mm_castsi128_ps(x0), 68)); \
448 tmp = x3
478 __m128i x0, x1, x2, x3, tmp, *kp = (__m128i*)key;
480 tmp = x3 = _mm_loadu_si128((__m128i*)(userkey + 16));
557 __m128i tmp = _mm_load_si128((__m128i*)in);
558 tmp = _mm_xor_si128(tmp, sched[0]);
560 tmp = _mm_aesenc_si128(tmp, sched[j]);
561 tmp = _mm_aesenclast_si128(tmp, sched[j]);
562 _mm_store_si128((__m128i*)out, tmp);
568 __m128i tmp = _mm_load_si128((__m128i*)in);
569 tmp = _mm_xor_si128(tmp, sched[0]);
571 tmp = _mm_aesdec_si128(tmp, sched[j]);
572 tmp = _mm_aesdeclast_si128(tmp, sched[j]);
573 _mm_store_si128((__m128i*)out, tmp);
757 } tmp;
763 tmp.u32[0] = 0x01000000 + ((OCB_TAG_LEN * 8 % 128) << 1);
765 tmp.u32[0] = 0x00000001 + ((OCB_TAG_LEN * 8 % 128) << 25);
768 tmp.u32[0] = 0x01000000 + ((ctx->tag_len * 8 % 128) << 1);
770 tmp.u32[0] = 0x00000001 + ((ctx->tag_len * 8 % 128) << 25);
772 tmp.u32[1] = ((uint32_t*)nonce)[0];
773 tmp.u32[2] = ((uint32_t*)nonce)[1];
774 tmp.u32[3] = ((uint32_t*)nonce)[2];
775 idx = (unsigned)(tmp.u8[15] & 0x3f); /* Get low 6 bits of nonce */
776 tmp.u8[15] = tmp.u8[15] & 0xc0; /* Zero low 6 bits of nonce */
777 if (unequal_blocks(tmp.bl, ctx->cached_Top)) { /* Cached? */
778 ctx->cached_Top = tmp.bl; /* Update cache, KtopStr */
779 AES_encrypt(tmp.u8, (unsigned char*)&ctx->KtopStr, &ctx->encrypt_key);
794 } tmp;
856 tmp.bl = xor_block(ad_offset, ctx->L[0]);
857 ta[0] = xor_block(tmp.bl, adp[0]);
858 tmp.bl = xor_block(tmp.bl, ctx->L[1]);
859 ta[1] = xor_block(tmp.bl, adp[1]);
884 tmp.bl = zero_block();
885 memcpy(tmp.u8, adp + k, remaining);
886 tmp.u8[remaining] = (unsigned char)0x80u;
887 ta[k] = xor_block(ad_offset, tmp.bl);
924 } tmp;
1046 tmp.bl = zero_block();
1047 memcpy(tmp.u8, ptp + k, remaining);
1048 tmp.u8[remaining] = (unsigned char)0x80u;
1049 checksum = xor_block(checksum, tmp.bl);
1060 tmp.bl = xor_block(tmp.bl, ta[k]);
1061 memcpy(ctp + k, tmp.u8, remaining);
1135 } tmp;
1259 AES_encrypt((unsigned char*)&offset, tmp.u8, &ctx->encrypt_key);
1260 pad = tmp.bl;
1261 memcpy(tmp.u8, ctp + k, remaining);
1262 tmp.bl = xor_block(tmp.bl, pad);
1263 tmp.u8[remaining] = (unsigned char)0x80u;
1264 memcpy(ptp + k, tmp.u8, remaining);
1265 checksum = xor_block(checksum, tmp.bl);
1297 tmp.bl = xor_block(offset, checksum);
1298 AES_encrypt(tmp.u8, tmp.u8, &ctx->encrypt_key);
1299 tmp.bl = xor_block(tmp.bl, ctx->ad_checksum); /* Full tag */
1303 if (unequal_blocks(tmp.bl, *(block*)tag))
1312 if (constant_time_memcmp(tag, tmp.u8, len) != 0)
1315 if (constant_time_memcmp((char*)ct + ct_len, tmp.u8, len) != 0)