/arch/ia64/sn/kernel/sn2/ |
H A D | cache.c | 16 * @bytes: number of bytes to flush 20 * @flush_addr to @flush_addr + @bytes are flushed 24 sn_flush_all_caches(long flush_addr, long bytes) argument 32 flush_icache_range(addr, addr + bytes); 38 flush_icache_range(addr, addr + bytes);
|
/arch/sparc/include/asm/ |
H A D | xor_32.h | 24 sparc_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) argument 26 int lines = bytes / (sizeof (long)) / 8; 61 sparc_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 64 int lines = bytes / (sizeof (long)) / 8; 112 sparc_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 115 int lines = bytes / (sizeof (long)) / 8; 176 sparc_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 179 int lines = bytes / (sizeof (long)) / 8;
|
/arch/hexagon/mm/ |
H A D | copy_user_template.S | 21 * W: width in bytes 32 p0 = cmp.gtu(bytes,#0) 38 p1 = cmp.gtu(bytes,#15) 45 loopcount = lsr(bytes,#3) 59 bytes -= asl(loopcount,#3) 71 p1 = cmp.gtu(bytes,#7) 76 loopcount = lsr(bytes,#2) 89 bytes -= asl(loopcount,#2) 97 p1 = cmp.gtu(bytes,#3) 102 loopcount = lsr(bytes,# [all...] |
H A D | copy_to_user.S | 43 #define bytes r2 define
|
/arch/arm/plat-omap/include/plat/ |
H A D | iommu.h | 157 #define iopgsz_max(bytes) \ 158 (((bytes) >= SZ_16M) ? SZ_16M : \ 159 ((bytes) >= SZ_1M) ? SZ_1M : \ 160 ((bytes) >= SZ_64K) ? SZ_64K : \ 161 ((bytes) >= SZ_4K) ? SZ_4K : 0) 163 #define bytes_to_iopgsz(bytes) \ 164 (((bytes) == SZ_16M) ? MMU_CAM_PGSZ_16M : \ 165 ((bytes) == SZ_1M) ? MMU_CAM_PGSZ_1M : \ 166 ((bytes) == SZ_64K) ? MMU_CAM_PGSZ_64K : \ 167 ((bytes) [all...] |
/arch/x86/oprofile/ |
H A D | backtrace.c | 47 unsigned long bytes; local 49 bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead)); 50 if (bytes != sizeof(bufhead)) 93 unsigned long bytes; local 95 bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead)); 96 if (bytes != sizeof(bufhead))
|
/arch/arm/include/asm/ |
H A D | xor.h | 47 xor_arm4regs_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) argument 49 unsigned int lines = bytes / sizeof(unsigned long) / 4; 67 xor_arm4regs_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 70 unsigned int lines = bytes / sizeof(unsigned long) / 4; 89 xor_arm4regs_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 92 unsigned int lines = bytes / sizeof(unsigned long) / 2; 108 xor_arm4regs_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 111 unsigned int lines = bytes / sizeof(unsigned long) / 2;
|
/arch/alpha/lib/ |
H A D | ev6-memcpy.S | 8 * - uses bcmpge to compare 8 bytes in parallel 67 ldq $1, 0($17) # L : get 8 bytes 84 wh64 ($7) # L1 : memory subsystem hint: 64 bytes at 86 ldq $6, 0($17) # L0 : bytes 0..7 90 ldq $4, 8($17) # L : bytes 8..15 91 ldq $5, 16($17) # L : bytes 16..23 95 ldq $3, 24($17) # L : bytes 24..31 100 addq $17, 32, $17 # E : src += 32 bytes 101 stq $6, 0($16) # L : bytes 0..7 105 stq $4, 8($16) # L : bytes [all...] |
H A D | ev6-clear_user.S | 8 * right "bytes left to zero" value (and that it is updated only _after_ 22 * bytes left to copy in $0 69 and $1, 7, $2 # .. .. E .. : number of misaligned bytes in tail 76 * Head is not aligned. Write (8 - $4) bytes to head of destination 86 addq $0, $4, $0 # .. E .. .. : bytes left -= 8 - misalignment 94 * $2 is number of trailing bytes (0..7) ($2 never used...) 128 * $0 is the number of bytes left 183 # zero to 16 quadwords left to store, plus any trailing bytes 189 beq $1, $trailbytes # U .. .. .. : U L U L : Only 0..7 bytes to go 202 # We have an unknown number of bytes lef [all...] |
/arch/s390/crypto/ |
H A D | ghash_s390.c | 25 u32 bytes; member in struct:ghash_desc_ctx 62 if (dctx->bytes) { 63 u8 *pos = buf + (GHASH_BLOCK_SIZE - dctx->bytes); 65 n = min(srclen, dctx->bytes); 66 dctx->bytes -= n; 72 if (!dctx->bytes) { 88 dctx->bytes = GHASH_BLOCK_SIZE - srclen; 100 if (dctx->bytes) { 101 u8 *pos = buf + (GHASH_BLOCK_SIZE - dctx->bytes); 103 memset(pos, 0, dctx->bytes); [all...] |
/arch/x86/include/asm/ |
H A D | xor_32.h | 32 xor_pII_mmx_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) argument 34 unsigned long lines = bytes >> 7; 75 xor_pII_mmx_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 78 unsigned long lines = bytes >> 7; 124 xor_pII_mmx_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 127 unsigned long lines = bytes >> 7; 179 xor_pII_mmx_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 182 unsigned long lines = bytes >> 7; 259 xor_p5_mmx_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) argument 261 unsigned long lines = bytes >> 306 xor_p5_mmx_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, unsigned long *p3) argument 363 xor_p5_mmx_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, unsigned long *p3, unsigned long *p4) argument 429 xor_p5_mmx_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, unsigned long *p3, unsigned long *p4, unsigned long *p5) argument 587 xor_sse_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) argument 640 xor_sse_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, unsigned long *p3) argument 701 xor_sse_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, unsigned long *p3, unsigned long *p4) argument 769 xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, unsigned long *p3, unsigned long *p4, unsigned long *p5) argument [all...] |
H A D | insn.h | 29 insn_byte_t bytes[4]; member in union:insn_field::__anon3008 39 * prefixes.bytes[3]: last prefix 44 * opcode.bytes[0]: opcode1 45 * opcode.bytes[1]: opcode2 46 * opcode.bytes[2]: opcode3 143 if (insn->vex_prefix.nbytes == 2) /* 2 bytes VEX */ 146 return X86_VEX3_M(insn->vex_prefix.bytes[1]); 151 if (insn->vex_prefix.nbytes == 2) /* 2 bytes VEX */ 152 return X86_VEX_P(insn->vex_prefix.bytes[1]); 154 return X86_VEX_P(insn->vex_prefix.bytes[ [all...] |
H A D | xor_64.h | 91 xor_sse_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) argument 93 unsigned int lines = bytes >> 8; 142 xor_sse_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 145 unsigned int lines = bytes >> 8; 201 xor_sse_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 204 unsigned int lines = bytes >> 8; 268 xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, argument 271 unsigned int lines = bytes >> 8;
|
H A D | kvm_emulate.h | 72 * 3. Valid access sizes are 1, 2, 4 and 8 bytes. On x86/32 systems only 89 * read_std: Read bytes of standard (non-emulated/special) memory. 93 * @bytes: [IN ] Number of bytes to read from memory. 97 unsigned int bytes, 101 * write_std: Write bytes of standard (non-emulated/special) memory. 105 * @bytes: [IN ] Number of bytes to write to memory. 108 unsigned long addr, void *val, unsigned int bytes, 111 * fetch: Read bytes o 204 unsigned int bytes; member in struct:operand [all...] |
/arch/s390/include/asm/ |
H A D | bitops.h | 412 unsigned long bytes = 0; local 434 : "+&a" (bytes), "+&d" (size) 437 return bytes; 449 unsigned long bytes = 0; local 471 : "+&a" (bytes), "+&a" (size) 474 return bytes; 616 unsigned long bytes, bits; local 620 bytes = __ffz_word_loop(addr, size); 621 bits = __ffz_word(bytes*8, __load_ulong_be(addr, bytes)); 637 unsigned long bytes, bits; local 751 unsigned long bytes, bits; local 793 unsigned long bytes, bits; local [all...] |
/arch/x86/kernel/ |
H A D | ioport.c | 27 unsigned int i, max_long, bytes, bytes_updated; local 73 bytes = (max_long + 1) * sizeof(unsigned long); 74 bytes_updated = max(bytes, t->io_bitmap_max); 76 t->io_bitmap_max = bytes;
|
H A D | msr.c | 18 * and then read/write in chunks of 8 bytes. A larger size means multiple 76 ssize_t bytes = 0; local 90 bytes += 8; 93 return bytes ? bytes : err; 104 ssize_t bytes = 0; local 118 bytes += 8; 121 return bytes ? bytes : err;
|
/arch/arm/lib/ |
H A D | memzero.S | 17 * Align the pointer in r0. r3 contains the number of bytes that we are 18 * mis-aligned by, and r1 is the number of bytes. If r1 < 4, then we 22 blt 5f @ 1 bytes to align with? 41 blt 4f @ 1 have < 16 bytes 53 3: subs r1, r1, #64 @ 1 write 32 bytes out per loop 66 tst r1, #16 @ 1 16 bytes or more? 111 4: tst r1, #8 @ 1 8 bytes or more? 113 tst r1, #4 @ 1 4 bytes or more? 116 * When we get here, we've got less than 4 bytes to zero. We 119 5: tst r1, #2 @ 1 2 bytes o [all...] |
/arch/x86/crypto/ |
H A D | ghash-clmulni-intel_glue.c | 45 u32 bytes; member in struct:ghash_desc_ctx 80 if (dctx->bytes) { 81 int n = min(srclen, dctx->bytes); 82 u8 *pos = dst + (GHASH_BLOCK_SIZE - dctx->bytes); 84 dctx->bytes -= n; 90 if (!dctx->bytes) 100 dctx->bytes = GHASH_BLOCK_SIZE - srclen; 112 if (dctx->bytes) { 113 u8 *tmp = dst + (GHASH_BLOCK_SIZE - dctx->bytes); 115 while (dctx->bytes [all...] |
/arch/powerpc/platforms/powermac/ |
H A D | pfunc_core.c | 81 static void print_blob(const char *title, const void *blob, int bytes) 84 while(bytes--) { 271 u32 bytes = pmf_next32(cmd); local 273 LOG_PARSE("pmf: read_i2c(bytes: %ud)\n", bytes); 275 PMF_PARSE_CALL(read_i2c, cmd, h, bytes); 280 u32 bytes = pmf_next32(cmd); local 281 const void *blob = pmf_next_blob(cmd, bytes); 283 LOG_PARSE("pmf: write_i2c(bytes: %ud) ...\n", bytes); 311 u32 bytes = pmf_next32(cmd); local 322 u32 bytes = pmf_next32(cmd); local 354 u32 bytes = pmf_next32(cmd); local 365 u32 bytes = pmf_next32(cmd); local 487 u32 bytes = pmf_next32(cmd); local [all...] |
/arch/arm/mach-msm/include/mach/ |
H A D | msm_smd.h | 62 ** number of bytes are readable or writable. 64 int smd_wait_until_readable(smd_channel_t *ch, int bytes); 65 int smd_wait_until_writable(smd_channel_t *ch, int bytes);
|
/arch/x86/lib/ |
H A D | insn.c | 25 /* Verify next sizeof(t) bytes can be on the same instruction */ 63 * insn_get_prefixes - scan x86 instruction prefix bytes 87 if (prefixes->bytes[i] == b) 92 prefixes->bytes[nb++] = b; 111 if (lb && lb != insn->prefixes.bytes[3]) { 112 if (unlikely(insn->prefixes.bytes[3])) { 114 b = insn->prefixes.bytes[3]; 116 if (prefixes->bytes[i] == lb) 117 prefixes->bytes[i] = b; 119 insn->prefixes.bytes[ [all...] |
/arch/xtensa/lib/ |
H A D | usercopy.S | 26 * The returned value is the number of bytes not copied. Implies zero 113 # copy 2 bytes 146 movi a2, 0 # return success for len bytes copied 152 # copy 16 bytes per iteration for word-aligned dst and word-aligned src 179 # copy 8 bytes 188 # copy 4 bytes 195 # copy 2 bytes 206 movi a2, 0 # return success for len bytes copied 217 # copy 16 bytes per iteration for word-aligned dst and unaligned src 248 # copy 8 bytes [all...] |
/arch/ia64/include/asm/sn/ |
H A D | arch.h | 83 extern void sn_flush_all_caches(long addr, long bytes);
|
/arch/x86/tools/ |
H A D | test_get_len.c | 63 fprintf(fp, "%s\t.value = %d, bytes[] = {%x, %x, %x, %x},\n", 64 indent, field->value, field->bytes[0], field->bytes[1], 65 field->bytes[2], field->bytes[3]); 159 fprintf(stderr, "Warning: objdump says %d bytes, but "
|