/arch/arm/lib/ |
H A D | csumpartialcopygeneric.S | 13 * csum_partial_copy_xxx(const char *src, char *dst, int len, int sum, ) 14 * r0 = src, r1 = dst, r2 = len, r3 = sum 21 dst .req r1 label 35 tst dst, #1 41 strb ip, [dst], #1 42 tst dst, #2 43 moveq pc, lr @ dst is now 32bit aligned 48 strb r8, [dst], #1 50 strb ip, [dst], #1 51 mov pc, lr @ dst i [all...] |
/arch/alpha/lib/ |
H A D | srm_printk.c | 14 char *src, *dst; local 31 for (dst = src + num_lf; src >= buf; ) { 33 *dst-- = '\r'; 35 *dst-- = *src--;
|
/arch/arm/kernel/ |
H A D | io.c | 39 void _memset_io(volatile void __iomem *dst, int c, size_t count) argument 43 writeb(c, dst); 44 dst++;
|
/arch/frv/kernel/ |
H A D | uaccess.c | 20 long strncpy_from_user(char *dst, const char __user *src, long count) argument 28 p = dst; 40 memset(dst + max, 0, count - max); 54 err = p - dst; /* return length excluding NUL */
|
/arch/hexagon/lib/ |
H A D | io.c | 34 short int *dst = (short int *) data; local 40 *dst++ = *src; 53 volatile short int *dst = (short int *)addr; local 59 *dst = *src++; 68 long *dst = (long *) data; local 74 *dst++ = *src; 82 volatile long *dst = (long *)addr; local 88 *dst = *src++;
|
/arch/m32r/lib/ |
H A D | csum_partial_copy.c | 31 csum_partial_copy_nocheck (const void *src, void *dst, int len, __wsum sum) argument 34 memcpy(dst, src, len); 45 csum_partial_copy_from_user (const void __user *src, void *dst, argument 50 missing = copy_from_user(dst, src, len); 52 memset(dst + len - missing, 0, missing); 56 return csum_partial(dst, len-missing, sum);
|
/arch/microblaze/lib/ |
H A D | memcpy.c | 39 char *dst = v_dst; local 43 *dst++ = *src++; 51 char *dst = v_dst; local 67 switch ((unsigned long)dst & 3) { 69 *dst++ = *src++; 72 *dst++ = *src++; 75 *dst++ = *src++; 79 i_dst = (void *)dst; 171 dst = (void *)i_dst; 178 *dst [all...] |
H A D | memmove.c | 38 char *dst = v_dst; local 49 dst += c; 53 *--dst = *--src; 61 char *dst = v_dst; local 80 dst += c; 89 switch ((unsigned long)dst & 3) { 91 *--dst = *--src; 94 *--dst = *--src; 97 *--dst = *--src; 101 i_dst = (void *)dst; [all...] |
/arch/powerpc/lib/ |
H A D | checksum_wrappers_64.c | 26 __wsum csum_and_copy_from_user(const void __user *src, void *dst, argument 46 csum = csum_partial_copy_generic((void __force *)src, dst, 50 int missing = __copy_from_user(dst, src, len); 53 memset(dst + len - missing, 0, missing); 59 csum = csum_partial(dst, len, sum); 67 __wsum csum_and_copy_to_user(const void *src, void __user *dst, int len, argument 81 if (unlikely((len < 0) || !access_ok(VERIFY_WRITE, dst, len))) { 87 csum = csum_partial_copy_generic(src, (void __force *)dst, 93 if (copy_to_user(dst, src, len)) {
|
/arch/score/lib/ |
H A D | checksum_copy.c | 30 unsigned int csum_partial_copy(const char *src, char *dst, argument 34 memcpy(dst, src, len); 39 unsigned int csum_partial_copy_from_user(const char *src, char *dst, argument 45 missing = copy_from_user(dst, src, len); 47 memset(dst + len - missing, 0, missing); 51 return csum_partial(dst, len, sum);
|
/arch/x86/crypto/ |
H A D | twofish_glue.c | 47 asmlinkage void twofish_enc_blk(struct twofish_ctx *ctx, u8 *dst, 50 asmlinkage void twofish_dec_blk(struct twofish_ctx *ctx, u8 *dst, 54 static void twofish_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) argument 56 twofish_enc_blk(crypto_tfm_ctx(tfm), dst, src); local 59 static void twofish_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) argument 61 twofish_dec_blk(crypto_tfm_ctx(tfm), dst, src); local
|
H A D | aes_glue.c | 13 void crypto_aes_encrypt_x86(struct crypto_aes_ctx *ctx, u8 *dst, const u8 *src) argument 15 aes_enc_blk(ctx, dst, src); 19 void crypto_aes_decrypt_x86(struct crypto_aes_ctx *ctx, u8 *dst, const u8 *src) argument 21 aes_dec_blk(ctx, dst, src); 25 static void aes_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) argument 27 aes_enc_blk(crypto_tfm_ctx(tfm), dst, src); local 30 static void aes_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) argument 32 aes_dec_blk(crypto_tfm_ctx(tfm), dst, src); local
|
/arch/x86/include/asm/ |
H A D | serpent.h | 11 asmlinkage void __serpent_enc_blk_4way(struct serpent_ctx *ctx, u8 *dst, 13 asmlinkage void serpent_dec_blk_4way(struct serpent_ctx *ctx, u8 *dst, 16 static inline void serpent_enc_blk_xway(struct serpent_ctx *ctx, u8 *dst, argument 19 __serpent_enc_blk_4way(ctx, dst, src, false); 22 static inline void serpent_enc_blk_xway_xor(struct serpent_ctx *ctx, u8 *dst, argument 25 __serpent_enc_blk_4way(ctx, dst, src, true); 28 static inline void serpent_dec_blk_xway(struct serpent_ctx *ctx, u8 *dst, argument 31 serpent_dec_blk_4way(ctx, dst, src); 38 asmlinkage void __serpent_enc_blk_8way(struct serpent_ctx *ctx, u8 *dst, 40 asmlinkage void serpent_dec_blk_8way(struct serpent_ctx *ctx, u8 *dst, 43 serpent_enc_blk_xway(struct serpent_ctx *ctx, u8 *dst, const u8 *src) argument 49 serpent_enc_blk_xway_xor(struct serpent_ctx *ctx, u8 *dst, const u8 *src) argument 55 serpent_dec_blk_xway(struct serpent_ctx *ctx, u8 *dst, const u8 *src) argument [all...] |
/arch/xtensa/kernel/ |
H A D | io.c | 47 void insb(unsigned long addr, void *dst, unsigned long count) { argument 50 *(unsigned char *)dst = readb(addr); 51 dst += 1; 56 void insw(unsigned long addr, void *dst, unsigned long count) { argument 59 *(unsigned short *)dst = readw(addr); 60 dst += 2; 65 void insl(unsigned long addr, void *dst, unsigned long count) { argument 71 *(unsigned long *)dst = readl(addr); 72 dst += 4;
|
/arch/blackfin/mm/ |
H A D | maccess.c | 19 long probe_kernel_read(void *dst, const void *src, size_t size) argument 31 memcpy(dst, &mmr, sizeof(mmr)); 35 memcpy(dst, &mmr, sizeof(mmr)); 42 return __probe_kernel_read(dst, src, size); 45 if (dma_memcpy(dst, src, size)) 49 if (isram_memcpy(dst, src, size)) 58 long probe_kernel_write(void *dst, const void *src, size_t size) argument 60 unsigned long ldst = (unsigned long)dst; 71 bfin_write16(dst, mmr); 76 bfin_write32(dst, mm [all...] |
/arch/c6x/lib/ |
H A D | checksum.c | 16 csum_partial_copy_from_user(const void __user *src, void *dst, int len, argument 21 missing = __copy_from_user(dst, src, len); 23 memset(dst + len - missing, 0, missing); 28 return csum_partial(dst, len, sum);
|
/arch/hexagon/include/asm/ |
H A D | cacheflush.h | 86 void *dst, void *src, int len) 88 memcpy(dst, src, len); 90 flush_icache_range((unsigned long) dst, 91 (unsigned long) dst + len); 83 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
|
H A D | uaccess.h | 87 #define __strncpy_from_user(dst, src, n) hexagon_strncpy_from_user(dst, src, n) 94 static inline long hexagon_strncpy_from_user(char *dst, const char __user *src, 100 static inline long hexagon_strncpy_from_user(char *dst, const char __user *src, argument 108 copy_from_user(dst, src, n); 111 copy_from_user(dst, src, res);
|
/arch/hexagon/mm/ |
H A D | copy_from_user.S | 42 #define dst r0 define 66 memd(dst) = d_dbuf 76 memw(dst) = w_dbuf 86 memh(dst) = w_dbuf 96 memb(dst) = w_dbuf
|
H A D | copy_to_user.S | 41 #define dst r0 define 59 /* Normal copy loops. Use dst-dst_sav to compute distance */ 60 /* dst holds best write, no need to unwind any loops */ 72 r2 += sub(dst_sav,dst)
|
/arch/m68k/lib/ |
H A D | checksum.c | 136 csum_partial_copy_from_user(const void __user *src, void *dst, argument 313 : "=d" (sum), "=d" (len), "=a" (src), "=a" (dst), 315 : "0" (sum), "1" (len), "2" (src), "3" (dst) 331 csum_partial_copy_nocheck(const void *src, void *dst, int len, __wsum sum) argument 419 : "=d" (sum), "=d" (len), "=a" (src), "=a" (dst), 421 : "0" (sum), "1" (len), "2" (src), "3" (dst)
|
H A D | uaccess.c | 109 long strncpy_from_user(char *dst, const char __user *src, long count) argument 135 : "=d" (res), "+a" (dst), "+a" (src), "+r" (count), "=&d" (c)
|
/arch/mips/jz4740/ |
H A D | prom.c | 30 char *dst = &(arcs_cmdline[0]); local 36 *dst++ = *src++; 39 *dst++ = ' '; 42 --dst; 44 *dst = 0;
|
/arch/sh/kernel/ |
H A D | io.c | 106 void memset_io(volatile void __iomem *dst, int c, unsigned long count) argument 110 writeb(c, dst); 111 dst++;
|
/arch/x86/lib/ |
H A D | csum-wrappers_64.c | 13 * @dst: destination address 19 * src and dst are best aligned to 64bits. 22 csum_partial_copy_from_user(const void __user *src, void *dst, argument 47 *(__u16 *)dst = val16; 51 dst += 2; 56 dst, len, isum, errp, NULL); 64 memset(dst, 0, len); 73 * @dst: destination address (user space) 79 * src and dst are best aligned to 64bits. 82 csum_partial_copy_to_user(const void *src, void __user *dst, argument 123 csum_partial_copy_nocheck(const void *src, void *dst, int len, __wsum sum) argument [all...] |