/arch/s390/lib/ |
H A D | string.c | 66 * @src: Where to copy the string from 70 char *strcpy(char *dest, const char *src) argument 77 : "+&a" (dest), "+&a" (src) : "d" (r0) 86 * @src: Where to copy the string from 94 size_t strlcpy(char *dest, const char *src, size_t size) argument 96 size_t ret = __strend(src) - src; 101 memcpy(dest, src, len); 110 * @src: Where to copy the string from 116 char *strncpy(char *dest, const char *src, size_ argument 132 strcat(char *dest, const char *src) argument 154 strlcat(char *dest, const char *src, size_t n) argument 183 strncat(char *dest, const char *src, size_t n) argument [all...] |
/arch/mn10300/lib/ |
H A D | checksum.c | 54 __wsum csum_partial_copy(const void *src, void *dst, int len, __wsum sum) argument 56 copy_from_user(dst, src, len); 61 __wsum csum_partial_copy_nocheck(const void *src, void *dst, argument 64 sum = csum_partial(src, len, sum); 65 memcpy(dst, src, len); 70 __wsum csum_partial_copy_from_user(const void *src, void *dst, argument 76 missing = copy_from_user(dst, src, len); 86 __wsum csum_and_copy_to_user(const void *src, void *dst, argument 92 missing = copy_to_user(dst, src, len); 98 return csum_partial(src, le [all...] |
H A D | usercopy.c | 33 #define __do_strncpy_from_user(dst, src, count, res) \ 64 :"i"(-EFAULT), "1"(count), "a"(src), "a"(dst) \ 69 __strncpy_from_user(char *dst, const char *src, long count) argument 72 __do_strncpy_from_user(dst, src, count, res); 77 strncpy_from_user(char *dst, const char *src, long count) argument 80 if (access_ok(VERIFY_READ, src, 1)) 81 __do_strncpy_from_user(dst, src, count, res);
|
/arch/parisc/math-emu/ |
H A D | fcnvfu.c | 65 register unsigned int src, result; local 69 src = *srcptr; 70 src_exponent = Sgl_exponent(src) - SGL_BIAS; 76 if (Sgl_isone_sign(src)) { 96 if (Sgl_isone_sign(src)) { 105 Sgl_clear_signexponent_set_hidden(src); 106 Suint_from_sgl_mantissa(src,src_exponent,result); 109 if (Sgl_isinexact_to_unsigned(src,src_exponent)) { 119 if (Sgl_isone_roundbit(src,src_exponent) && 120 (Sgl_isone_stickybit(src,src_exponen 187 register unsigned int src, resultp1, resultp2; local [all...] |
H A D | frnd.c | 53 register unsigned int src, result; local 57 src = *srcptr; 61 if ((src_exponent = Sgl_exponent(src)) == SGL_INFINITY_EXPONENT) { 65 if (Sgl_isone_signaling(src)) { 70 Sgl_set_quiet(src); 75 *dstptr = src; 82 *dstptr = src; 89 Sgl_clear_exponent_set_hidden(src); 90 result = src; 93 if (Sgl_isinexact_to_fix(src,src_exponen [all...] |
H A D | fcnvfx.c | 64 register unsigned int src, temp; local 68 src = *srcptr; 69 src_exponent = Sgl_exponent(src) - SGL_BIAS; 77 Sgl_isnotzero_mantissa(src) || Sgl_iszero_sign(src)) { 78 if (Sgl_iszero_sign(src)) result = 0x7fffffff; 93 temp = src; 96 if (Sgl_isone_sign(src)) result = -Sgl_all(temp); 100 if (Sgl_isinexact_to_fix(src,src_exponent)) { 105 if (Sgl_iszero_sign(src)) resul 162 register unsigned int src, temp, resultp2; local [all...] |
H A D | fcnvfut.c | 63 register unsigned int src, result; local 66 src = *srcptr; 67 src_exponent = Sgl_exponent(src) - SGL_BIAS; 73 if (Sgl_isone_sign(src)) { 93 if (Sgl_isone_sign(src)) { 102 Sgl_clear_signexponent_set_hidden(src); 103 Suint_from_sgl_mantissa(src,src_exponent,result); 107 if (Sgl_isinexact_to_unsigned(src,src_exponent)) { 116 if (Sgl_isnotzero_exponentmantissa(src)) { 133 register unsigned int src, resultp local [all...] |
/arch/x86/crypto/ |
H A D | des3_ede_glue.c | 38 const u8 *src); 42 const u8 *src); 45 const u8 *src) 49 des3_ede_x86_64_crypt_blk(enc_ctx, dst, src); 53 const u8 *src) 57 des3_ede_x86_64_crypt_blk(dec_ctx, dst, src); 61 const u8 *src) 65 des3_ede_x86_64_crypt_blk_3way(enc_ctx, dst, src); 69 const u8 *src) 73 des3_ede_x86_64_crypt_blk_3way(dec_ctx, dst, src); 44 des3_ede_enc_blk(struct des3_ede_x86_ctx *ctx, u8 *dst, const u8 *src) argument 52 des3_ede_dec_blk(struct des3_ede_x86_ctx *ctx, u8 *dst, const u8 *src) argument 60 des3_ede_enc_blk_3way(struct des3_ede_x86_ctx *ctx, u8 *dst, const u8 *src) argument 68 des3_ede_dec_blk_3way(struct des3_ede_x86_ctx *ctx, u8 *dst, const u8 *src) argument 76 des3_ede_x86_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) argument 78 des3_ede_enc_blk(crypto_tfm_ctx(tfm), dst, src); local 81 des3_ede_x86_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) argument 83 des3_ede_dec_blk(crypto_tfm_ctx(tfm), dst, src); local 130 ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) argument 140 ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) argument 156 u64 *src = (u64 *)walk->src.virt.addr; local 174 cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) argument 197 u64 *src = (u64 *)walk->src.virt.addr; local 253 cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) argument 275 u8 *src = walk->src.virt.addr; local 292 __be64 *src = (__be64 *)walk->src.virt.addr; local 337 ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) argument [all...] |
H A D | glue_helper-asm-avx2.S | 13 #define load_16way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ 14 vmovdqu (0*32)(src), x0; \ 15 vmovdqu (1*32)(src), x1; \ 16 vmovdqu (2*32)(src), x2; \ 17 vmovdqu (3*32)(src), x3; \ 18 vmovdqu (4*32)(src), x4; \ 19 vmovdqu (5*32)(src), x5; \ 20 vmovdqu (6*32)(src), x6; \ 21 vmovdqu (7*32)(src), x7; 33 #define store_cbc_16way(src, ds [all...] |
/arch/hexagon/mm/ |
H A D | copy_user_template.S | 34 r3 = or(dst,src) 35 r4 = xor(dst,src) 41 src_dst_sav = combine(src,dst) 54 d_dbuf = memd(src++#8) 84 w_dbuf = memw(src++#4) 110 w_dbuf = memuh(src++#2) 126 w_dbuf = memub(src++#1) 150 if (p0.new) w_dbuf = memub(src) 151 p0 = tstbit(src,#0) 158 if (p0) src [all...] |
/arch/x86/include/asm/crypto/ |
H A D | glue_helper.h | 13 typedef void (*common_glue_func_t)(void *ctx, u8 *dst, const u8 *src); 14 typedef void (*common_glue_cbc_func_t)(void *ctx, u128 *dst, const u128 *src); 15 typedef void (*common_glue_ctr_func_t)(void *ctx, u128 *dst, const u128 *src, 17 typedef void (*common_glue_xts_func_t)(void *ctx, u128 *dst, const u128 *src, 78 static inline void le128_to_be128(be128 *dst, const le128 *src) argument 80 dst->a = cpu_to_be64(le64_to_cpu(src->a)); 81 dst->b = cpu_to_be64(le64_to_cpu(src->b)); 84 static inline void be128_to_le128(le128 *dst, const be128 *src) argument 86 dst->a = cpu_to_le64(be64_to_cpu(src->a)); 87 dst->b = cpu_to_le64(be64_to_cpu(src 103 le128_gf128mul_x_ble(le128 *dst, const le128 *src) argument [all...] |
H A D | twofish.h | 21 const u8 *src); 23 const u8 *src); 27 const u8 *src, bool xor); 29 const u8 *src); 32 extern void twofish_dec_blk_cbc_3way(void *ctx, u128 *dst, const u128 *src); 33 extern void twofish_enc_blk_ctr(void *ctx, u128 *dst, const u128 *src, 35 extern void twofish_enc_blk_ctr_3way(void *ctx, u128 *dst, const u128 *src,
|
/arch/x86/lib/ |
H A D | csum-wrappers_64.c | 13 * @src: source address (user space) 20 * src and dst are best aligned to 64bits. 23 csum_partial_copy_from_user(const void __user *src, void *dst, argument 29 if (!likely(access_ok(VERIFY_READ, src, len))) 40 if (unlikely((unsigned long)src & 6)) { 41 while (((unsigned long)src & 6) && len >= 2) { 44 if (__get_user(val16, (const __u16 __user *)src)) 50 src += 2; 56 isum = csum_partial_copy_generic((__force const void *)src, 74 * @src 84 csum_partial_copy_to_user(const void *src, void __user *dst, int len, __wsum isum, int *errp) argument 130 csum_partial_copy_nocheck(const void *src, void *dst, int len, __wsum sum) argument [all...] |
/arch/arm64/lib/ |
H A D | memmove.S | 10 * files/head:/src/aarch64/ 30 * Move a buffer from src to test (alignment handled by the hardware). 31 * If dest <= src, call memcpy, otherwise copy in reverse order. 35 * x1 - src 41 src .req x1 label 61 cmp dstin, src 63 add tmp1, src, count 68 add src, src, count 72 ands tmp2, src, #1 [all...] |
H A D | memcpy.S | 10 * files/head:/src/aarch64/ 30 * Copy a buffer from src to dest (alignment handled by the hardware) 34 * x1 - src 40 src .req x1 label 65 neg tmp2, src 70 * Copy the leading memory data from src to dst in an increasing 72 * memory data is eliminated when the distance between src and 76 ldrb tmp1w, [src], #1 80 ldrh tmp1w, [src], #2 84 ldr tmp1w, [src], # [all...] |
/arch/mips/cavium-octeon/ |
H A D | octeon-memcpy.S | 21 #define src a1 define 27 * memcpy copies len bytes from src to dst and sets v0 to dst. 29 * - src and dst don't overlap 30 * - src is readable 34 * __copy_user copies up to len bytes from src to dst and sets a2 (len) to 36 * __copy_user assumes that src and dst don't overlap, and that the call is 39 * - src is readable (no exceptions when reading src) 58 * 2- src_entry <= src < AT, and 59 * 3- (dst - src) [all...] |
/arch/mips/lib/ |
H A D | csum_partial.S | 74 #define CSUM_BIGCHUNK1(src, offset, sum, _t0, _t1, _t2, _t3) \ 75 LOAD _t0, (offset + UNIT(0))(src); \ 76 LOAD _t1, (offset + UNIT(1))(src); \ 77 LOAD _t2, (offset + UNIT(2))(src); \ 78 LOAD _t3, (offset + UNIT(3))(src); \ 85 #define CSUM_BIGCHUNK(src, offset, sum, _t0, _t1, _t2, _t3) \ 86 CSUM_BIGCHUNK1(src, offset, sum, _t0, _t1, _t2, _t3) 88 #define CSUM_BIGCHUNK(src, offset, sum, _t0, _t1, _t2, _t3) \ 89 CSUM_BIGCHUNK1(src, offset, sum, _t0, _t1, _t2, _t3); \ 90 CSUM_BIGCHUNK1(src, offse 99 #define src a0 define 315 #define src a0 define [all...] |
/arch/m68k/math-emu/ |
H A D | multi_arith.h | 82 static inline int fp_addmant(struct fp_ext *dest, struct fp_ext *src) argument 88 : "g,d" (src->lowmant), "0,0" (dest->lowmant)); 90 : "d" (src->mant.m32[1]), "0" (dest->mant.m32[1])); 92 : "d" (src->mant.m32[0]), "0" (dest->mant.m32[0])); 140 #define fp_addx96(dest, src) ({ \ 149 #define fp_sub64(dest, src) ({ \ 151 : "dm,d" (src.m32[1]), "0,0" (dest.m32[1])); \ 153 : "d" (src.m32[0]), "0" (dest.m32[0])); \ 181 static inline void fp_dividemant(union fp_mant128 *dest, struct fp_ext *src, argument 192 if (src 244 fp_putmant128(struct fp_ext *dest, union fp_mant128 *src, int shift) argument [all...] |
/arch/blackfin/include/asm/ |
H A D | string.h | 15 extern char *strcpy(char *dest, const char *src); 18 extern char *strncpy(char *dest, const char *src, size_t n); 35 extern void *memmove(void *dest, const void *src, size_t count);
|
/arch/arc/mm/ |
H A D | extable.c | 51 long arc_strncpy_from_user_noinline(char *dst, const char __user *src, argument 54 return __arc_strncpy_from_user(dst, src, count); 58 long arc_strnlen_user_noinline(const char __user *src, long n) argument 60 return __arc_strnlen_user(src, n);
|
/arch/cris/include/asm/ |
H A D | cacheflush.h | 25 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 26 memcpy(dst, src, len) 27 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 28 memcpy(dst, src, len)
|
/arch/metag/mm/ |
H A D | maccess.c | 18 long probe_kernel_write(void *dst, const void *src, size_t size) argument 22 unsigned long lsrc = (unsigned long)src; 23 const u8 *psrc = (u8 *)src; 43 return __probe_kernel_write(dst, src, size); 45 /* If src is unaligned, copy to the aligned bounce buffer first. */
|
/arch/mips/jz4740/ |
H A D | prom.c | 31 char *src; local 34 src = argv[i]; 35 while (*src && count) { 36 *dst++ = *src++;
|
/arch/sparc/include/asm/ |
H A D | uaccess.h | 12 long strncpy_from_user(char *dest, const char __user *src, long count);
|
/arch/mips/boot/compressed/ |
H A D | string.c | 9 void *memcpy(void *dest, const void *src, size_t n) argument 12 const char *s = src;
|