/net/sunrpc/ |
H A D | socklib.c | 67 * @base: starting offset 72 ssize_t xdr_partial_copy_from_skb(struct xdr_buf *xdr, unsigned int base, struct xdr_skb_reader *desc, xdr_skb_read_actor copy_actor) argument 80 if (base < len) { 81 len -= base; 82 ret = copy_actor(desc, (char *)xdr->head[0].iov_base + base, len); 86 base = 0; 88 base -= len; 92 if (unlikely(base >= pglen)) { 93 base -= pglen; 96 if (base || xd [all...] |
H A D | xdr.c | 133 struct page **pages, unsigned int base, unsigned int len) 143 xdr->page_base = base; 166 * if a memory area starts at byte 'base' in page 'pages[i]', 167 * then its address is given as (i << PAGE_CACHE_SHIFT) + base 686 * @base: offset of first byte 690 void xdr_write_pages(struct xdr_stream *xdr, struct page **pages, unsigned int base, argument 696 buf->page_base = base; 729 unsigned int base, unsigned int len) 738 if (base >= maxlen) 740 maxlen -= base; 132 xdr_inline_pages(struct xdr_buf *xdr, unsigned int offset, struct page **pages, unsigned int base, unsigned int len) argument 728 xdr_set_page_base(struct xdr_stream *xdr, unsigned int base, unsigned int len) argument 1022 xdr_buf_subsegment(struct xdr_buf *buf, struct xdr_buf *subbuf, unsigned int base, unsigned int len) argument 1125 read_bytes_from_xdr_buf(struct xdr_buf *buf, unsigned int base, void *obj, unsigned int len) argument 1156 write_bytes_to_xdr_buf(struct xdr_buf *buf, unsigned int base, void *obj, unsigned int len) argument 1170 xdr_decode_word(struct xdr_buf *buf, unsigned int base, u32 *obj) argument 1184 xdr_encode_word(struct xdr_buf *buf, unsigned int base, u32 obj) argument 1232 xdr_xcode_array2(struct xdr_buf *buf, unsigned int base, struct xdr_array2_desc *desc, int encode) argument 1426 xdr_decode_array2(struct xdr_buf *buf, unsigned int base, struct xdr_array2_desc *desc) argument 1437 xdr_encode_array2(struct xdr_buf *buf, unsigned int base, struct xdr_array2_desc *desc) argument [all...] |
H A D | auth_unix.c | 148 __be32 *base, *hold; local 152 base = p++; 166 *base = htonl((p - base - 1) << 2); /* cred length */
|
H A D | xprtsock.c | 385 static int xs_send_kvec(struct socket *sock, struct sockaddr *addr, int addrlen, struct kvec *vec, unsigned int base, int more) argument 393 .iov_base = vec->iov_base + base, 394 .iov_len = vec->iov_len - base, 402 static int xs_send_pagedata(struct socket *sock, struct xdr_buf *xdr, unsigned int base, int more, bool zerocopy, int *sent_p) argument 410 remainder = xdr->page_len - base; 411 base += xdr->page_base; 412 ppage = xdr->pages + (base >> PAGE_SHIFT); 413 base &= ~PAGE_MASK; 418 unsigned int len = min_t(unsigned int, PAGE_SIZE - base, remainder); 424 err = do_sendpage(sock, *ppage, base, le 449 xs_sendpages(struct socket *sock, struct sockaddr *addr, int addrlen, struct xdr_buf *xdr, unsigned int base, bool zerocopy, int *sent_p) argument 555 rpc_fraghdr *base = buf->head[0].iov_base; local [all...] |
H A D | svcsock.c | 182 size_t base = xdr->page_base; local 202 size = PAGE_SIZE - base < pglen ? PAGE_SIZE - base : pglen; 206 result = kernel_sendpage(sock, *ppage, base, size, flags); 214 base = 0; 345 int buflen, unsigned int base) 352 if (base == 0) 356 if (iov[i].iov_len > base) 358 base -= iov[i].iov_len; 362 iov[i].iov_len -= base; 343 svc_partial_recvfrom(struct svc_rqst *rqstp, struct kvec *iov, int nr, int buflen, unsigned int base) argument 1089 unsigned int want, base; local [all...] |
/net/sctp/ |
H A D | endpointola.c | 55 * Initialize the base fields of the endpoint structure. 112 /* Initialize the base structure. */ 114 ep->base.type = SCTP_EP_TYPE_SOCKET; 117 atomic_set(&ep->base.refcnt, 1); 118 ep->base.dead = false; 121 sctp_inq_init(&ep->base.inqueue); 124 sctp_inq_set_th_handler(&ep->base.inqueue, sctp_endpoint_bh_rcv); 127 sctp_bind_addr_init(&ep->base.bind_addr, 0); 130 ep->base.sk = sk; 131 sock_hold(ep->base [all...] |
H A D | associola.c | 83 asoc->base.sk = (struct sock *)sk; 86 sock_hold(asoc->base.sk); 88 /* Initialize the common base substructure. */ 89 asoc->base.type = SCTP_EP_TYPE_ASSOCIATION; 92 atomic_set(&asoc->base.refcnt, 1); 95 sctp_bind_addr_init(&asoc->base.bind_addr, ep->base.bind_addr.port); 188 asoc->c.my_port = ep->base.bind_addr.port; 240 sctp_inq_init(&asoc->base.inqueue); 241 sctp_inq_set_th_handler(&asoc->base [all...] |
H A D | sm_sideeffect.c | 247 struct net *net = sock_net(asoc->base.sk); 251 bh_lock_sock(asoc->base.sk); 252 if (sock_owned_by_user(asoc->base.sk)) { 275 asoc->base.sk->sk_err = -error; 278 bh_unlock_sock(asoc->base.sk); 288 struct net *net = sock_net(asoc->base.sk); 291 bh_lock_sock(asoc->base.sk); 292 if (sock_owned_by_user(asoc->base.sk)) { 305 if (asoc->base.dead) 315 asoc->base [all...] |
H A D | ulpqueue.c | 176 struct sctp_sock *sp = sctp_sk(ulpq->asoc->base.sk); 187 return sctp_clear_pd(ulpq->asoc->base.sk, ulpq->asoc); 195 struct sock *sk = ulpq->asoc->base.sk; 489 if (!sctp_sk(asoc->base.sk)->frag_interleave && 490 atomic_read(&sctp_sk(asoc->base.sk)->pd_mode)) 494 pd_point = sctp_sk(asoc->base.sk)->pd_point; 496 retval = sctp_make_reassembled_event(sock_net(asoc->base.sk), 507 retval = sctp_make_reassembled_event(sock_net(ulpq->asoc->base.sk), 573 retval = sctp_make_reassembled_event(sock_net(ulpq->asoc->base.sk), 674 retval = sctp_make_reassembled_event(sock_net(ulpq->asoc->base [all...] |
H A D | chunk.c | 111 sp = sctp_sk(asoc->base.sk); 198 sctp_sk(asoc->base.sk)->pf->af->net_header_len - 254 SCTP_INC_STATS_USER(sock_net(asoc->base.sk), SCTP_MIB_FRAGUSRMSGS);
|
H A D | input.c | 180 rcvr = asoc ? &asoc->base : &ep->base; 198 rcvr = &ep->base; 503 sk = asoc->base.sk; 711 struct net *net = sock_net(ep->base.sk); 715 epb = &ep->base; 736 struct net *net = sock_net(ep->base.sk); 740 epb = &ep->base; 788 struct net *net = sock_net(asoc->base.sk); 792 epb = &asoc->base; [all...] |
H A D | ulpevent.c | 103 sctp_skb_set_owner_r(skb, asoc->base.sk); 344 sctp_get_pf_specific(asoc->base.sk->sk_family)->addr_to_user( 345 sctp_sk(asoc->base.sk), 652 rx_count = atomic_read(&asoc->base.sk->sk_rmem_alloc); 654 if (rx_count >= asoc->base.sk->sk_rcvbuf) { 656 if ((asoc->base.sk->sk_userlocks & SOCK_RCVBUF_LOCK) || 657 (!sk_rmem_schedule(asoc->base.sk, chunk->skb,
|
H A D | sm_make_chunk.c | 102 skb->sk = asoc ? asoc->base.sk : NULL; 221 struct net *net = sock_net(asoc->base.sk); 255 sp = sctp_sk(asoc->base.sk); 405 addrs = sctp_bind_addrs_to_raw(&asoc->base.bind_addr, &addrs_len, gfp); 424 sp = sctp_sk(asoc->base.sk); 1383 sk = asoc ? asoc->base.sk : NULL; 1580 asoc = sctp_association_new(ep, ep->base.sk, scope, gfp); 1665 if (sctp_sk(ep->base.sk)->hmac) { 1670 desc.tfm = sctp_sk(ep->base.sk)->hmac; 1732 if (!sctp_sk(ep->base [all...] |
H A D | probe.c | 147 ep->base.bind_addr.port == port || 159 ep->base.bind_addr.port, asoc->peer.port,
|
/net/ipv4/ |
H A D | inetpeer.c | 212 struct inet_peer_base *base) 214 struct inet_peer *u = rcu_dereference(base->root); 238 #define lookup_rightempty(start, base) \ 244 for (u = rcu_deref_locked(*v, base); \ 248 u = rcu_deref_locked(*v, base); \ 259 struct inet_peer_base *base) 267 node = rcu_deref_locked(*nodep, base); 268 l = rcu_deref_locked(node->avl_left, base); 269 r = rcu_deref_locked(node->avl_right, base); 275 ll = rcu_deref_locked(l->avl_left, base); 211 lookup_rcu(const struct inetpeer_addr *daddr, struct inet_peer_base *base) argument 257 peer_avl_rebalance(struct inet_peer __rcu **stack[], struct inet_peer __rcu ***stackend, struct inet_peer_base *base) argument 350 unlink_from_pool(struct inet_peer *p, struct inet_peer_base *base, struct inet_peer __rcu **stack[PEER_MAXDEPTH]) argument 383 inet_peer_gc(struct inet_peer_base *base, struct inet_peer __rcu **stack[PEER_MAXDEPTH], struct inet_peer __rcu ***stackptr) argument 419 inet_getpeer(struct inet_peer_base *base, const struct inetpeer_addr *daddr, int create) argument 543 inetpeer_invalidate_tree(struct inet_peer_base *base) argument [all...] |
/net/core/ |
H A D | iovec.c | 98 u8 __user *base = iov->iov_base + offset; local 109 if (copy_from_user(kdata, base, copy)) 112 base += copy; 122 if (copy_from_user(kdata, base, par_len)) 126 base += par_len; 136 if (copy_from_user(kdata + copy, base + copy, 143 csum = csum_and_copy_from_user(base, kdata, copy, 165 unsigned long seg, base; local 175 base = (unsigned long)iov[seg].iov_base + offset; 177 size = ((base [all...] |
H A D | datagram.c | 609 unsigned long base; local 619 base = (unsigned long)from->iov_base + offset; 620 size = ((base & ~PAGE_MASK) + len + ~PAGE_MASK) >> PAGE_SHIFT; 623 num_pages = get_user_pages_fast(base, size, 0, &page[i]); 634 int off = base & ~PAGE_MASK; 637 base += size;
|
/net/netfilter/ |
H A D | nf_conntrack_h323_asn1.c | 114 static int decode_nul(bitstr_t *bs, const struct field_t *f, char *base, int level); 115 static int decode_bool(bitstr_t *bs, const struct field_t *f, char *base, int level); 116 static int decode_oid(bitstr_t *bs, const struct field_t *f, char *base, int level); 117 static int decode_int(bitstr_t *bs, const struct field_t *f, char *base, int level); 118 static int decode_enum(bitstr_t *bs, const struct field_t *f, char *base, int level); 119 static int decode_bitstr(bitstr_t *bs, const struct field_t *f, char *base, int level); 120 static int decode_numstr(bitstr_t *bs, const struct field_t *f, char *base, int level); 121 static int decode_octstr(bitstr_t *bs, const struct field_t *f, char *base, int level); 122 static int decode_bmpstr(bitstr_t *bs, const struct field_t *f, char *base, int level); 123 static int decode_seq(bitstr_t *bs, const struct field_t *f, char *base, in 267 decode_nul(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 276 decode_bool(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 288 decode_oid(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 305 decode_int(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 349 decode_enum(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 365 decode_bitstr(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 399 decode_numstr(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 417 decode_octstr(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 466 decode_bmpstr(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 492 decode_seq(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 609 decode_seqof(bitstr_t *bs, const struct field_t *f, char *base, int level) argument 699 decode_choice(bitstr_t *bs, const struct field_t *f, char *base, int level) argument [all...] |
H A D | nft_payload.c | 29 switch (priv->base) { 67 priv->base = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_BASE])); 83 nla_put_be32(skb, NFTA_PAYLOAD_BASE, htonl(priv->base)) || 114 enum nft_payload_bases base; local 123 base = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_BASE])); 124 switch (base) { 139 base != NFT_PAYLOAD_LL_HEADER)
|
/net/sched/ |
H A D | cls_basic.c | 136 struct basic_filter *f, unsigned long base, 155 tcf_bind_filter(tp, &f->res, base); 169 struct tcf_proto *tp, unsigned long base, u32 handle, 217 err = basic_set_parms(net, tp, fnew, base, tb, tca[TCA_RATE], ovr); 135 basic_set_parms(struct net *net, struct tcf_proto *tp, struct basic_filter *f, unsigned long base, struct nlattr **tb, struct nlattr *est, bool ovr) argument 168 basic_change(struct net *net, struct sk_buff *in_skb, struct tcf_proto *tp, unsigned long base, u32 handle, struct nlattr **tca, unsigned long *arg, bool ovr) argument
|
H A D | cls_bpf.c | 167 unsigned long base, struct nlattr **tb, 214 tcf_bind_filter(tp, &prog->res, base); 241 struct tcf_proto *tp, unsigned long base, 280 ret = cls_bpf_modify_existing(net, tp, prog, base, tb, tca[TCA_RATE], ovr); 165 cls_bpf_modify_existing(struct net *net, struct tcf_proto *tp, struct cls_bpf_prog *prog, unsigned long base, struct nlattr **tb, struct nlattr *est, bool ovr) argument 240 cls_bpf_change(struct net *net, struct sk_buff *in_skb, struct tcf_proto *tp, unsigned long base, u32 handle, struct nlattr **tca, unsigned long *arg, bool ovr) argument
|
H A D | cls_fw.c | 185 struct nlattr **tb, struct nlattr **tca, unsigned long base, bool ovr) 199 tcf_bind_filter(tp, &f->res, base); 231 struct tcf_proto *tp, unsigned long base, 269 err = fw_change_attrs(net, tp, fnew, tb, tca, base, ovr); 314 err = fw_change_attrs(net, tp, f, tb, tca, base, ovr); 184 fw_change_attrs(struct net *net, struct tcf_proto *tp, struct fw_filter *f, struct nlattr **tb, struct nlattr **tca, unsigned long base, bool ovr) argument 230 fw_change(struct net *net, struct sk_buff *in_skb, struct tcf_proto *tp, unsigned long base, u32 handle, struct nlattr **tca, unsigned long *arg, bool ovr) argument
|
/net/ipv6/ |
H A D | ah6.c | 79 static inline struct tmp_ext *ah_tmp_ext(void *base) argument 81 return base + IPV6HDR_BASELEN; 298 static void ah6_output_done(struct crypto_async_request *base, int err) argument 303 struct sk_buff *skb = base->data; 393 * 8 bytes of the base IP header. 465 static void ah6_input_done(struct crypto_async_request *base, int err) argument 470 struct sk_buff *skb = base->data;
|
/net/ipv4/netfilter/ |
H A D | arp_tables.c | 238 get_entry(const void *base, unsigned int offset) argument 240 return (struct arpt_entry *)(base + offset); 275 * Ensure we load private-> members after we've fetched the base 562 const unsigned char *base, 587 if ((unsigned char *)e - base == hook_entries[h]) 589 if ((unsigned char *)e - base == underflows[h]) { 836 const void *base, struct xt_table_info *newinfo) 843 entry_offset = (void *)e - base; 854 (e < (struct arpt_entry *)(base + info->hook_entry[i]))) 857 (e < (struct arpt_entry *)(base 560 check_entry_size_and_hooks(struct arpt_entry *e, struct xt_table_info *newinfo, const unsigned char *base, const unsigned char *limit, const unsigned int *hook_entries, const unsigned int *underflows, unsigned int valid_hooks) argument 834 compat_calc_entry(const struct arpt_entry *e, const struct xt_table_info *info, const void *base, struct xt_table_info *newinfo) argument 1211 check_compat_entry_size_and_hooks(struct compat_arpt_entry *e, struct xt_table_info *newinfo, unsigned int *size, const unsigned char *base, const unsigned char *limit, const unsigned int *hook_entries, const unsigned int *underflows, const char *name) argument 1284 compat_copy_entry_from_user(struct compat_arpt_entry *e, void **dstptr, unsigned int *size, const char *name, struct xt_table_info *newinfo, unsigned char *base) argument [all...] |
/net/bridge/netfilter/ |
H A D | ebtables.c | 197 const char *base; local 219 /* base for chain jumps */ 220 base = private->entries; 263 BUGPRINT("RETURN on base chain"); 292 chaininfo = (struct ebt_entries *) (base + verdict); 688 a base chain */ 761 * the hook mask for udc tells us from which base chains the udc can be 765 unsigned int udc_cnt, unsigned int hooknr, char *base) 800 (struct ebt_entries *)(base + verdict); 823 /* this udc is accessible from the base chai 764 check_chainloops(const struct ebt_entries *chain, struct ebt_cl_stack *cl_s, unsigned int udc_cnt, unsigned int hooknr, char *base) argument 1342 ebt_make_matchname(const struct ebt_entry_match *m, const char *base, char __user *ubase) argument 1356 ebt_make_watchername(const struct ebt_entry_watcher *w, const char *base, char __user *ubase) argument 1369 ebt_make_names(struct ebt_entry *e, const char *base, char __user *ubase) argument 1738 compat_calc_entry(const struct ebt_entry *e, const struct ebt_table_info *info, const void *base, struct compat_ebt_replace *newinfo) argument 1897 compat_mtw_from_user(struct compat_ebt_entry_mwt *mwt, enum compat_mwt compat_mwt, struct ebt_entries_buf_state *state, const unsigned char *base) argument 1971 ebt_size_mwt(struct compat_ebt_entry_mwt *match32, unsigned int size_left, enum compat_mwt type, struct ebt_entries_buf_state *state, const void *base) argument 2030 size_entry_mwt(struct ebt_entry *entry, const unsigned char *base, unsigned int *total, struct ebt_entries_buf_state *state) argument [all...] |