/net/ipv4/netfilter/ |
H A D | ipt_MASQUERADE.c | 34 const struct nf_nat_ipv4_multi_range_compat *mr = par->targinfo; local 36 if (mr->range[0].flags & NF_NAT_RANGE_MAP_IPS) { 40 if (mr->rangesize != 1) { 41 pr_debug("bad rangesize %u\n", mr->rangesize); 51 const struct nf_nat_ipv4_multi_range_compat *mr; local 53 mr = par->targinfo; 54 range.flags = mr->range[0].flags; 55 range.min_proto = mr->range[0].min; 56 range.max_proto = mr->range[0].max;
|
/net/rds/ |
H A D | rdma.c | 43 * - should we limit the size of a mr region? let transport return failure? 71 struct rds_mr *mr; local 75 mr = rb_entry(parent, struct rds_mr, r_rb_node); 77 if (key < mr->r_key) 79 else if (key > mr->r_key) 82 return mr; 96 static void rds_destroy_mr(struct rds_mr *mr) argument 98 struct rds_sock *rs = mr->r_sock; 102 rdsdebug("RDS: destroy mr key is %x refcnt %u\n", 103 mr 119 __rds_put_mr_final(struct rds_mr *mr) argument 131 struct rds_mr *mr; local 176 struct rds_mr *mr = NULL, *found; local 359 struct rds_mr *mr; local 411 struct rds_mr *mr; local 716 struct rds_mr *mr; local [all...] |
H A D | iw.c | 94 rds_iwdev->mr = ib_get_dma_mr(rds_iwdev->pd, 98 if (IS_ERR(rds_iwdev->mr)) 101 rds_iwdev->mr = NULL; 118 if (rds_iwdev->mr) 119 ib_dereg_mr(rds_iwdev->mr); 149 if (rds_iwdev->mr) 150 ib_dereg_mr(rds_iwdev->mr);
|
H A D | ib.c | 102 if (rds_ibdev->mr) 103 ib_dereg_mr(rds_ibdev->mr); 167 rds_ibdev->mr = ib_get_dma_mr(rds_ibdev->pd, IB_ACCESS_LOCAL_WRITE); 168 if (IS_ERR(rds_ibdev->mr)) { 169 rds_ibdev->mr = NULL;
|
H A D | iw_rdma.c | 42 * This is stored as mr->r_trans_private. 49 struct ib_mr *mr; member in struct:rds_iw_mr 623 *key_ret = ibmr->mr->rkey; 625 printk(KERN_WARNING "RDS/IW: failed to map mr (errno=%d)\n", ret); 663 struct ib_mr *mr; local 666 mr = ib_alloc_fast_reg_mr(rds_iwdev->pd, pool->max_message_size); 667 if (IS_ERR(mr)) { 668 err = PTR_ERR(mr); 682 ib_dereg_mr(mr); 687 ibmr->mr [all...] |
H A D | rds.h | 229 /* Flags for mr->r_state */ 752 void __rds_put_mr_final(struct rds_mr *mr); 753 static inline void rds_mr_put(struct rds_mr *mr) argument 755 if (atomic_dec_and_test(&mr->r_refcount)) 756 __rds_put_mr_final(mr);
|
H A D | ib.h | 176 struct ib_mr *mr; member in struct:rds_ib_device
|
H A D | iw.h | 184 struct ib_mr *mr; member in struct:rds_iw_device
|
H A D | ib_cm.c | 292 ic->i_mr = rds_ibdev->mr; 396 rdsdebug("conn %p pd %p mr %p cq %p %p\n", conn, ic->i_pd, ic->i_mr,
|
H A D | iw_cm.c | 269 ic->i_mr = rds_iwdev->mr; 341 rdsdebug("conn %p pd %p mr %p cq %p %p\n", conn, ic->i_pd, ic->i_mr,
|
/net/mac80211/ |
H A D | rc80211_minstrel_ht_debugfs.c | 37 struct minstrel_rate_stats *mr = &mi->groups[i].rates[j]; local 62 tp = mr->cur_tp / 10; 63 prob = MINSTREL_TRUNC(mr->cur_prob * 1000); 64 eprob = MINSTREL_TRUNC(mr->probability * 1000); 71 mr->retry_count, 72 mr->last_success, 73 mr->last_attempts, 74 (unsigned long long)mr->succ_hist, 75 (unsigned long long)mr->att_hist);
|
H A D | rc80211_minstrel.c | 142 struct minstrel_rate *mr = &mi->r[i]; local 145 usecs = mr->perfect_tx_time; 176 mr->adjusted_retry_count = mrs->retry_count >> 1; 177 if (mr->adjusted_retry_count > 2) 178 mr->adjusted_retry_count = 2; 179 mr->sample_limit = 4; 181 mr->sample_limit = -1; 182 mr->adjusted_retry_count = mrs->retry_count; 184 if (!mr->adjusted_retry_count) 185 mr 264 minstrel_get_retry_count(struct minstrel_rate *mr, struct ieee80211_tx_info *info) argument 301 struct minstrel_rate *msr, *mr; local 463 struct minstrel_rate *mr = &mi->r[n]; local 518 struct minstrel_rate *mr = &mi->r[i]; local [all...] |
H A D | rc80211_minstrel_ht.c | 177 minstrel_calc_rate_ewma(struct minstrel_rate_stats *mr) argument 179 if (unlikely(mr->attempts > 0)) { 180 mr->sample_skipped = 0; 181 mr->cur_prob = MINSTREL_FRAC(mr->success, mr->attempts); 182 if (!mr->att_hist) 183 mr->probability = mr->cur_prob; 185 mr 205 struct minstrel_rate_stats *mr; local 281 struct minstrel_rate_stats *mr; local 352 struct minstrel_rate_stats *mr; local 384 struct minstrel_rate_stats *mr; local 652 struct minstrel_rate_stats *mr; local 712 struct minstrel_rate_stats *mr; local 785 struct minstrel_rate_stats *mr; local [all...] |
H A D | rc80211_minstrel_debugfs.c | 74 struct minstrel_rate *mr = &mi->r[i]; local 82 p += sprintf(p, "%3u%s", mr->bitrate / 2, 83 (mr->bitrate & 1 ? ".5" : " "));
|
/net/netfilter/ |
H A D | xt_REDIRECT.c | 85 const struct nf_nat_ipv4_multi_range_compat *mr = par->targinfo; local 87 if (mr->range[0].flags & NF_NAT_RANGE_MAP_IPS) { 91 if (mr->rangesize != 1) { 92 pr_debug("bad rangesize %u.\n", mr->rangesize); 104 const struct nf_nat_ipv4_multi_range_compat *mr = par->targinfo; local 135 newrange.flags = mr->range[0].flags | NF_NAT_RANGE_MAP_IPS; 138 newrange.min_proto = mr->range[0].min; 139 newrange.max_proto = mr->range[0].max;
|
H A D | xt_NETMAP.c | 72 const struct nf_nat_ipv4_multi_range_compat *mr = par->targinfo; local 81 netmask = ~(mr->range[0].min_ip ^ mr->range[0].max_ip); 88 new_ip |= mr->range[0].min_ip & netmask; 92 newrange.flags = mr->range[0].flags | NF_NAT_RANGE_MAP_IPS; 95 newrange.min_proto = mr->range[0].min; 96 newrange.max_proto = mr->range[0].max; 104 const struct nf_nat_ipv4_multi_range_compat *mr = par->targinfo; local 106 if (!(mr->range[0].flags & NF_NAT_RANGE_MAP_IPS)) { 110 if (mr [all...] |
H A D | xt_nat.c | 19 const struct nf_nat_ipv4_multi_range_compat *mr = par->targinfo; local 21 if (mr->rangesize != 1) { 45 const struct nf_nat_ipv4_multi_range_compat *mr = par->targinfo; local 55 xt_nat_convert_range(&range, &mr->range[0]); 62 const struct nf_nat_ipv4_multi_range_compat *mr = par->targinfo; local 71 xt_nat_convert_range(&range, &mr->range[0]);
|
/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_recvfrom.c | 302 key = (u8)(frmr->mr->lkey & 0x000000FF); 303 ib_update_fast_reg_key(frmr->mr, ++key); 306 ctxt->sge[0].lkey = frmr->mr->lkey; 321 fastreg_wr.wr.fast_reg.rkey = frmr->mr->lkey; 334 read_wr.ex.invalidate_rkey = ctxt->frmr->mr->lkey; 343 inv_wr.ex.invalidate_rkey = frmr->mr->lkey;
|
H A D | svc_rdma_transport.c | 742 struct ib_mr *mr; local 750 mr = ib_alloc_fast_reg_mr(xprt->sc_pd, RPCSVC_MAXPAGES); 751 if (IS_ERR(mr)) 759 frmr->mr = mr; 765 ib_dereg_mr(mr); 780 ib_dereg_mr(frmr->mr); 811 if (ib_dma_mapping_error(frmr->mr->device, addr)) 814 ib_dma_unmap_page(frmr->mr->device, addr, PAGE_SIZE, 1159 /* De-allocate fastreg mr */ [all...] |
H A D | verbs.c | 1255 * 1. recv mr memory (mr free, then kfree) 1256 * 2. send mr memory (mr free, then kfree) 1612 struct ib_mr *mr; local 1637 mr = ib_reg_phys_mr(ia->ri_pd, &ipb, 1, 1645 if (IS_ERR(mr)) { 1647 rc = PTR_ERR(mr); 1650 *mrp = mr; 1651 iov->lkey = mr 1659 rpcrdma_deregister_internal(struct rpcrdma_ia *ia, struct ib_mr *mr, struct ib_sge *iov) argument 1720 struct ib_mr *mr = frmr->fr_mr; local [all...] |