/net/sched/ |
H A D | sch_plug.c | 140 struct tc_plug_qopt *ctl = nla_data(opt); local 142 if (nla_len(opt) < sizeof(*ctl)) 145 q->limit = ctl->limit;
|
H A D | sch_fifo.c | 64 struct tc_fifo_qopt *ctl = nla_data(opt); local 66 if (nla_len(opt) < sizeof(*ctl)) 69 sch->limit = ctl->limit;
|
H A D | sch_red.c | 183 struct tc_red_qopt *ctl; local 201 ctl = nla_data(tb[TCA_RED_PARMS]); 203 if (ctl->limit > 0) { 204 child = fifo_create_dflt(sch, &bfifo_qdisc_ops, ctl->limit); 210 q->flags = ctl->flags; 211 q->limit = ctl->limit; 219 ctl->qth_min, ctl->qth_max, ctl->Wlog, 220 ctl [all...] |
H A D | sch_gred.c | 384 struct tc_gred_qopt *ctl, int prio, 400 q->limit = ctl->limit; 406 ctl->qth_min, ctl->qth_max, ctl->Wlog, ctl->Plog, 407 ctl->Scell_log, stab, max_P); 422 struct tc_gred_qopt *ctl; local 446 ctl = nla_data(tb[TCA_GRED_PARMS]); 449 if (ctl 383 gred_change_vq(struct Qdisc *sch, int dp, struct tc_gred_qopt *ctl, int prio, u8 *stab, u32 max_P, struct gred_sched_data **prealloc) argument [all...] |
H A D | sch_choke.c | 413 const struct tc_red_qopt *ctl; local 432 ctl = nla_data(tb[TCA_CHOKE_PARMS]); 434 if (ctl->limit > CHOKE_MAX_QUEUE) 437 mask = roundup_pow_of_two(ctl->limit + 1) - 1; 476 q->flags = ctl->flags; 477 q->limit = ctl->limit; 479 red_set_parms(&q->parms, ctl->qth_min, ctl->qth_max, ctl->Wlog, 480 ctl [all...] |
H A D | sch_sfb.c | 494 const struct tc_sfb_qopt *ctl = &sfb_default_ops; local 506 ctl = nla_data(tb[TCA_SFB_PARMS]); 509 limit = ctl->limit; 523 q->rehash_interval = msecs_to_jiffies(ctl->rehash_interval); 524 q->warmup_time = msecs_to_jiffies(ctl->warmup_time); 527 q->increment = ctl->increment; 528 q->decrement = ctl->decrement; 529 q->max = ctl->max; 530 q->bin_size = ctl->bin_size; 531 q->penalty_rate = ctl [all...] |
H A D | sch_sfq.c | 642 struct tc_sfq_qopt *ctl = nla_data(opt); local 647 if (opt->nla_len < nla_attr_size(sizeof(*ctl))) 651 if (ctl->divisor && 652 (!is_power_of_2(ctl->divisor) || ctl->divisor > 65536)) 660 if (ctl->quantum) { 661 q->quantum = ctl->quantum; 664 q->perturb_period = ctl->perturb_period * HZ; 665 if (ctl->flows) 666 q->maxflows = min_t(u32, ctl [all...] |
/net/sctp/ |
H A D | sysctl.c | 65 static int proc_sctp_do_hmac_alg(ctl_table *ctl, 303 static int proc_sctp_do_hmac_alg(ctl_table *ctl, argument
|
/net/ipv4/ |
H A D | sysctl_net_ipv4.c | 139 static int proc_tcp_default_init_rwnd(ctl_table *ctl, int write, argument 143 int old_value = *(int *)ctl->data; 144 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos); 145 int new_value = *(int *)ctl->data; 148 *(int *)ctl->data = old_value; 153 static int proc_tcp_congestion_control(ctl_table *ctl, int write, argument 171 static int proc_tcp_available_congestion_control(ctl_table *ctl, argument 188 static int proc_allowed_congestion_control(ctl_table *ctl, argument 208 static int ipv4_tcp_mem(ctl_table *ctl, int write, argument 222 .mode = ctl 251 proc_tcp_fastopen_key(ctl_table *ctl, int write, void __user *buffer, size_t *lenp, loff_t *ppos) argument [all...] |
H A D | devinet.c | 1950 static int devinet_conf_proc(ctl_table *ctl, int write, argument 1954 int old_value = *(int *)ctl->data; 1955 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos); 1956 int new_value = *(int *)ctl->data; 1959 struct ipv4_devconf *cnf = ctl->extra1; 1960 struct net *net = ctl->extra2; 1961 int i = (int *)ctl->data - cnf->data; 1993 static int devinet_sysctl_forward(ctl_table *ctl, int write, argument 1997 int *valp = ctl->data; 2000 int ret = proc_dointvec(ctl, writ 2036 ipv4_doint_and_flush(ctl_table *ctl, int write, void __user *buffer, size_t *lenp, loff_t *ppos) argument [all...] |
/net/bridge/ |
H A D | br_netfilter.c | 995 int brnf_sysctl_call_tables(ctl_table * ctl, int write, argument 1000 ret = proc_dointvec(ctl, write, buffer, lenp, ppos); 1002 if (write && *(int *)(ctl->data)) 1003 *(int *)(ctl->data) = 1;
|
/net/ipv6/ |
H A D | ndisc.c | 1605 static void ndisc_warn_deprecated_sysctl(struct ctl_table *ctl, argument 1614 dev_name, ctl->procname, 1615 dev_name, ctl->procname); 1620 int ndisc_ifinfo_sysctl_change(struct ctl_table *ctl, int write, void __user *buffer, size_t *lenp, loff_t *ppos) argument 1622 struct net_device *dev = ctl->extra1; 1626 if ((strcmp(ctl->procname, "retrans_time") == 0) || 1627 (strcmp(ctl->procname, "base_reachable_time") == 0)) 1628 ndisc_warn_deprecated_sysctl(ctl, "syscall", dev ? dev->name : "default"); 1630 if (strcmp(ctl->procname, "retrans_time") == 0) 1631 ret = proc_dointvec(ctl, writ [all...] |
H A D | route.c | 2790 int ipv6_sysctl_rtcache_flush(ctl_table *ctl, int write, argument 2798 net = (struct net *)ctl->extra1; 2800 proc_dointvec(ctl, write, buffer, lenp, ppos);
|
H A D | addrconf.c | 4672 int addrconf_sysctl_forward(ctl_table *ctl, int write, argument 4675 int *valp = ctl->data; 4682 * ctl->data points to idev->cnf.forwarding, we should 4685 lctl = *ctl; 4691 ret = addrconf_fixup_forwarding(ctl, valp, val); 4754 int addrconf_sysctl_disable(ctl_table *ctl, int write, argument 4757 int *valp = ctl->data; 4764 * ctl->data points to idev->cnf.disable_ipv6, we should 4767 lctl = *ctl; 4773 ret = addrconf_disable_ipv6(ctl, val [all...] |
/net/core/ |
H A D | neighbour.c | 2771 static int proc_unres_qlen(ctl_table *ctl, int write, void __user *buffer, argument 2775 ctl_table tmp = *ctl; 2781 size = *(int *)ctl->data / SKB_TRUESIZE(ETH_FRAME_LEN); 2785 *(int *)ctl->data = size * SKB_TRUESIZE(ETH_FRAME_LEN);
|
/net/ |
H A D | socket.c | 1967 unsigned char ctl[sizeof(struct cmsghdr) + 20] local 1970 unsigned char *ctl_buf = ctl; 2007 cmsghdr_from_user_compat_to_kern(msg_sys, sock->sk, ctl, 2008 sizeof(ctl)); 2014 if (ctl_len > sizeof(ctl)) { 2061 if (ctl_buf != ctl)
|