/net/core/ |
H A D | request_sock.c | 24 * Maximum number of SYN_RECV sockets in queue per LISTEN socket. 40 int reqsk_queue_alloc(struct request_sock_queue *queue, argument 61 rwlock_init(&queue->syn_wait_lock); 62 queue->rskq_accept_head = NULL; 66 write_lock_bh(&queue->syn_wait_lock); 67 queue->listen_opt = lopt; 68 write_unlock_bh(&queue->syn_wait_lock); 73 void __reqsk_queue_destroy(struct request_sock_queue *queue) argument 76 kvfree(queue->listen_opt); 80 struct request_sock_queue *queue) 79 reqsk_queue_yank_listen_sk( struct request_sock_queue *queue) argument 92 reqsk_queue_destroy(struct request_sock_queue *queue) argument [all...] |
H A D | net-sysfs.c | 544 struct netdev_rx_queue *queue = to_rx_queue(kobj); local 549 return attribute->show(queue, attribute, buf); 556 struct netdev_rx_queue *queue = to_rx_queue(kobj); local 561 return attribute->store(queue, attribute, buf, count); 570 static ssize_t show_rps_map(struct netdev_rx_queue *queue, argument 582 map = rcu_dereference(queue->rps_map); 600 static ssize_t store_rps_map(struct netdev_rx_queue *queue, argument 641 old_map = rcu_dereference_protected(queue->rps_map, 643 rcu_assign_pointer(queue->rps_map, map); 656 static ssize_t show_rps_dev_flow_table_cnt(struct netdev_rx_queue *queue, argument 679 store_rps_dev_flow_table_cnt(struct netdev_rx_queue *queue, struct rx_queue_attribute *attr, const char *buf, size_t len) argument 757 struct netdev_rx_queue *queue = to_rx_queue(kobj); local 782 struct netdev_rx_queue *queue = to_rx_queue(kobj); local 801 struct netdev_rx_queue *queue = dev->_rx + index; local 879 struct netdev_queue *queue = to_netdev_queue(kobj); local 892 struct netdev_queue *queue = to_netdev_queue(kobj); local 905 show_trans_timeout(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, char *buf) argument 951 bql_show_hold_time(struct netdev_queue *queue, struct netdev_queue_attribute *attr, char *buf) argument 960 bql_set_hold_time(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, const char *buf, size_t len) argument 981 bql_show_inflight(struct netdev_queue *queue, struct netdev_queue_attribute *attr, char *buf) argument 1032 get_netdev_queue_index(struct netdev_queue *queue) argument 1044 show_xps_map(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, char *buf) argument 1089 store_xps_map(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, const char *buf, size_t len) argument 1133 struct netdev_queue *queue = to_netdev_queue(kobj); local 1141 struct netdev_queue *queue = to_netdev_queue(kobj); local 1160 struct netdev_queue *queue = dev->_tx + index; local 1202 struct netdev_queue *queue = dev->_tx + i; local [all...] |
H A D | datagram.c | 187 struct sk_buff_head *queue = &sk->sk_receive_queue; local 190 last = (struct sk_buff *)queue; 191 spin_lock_irqsave(&queue->lock, cpu_flags); 192 skb_queue_walk(queue, skb) { 204 __skb_unlink(skb, queue); 206 spin_unlock_irqrestore(&queue->lock, cpu_flags); 210 spin_unlock_irqrestore(&queue->lock, cpu_flags); 278 * receive queue of the socket, it will be taken off the queue 837 * sequenced packet sockets providing the socket receive queue [all...] |
/net/sunrpc/ |
H A D | sched.c | 61 * queue->lock and bh_disabled in order to avoid races within 65 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) argument 72 if (list_empty(&queue->timer_list.list)) 73 del_timer(&queue->timer_list.timer); 77 rpc_set_queue_timer(struct rpc_wait_queue *queue, unsigned long expires) argument 79 queue->timer_list.expires = expires; 80 mod_timer(&queue->timer_list.timer, expires); 87 __rpc_add_timer(struct rpc_wait_queue *queue, struct rpc_task *task) argument 96 if (list_empty(&queue->timer_list.list) || time_before(task->u.tk_wait.expires, queue 101 rpc_rotate_queue_owner(struct rpc_wait_queue *queue) argument 113 rpc_set_waitqueue_priority(struct rpc_wait_queue *queue, int priority) argument 122 rpc_set_waitqueue_owner(struct rpc_wait_queue *queue, pid_t pid) argument 128 rpc_reset_waitqueue_priority(struct rpc_wait_queue *queue) argument 137 __rpc_add_wait_queue_priority(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument 167 __rpc_add_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument 209 __rpc_remove_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task) argument 220 __rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname, unsigned char nr_queues) argument 235 rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument 241 rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument 247 rpc_destroy_wait_queue(struct rpc_wait_queue *queue) argument 416 __rpc_do_wake_up_task(struct rpc_wait_queue *queue, struct rpc_task *task) argument 439 rpc_wake_up_task_queue_locked(struct rpc_wait_queue *queue, struct rpc_task *task) argument 451 rpc_wake_up_queued_task(struct rpc_wait_queue *queue, struct rpc_task *task) argument 462 __rpc_find_next_queued_priority(struct rpc_wait_queue *queue) argument 509 __rpc_find_next_queued(struct rpc_wait_queue *queue) argument 521 rpc_wake_up_first(struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument 550 rpc_wake_up_next(struct rpc_wait_queue *queue) argument 562 rpc_wake_up(struct rpc_wait_queue *queue) argument 591 rpc_wake_up_status(struct rpc_wait_queue *queue, int status) argument 616 struct rpc_wait_queue *queue = (struct rpc_wait_queue *)ptr; local 720 struct rpc_wait_queue *queue; local [all...] |
/net/sctp/ |
H A D | inqueue.c | 10 * An SCTP inqueue is a queue into which you push SCTP packets 47 void sctp_inq_init(struct sctp_inq *queue) argument 49 INIT_LIST_HEAD(&queue->in_chunk_list); 50 queue->in_progress = NULL; 53 INIT_WORK(&queue->immediate, NULL); 57 void sctp_inq_free(struct sctp_inq *queue) argument 61 /* Empty the queue. */ 62 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) { 70 if (queue->in_progress) { 71 sctp_chunk_free(queue 99 sctp_inq_peek(struct sctp_inq *queue) argument 122 sctp_inq_pop(struct sctp_inq *queue) argument [all...] |
H A D | ulpqueue.c | 55 /* Initialize a ULP queue from a block of memory. */ 196 struct sk_buff_head *queue, *skb_list; local 221 queue = &sk->sk_receive_queue; 232 queue = &sctp_sk(sk)->pd_lobby; 235 queue = &sk->sk_receive_queue; 240 * can queue this to the receive queue instead 244 queue = &sk->sk_receive_queue; 246 queue = &sctp_sk(sk)->pd_lobby; 254 sctp_skb_list_tail(skb_list, queue); 326 sctp_make_reassembled_event(struct net *net, struct sk_buff_head *queue, struct sk_buff *f_frag, struct sk_buff *l_frag) argument [all...] |
/net/netfilter/ |
H A D | nfnetlink_queue_core.c | 64 u_int16_t queue_num; /* number of this queue */ 74 struct list_head queue_list; /* packets in queue */ 158 static void nfqnl_flush(struct nfqnl_instance *queue, nfqnl_cmpfn cmpfn, 188 __enqueue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument 190 list_add_tail(&entry->list, &queue->queue_list); 191 queue->queue_total++; 195 __dequeue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument 198 queue->queue_total--; 202 find_dequeue_entry(struct nfqnl_instance *queue, unsigned int id) argument 206 spin_lock_bh(&queue 224 nfqnl_flush(struct nfqnl_instance *queue, nfqnl_cmpfn cmpfn, unsigned long data) argument 282 nfqnl_build_packet_message(struct net *net, struct nfqnl_instance *queue, struct nf_queue_entry *entry, __be32 **packet_id_ptr) argument 509 __nfqnl_enqueue_packet(struct net *net, struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument 600 __nfqnl_enqueue_packet_gso(struct net *net, struct nfqnl_instance *queue, struct sk_buff *skb, struct nf_queue_entry *entry) argument 633 struct nfqnl_instance *queue; local 727 nfqnl_set_mode(struct nfqnl_instance *queue, unsigned char mode, unsigned int range) argument 865 struct nfqnl_instance *queue; local 907 struct nfqnl_instance *queue; local 957 struct nfqnl_instance *queue; local 1035 struct nfqnl_instance *queue; local [all...] |
H A D | xt_NFQUEUE.c | 1 /* iptables module for using new netfilter netlink queue 42 u32 queue = info->queuenum; local 45 queue = nfqueue_hash(skb, queue, info->queues_total, 48 return NF_QUEUE_NR(queue); 91 u32 queue = info->queuenum; local 98 queue = info->queuenum + cpu % info->queues_total; 100 queue = nfqueue_hash(skb, queue, info->queues_total, 105 ret = NF_QUEUE_NR(queue); [all...] |
H A D | nft_queue.c | 35 u32 queue = priv->queuenum; local 42 queue = priv->queuenum + cpu % priv->queues_total; 44 queue = nfqueue_hash(pkt->skb, queue, 50 ret = NF_QUEUE_NR(queue); 110 .name = "queue", 132 MODULE_ALIAS_NFT_EXPR("queue");
|
/net/irda/ |
H A D | irqueue.c | 5 * Description: General queue implementation 228 * Function enqueue_first (queue, proc) 230 * Insert item first in queue. 233 static void enqueue_first(irda_queue_t **queue, irda_queue_t* element) argument 239 * Check if queue is empty. 241 if ( *queue == NULL ) { 243 * Queue is empty. Insert one element into the queue. 245 element->q_next = element->q_prev = *queue = element; 249 * Queue is not empty. Insert element into front of queue. 251 element->q_next = (*queue); 266 dequeue_first(irda_queue_t **queue) argument 307 dequeue_general(irda_queue_t **queue, irda_queue_t* element) argument 393 irda_queue_t* queue; local [all...] |
H A D | irlmp_frame.c | 418 * Function irlmp_find_handle (self, dlsap_sel, slsap_sel, status, queue) 436 hashbin_t *queue) 455 spin_lock_irqsave(&queue->hb_spinlock, flags); 457 lsap = (struct lsap_cb *) hashbin_get_first(queue); 480 lsap = (struct lsap_cb *) hashbin_get_next(queue); 486 spin_unlock_irqrestore(&queue->hb_spinlock, flags); 434 irlmp_find_lsap(struct lap_cb *self, __u8 dlsap_sel, __u8 slsap_sel, int status, hashbin_t *queue) argument
|
/net/mac80211/ |
H A D | tkip.h | 32 u8 *ra, int only_iv, int queue,
|
H A D | tkip.c | 245 u8 *ra, int only_iv, int queue, 268 if (key->u.tkip.rx[queue].state != TKIP_STATE_NOT_INIT && 269 (iv32 < key->u.tkip.rx[queue].iv32 || 270 (iv32 == key->u.tkip.rx[queue].iv32 && 271 iv16 <= key->u.tkip.rx[queue].iv16))) 276 key->u.tkip.rx[queue].state = TKIP_STATE_PHASE1_HW_UPLOADED; 280 if (key->u.tkip.rx[queue].state == TKIP_STATE_NOT_INIT || 281 key->u.tkip.rx[queue].iv32 != iv32) { 283 tkip_mixing_phase1(tk, &key->u.tkip.rx[queue], ta, iv32); 287 key->u.tkip.rx[queue] 242 ieee80211_tkip_decrypt_data(struct crypto_cipher *tfm, struct ieee80211_key *key, u8 *payload, size_t payload_len, u8 *ta, u8 *ra, int only_iv, int queue, u32 *out_iv32, u16 *out_iv16) argument [all...] |
H A D | util.c | 290 void ieee80211_propagate_queue_wake(struct ieee80211_local *local, int queue) argument 311 if (ac_queue == queue || 312 (sdata->vif.cab_queue == queue && 320 static void __ieee80211_wake_queue(struct ieee80211_hw *hw, int queue, argument 326 trace_wake_queue(local, queue, reason); 328 if (WARN_ON(queue >= hw->queues)) 331 if (!test_bit(reason, &local->queue_stop_reasons[queue])) 335 local->q_stop_reasons[queue][reason] = 0; 337 local->q_stop_reasons[queue][reason]--; 339 if (local->q_stop_reasons[queue][reaso 354 ieee80211_wake_queue_by_reason(struct ieee80211_hw *hw, int queue, enum queue_stop_reason reason, bool refcounted) argument 366 ieee80211_wake_queue(struct ieee80211_hw *hw, int queue) argument 374 __ieee80211_stop_queue(struct ieee80211_hw *hw, int queue, enum queue_stop_reason reason, bool refcounted) argument 414 ieee80211_stop_queue_by_reason(struct ieee80211_hw *hw, int queue, enum queue_stop_reason reason, bool refcounted) argument 426 ieee80211_stop_queue(struct ieee80211_hw *hw, int queue) argument 440 int queue = info->hw_queue; local 462 int queue, i; local 514 ieee80211_queue_stopped(struct ieee80211_hw *hw, int queue) argument [all...] |
H A D | agg-tx.c | 160 * refcount the global queue stop caused by that in order 171 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local 173 /* we do refcounting here, so don't use the queue reason refcounting */ 175 if (atomic_inc_return(&sdata->local->agg_queue_stop[queue]) == 1) 177 &sdata->local->hw, queue, 186 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local 188 if (atomic_dec_return(&sdata->local->agg_queue_stop[queue]) == 0) 190 &sdata->local->hw, queue, 205 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local 211 "TID %d gone but expected when splicing aggregates from the pending queue\ [all...] |
H A D | trace.h | 1995 TP_PROTO(struct ieee80211_local *local, u16 queue, 1998 TP_ARGS(local, queue, reason), 2002 __field(u16, queue) 2008 __entry->queue = queue; 2013 LOCAL_PR_FMT " queue:%d, reason:%d", 2014 LOCAL_PR_ARG, __entry->queue, __entry->reason 2019 TP_PROTO(struct ieee80211_local *local, u16 queue, 2022 TP_ARGS(local, queue, reason), 2026 __field(u16, queue) [all...] |
/net/ipv4/ |
H A D | tcp_yeah.c | 19 #define TCP_YEAH_GAMMA 1 /* fraction of queue to be removed per rtt */ 131 u32 rtt, queue; local 153 queue = bw; 155 if (queue > TCP_YEAH_ALPHA || 157 if (queue > TCP_YEAH_ALPHA && 159 u32 reduction = min(queue / TCP_YEAH_GAMMA , 188 yeah->lastQ = queue;
|
H A D | inet_connection_sock.c | 261 * beginning of the wait-queue. As such, it's ok to "drop" 263 * having to remove and re-insert us on the wait queue. 295 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local 310 if (reqsk_queue_empty(queue)) { 322 req = reqsk_queue_remove(queue); 326 if (sk->sk_protocol == IPPROTO_TCP && queue->fastopenq != NULL) { 327 spin_lock_bh(&queue->fastopenq->lock); 338 spin_unlock_bh(&queue->fastopenq->lock); 567 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local 568 struct listen_sock *lopt = queue 785 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local [all...] |
H A D | tcp_fastopen.c | 135 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; local 146 spin_lock(&queue->fastopenq->lock); 147 queue->fastopenq->qlen++; 148 spin_unlock(&queue->fastopenq->lock); 171 * because it's been added to the accept queue directly. 176 /* Add the child socket directly into the accept queue */ 190 * (any reason not to?) but no need to queue the skb since
|
H A D | ip_output.c | 479 * single device frame, and queue such a frame for sending. 731 * Put this fragment into the sending queue. 784 struct sk_buff_head *queue, 797 if ((skb = skb_peek_tail(queue)) == NULL) { 820 __skb_queue_tail(queue, skb); 837 struct sk_buff_head *queue, 860 skb = skb_peek_tail(queue); 894 err = ip_ufo_append_data(sk, queue, getfrag, from, length, 1019 * Put the packet on the pending queue. 1021 __skb_queue_tail(queue, sk 783 ip_ufo_append_data(struct sock *sk, struct sk_buff_head *queue, int getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb), void *from, int length, int hh_len, int fragheaderlen, int transhdrlen, int maxfraglen, unsigned int flags) argument 835 __ip_append_data(struct sock *sk, struct flowi4 *fl4, struct sk_buff_head *queue, struct inet_cork *cork, struct page_frag *pfrag, int getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb), void *from, int length, int transhdrlen, unsigned int flags) argument 1320 __ip_make_skb(struct sock *sk, struct flowi4 *fl4, struct sk_buff_head *queue, struct inet_cork *cork) argument 1441 __ip_flush_pending_frames(struct sock *sk, struct sk_buff_head *queue, struct inet_cork *cork) argument 1467 struct sk_buff_head queue; local [all...] |
/net/x25/ |
H A D | x25_link.c | 110 while ((skbn = skb_dequeue(&nb->queue)) != NULL) 201 skb_queue_tail(&nb->queue, skb); 207 skb_queue_tail(&nb->queue, skb); 254 skb_queue_head_init(&nb->queue); 285 skb_queue_purge(&nb->queue);
|
/net/nfc/ |
H A D | digital_core.c | 33 struct list_head queue; member in struct:digital_cmd 125 queue); 131 list_del(&cmd->queue); 169 queue); 220 list_del(&cmd->queue); 247 INIT_LIST_HEAD(&cmd->queue); 250 list_add_tail(&cmd->queue, &ddev->cmd_queue); 837 list_for_each_entry_safe(cmd, n, &ddev->cmd_queue, queue) { 838 list_del(&cmd->queue);
|
/net/802/ |
H A D | garp.c | 253 skb_queue_tail(&app->queue, app->pdu); 261 while ((skb = skb_dequeue(&app->queue))) 294 goto queue; 296 goto queue; 301 goto queue; 308 queue: 585 skb_queue_head_init(&app->queue);
|
H A D | mrp.c | 345 skb_queue_tail(&app->queue, app->pdu); 353 while ((skb = skb_dequeue(&app->queue))) 416 goto queue; 426 goto queue; 437 goto queue; 466 queue: 867 skb_queue_head_init(&app->queue);
|
/net/rose/ |
H A D | rose_link.c | 173 while ((skbn = skb_dequeue(&neigh->queue)) != NULL) 284 skb_queue_tail(&neigh->queue, skb);
|