/net/core/ |
H A D | request_sock.c | 24 * Maximum number of SYN_RECV sockets in queue per LISTEN socket. 40 int reqsk_queue_alloc(struct request_sock_queue *queue, argument 61 rwlock_init(&queue->syn_wait_lock); 62 queue->rskq_accept_head = NULL; 66 write_lock_bh(&queue->syn_wait_lock); 67 queue->listen_opt = lopt; 68 write_unlock_bh(&queue->syn_wait_lock); 73 void __reqsk_queue_destroy(struct request_sock_queue *queue) argument 76 kvfree(queue->listen_opt); 80 struct request_sock_queue *queue) 79 reqsk_queue_yank_listen_sk( struct request_sock_queue *queue) argument 92 reqsk_queue_destroy(struct request_sock_queue *queue) argument [all...] |
H A D | datagram.c | 187 struct sk_buff_head *queue = &sk->sk_receive_queue; local 190 last = (struct sk_buff *)queue; 191 spin_lock_irqsave(&queue->lock, cpu_flags); 192 skb_queue_walk(queue, skb) { 204 __skb_unlink(skb, queue); 206 spin_unlock_irqrestore(&queue->lock, cpu_flags); 210 spin_unlock_irqrestore(&queue->lock, cpu_flags); 278 * receive queue of the socket, it will be taken off the queue 837 * sequenced packet sockets providing the socket receive queue [all...] |
H A D | net-sysfs.c | 544 struct netdev_rx_queue *queue = to_rx_queue(kobj); local 549 return attribute->show(queue, attribute, buf); 556 struct netdev_rx_queue *queue = to_rx_queue(kobj); local 561 return attribute->store(queue, attribute, buf, count); 570 static ssize_t show_rps_map(struct netdev_rx_queue *queue, argument 582 map = rcu_dereference(queue->rps_map); 600 static ssize_t store_rps_map(struct netdev_rx_queue *queue, argument 641 old_map = rcu_dereference_protected(queue->rps_map, 643 rcu_assign_pointer(queue->rps_map, map); 656 static ssize_t show_rps_dev_flow_table_cnt(struct netdev_rx_queue *queue, argument 679 store_rps_dev_flow_table_cnt(struct netdev_rx_queue *queue, struct rx_queue_attribute *attr, const char *buf, size_t len) argument 757 struct netdev_rx_queue *queue = to_rx_queue(kobj); local 782 struct netdev_rx_queue *queue = to_rx_queue(kobj); local 801 struct netdev_rx_queue *queue = dev->_rx + index; local 879 struct netdev_queue *queue = to_netdev_queue(kobj); local 892 struct netdev_queue *queue = to_netdev_queue(kobj); local 905 show_trans_timeout(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, char *buf) argument 951 bql_show_hold_time(struct netdev_queue *queue, struct netdev_queue_attribute *attr, char *buf) argument 960 bql_set_hold_time(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, const char *buf, size_t len) argument 981 bql_show_inflight(struct netdev_queue *queue, struct netdev_queue_attribute *attr, char *buf) argument 1032 get_netdev_queue_index(struct netdev_queue *queue) argument 1044 show_xps_map(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, char *buf) argument 1089 store_xps_map(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, const char *buf, size_t len) argument 1133 struct netdev_queue *queue = to_netdev_queue(kobj); local 1141 struct netdev_queue *queue = to_netdev_queue(kobj); local 1160 struct netdev_queue *queue = dev->_tx + index; local 1202 struct netdev_queue *queue = dev->_tx + i; local [all...] |
H A D | pktgen.c | 201 #define F_QUEUE_MAP_RND (1<<13) /* queue map Random */ 202 #define F_QUEUE_MAP_CPU (1<<14) /* queue map mirrors smp_processor_id() */ 245 struct list_head list; /* chaining in the thread's run-queue */ 433 wait_queue_head_t queue; member in struct:pktgen_thread 2069 /* make sure that we don't pick a non-existing transmit queue */ 3425 init_waitqueue_head(&t->queue); 3440 wait_event_interruptible_timeout(t->queue,
|
/net/netfilter/ |
H A D | xt_NFQUEUE.c | 1 /* iptables module for using new netfilter netlink queue 42 u32 queue = info->queuenum; local 45 queue = nfqueue_hash(skb, queue, info->queues_total, 48 return NF_QUEUE_NR(queue); 91 u32 queue = info->queuenum; local 98 queue = info->queuenum + cpu % info->queues_total; 100 queue = nfqueue_hash(skb, queue, info->queues_total, 105 ret = NF_QUEUE_NR(queue); [all...] |
H A D | nft_queue.c | 35 u32 queue = priv->queuenum; local 42 queue = priv->queuenum + cpu % priv->queues_total; 44 queue = nfqueue_hash(pkt->skb, queue, 50 ret = NF_QUEUE_NR(queue); 110 .name = "queue", 132 MODULE_ALIAS_NFT_EXPR("queue");
|
H A D | nfnetlink_queue_core.c | 64 u_int16_t queue_num; /* number of this queue */ 74 struct list_head queue_list; /* packets in queue */ 158 static void nfqnl_flush(struct nfqnl_instance *queue, nfqnl_cmpfn cmpfn, 188 __enqueue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument 190 list_add_tail(&entry->list, &queue->queue_list); 191 queue->queue_total++; 195 __dequeue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument 198 queue->queue_total--; 202 find_dequeue_entry(struct nfqnl_instance *queue, unsigned int id) argument 206 spin_lock_bh(&queue 224 nfqnl_flush(struct nfqnl_instance *queue, nfqnl_cmpfn cmpfn, unsigned long data) argument 282 nfqnl_build_packet_message(struct net *net, struct nfqnl_instance *queue, struct nf_queue_entry *entry, __be32 **packet_id_ptr) argument 509 __nfqnl_enqueue_packet(struct net *net, struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument 600 __nfqnl_enqueue_packet_gso(struct net *net, struct nfqnl_instance *queue, struct sk_buff *skb, struct nf_queue_entry *entry) argument 633 struct nfqnl_instance *queue; local 727 nfqnl_set_mode(struct nfqnl_instance *queue, unsigned char mode, unsigned int range) argument 865 struct nfqnl_instance *queue; local 907 struct nfqnl_instance *queue; local 957 struct nfqnl_instance *queue; local 1035 struct nfqnl_instance *queue; local [all...] |
/net/sctp/ |
H A D | inqueue.c | 10 * An SCTP inqueue is a queue into which you push SCTP packets 47 void sctp_inq_init(struct sctp_inq *queue) argument 49 INIT_LIST_HEAD(&queue->in_chunk_list); 50 queue->in_progress = NULL; 53 INIT_WORK(&queue->immediate, NULL); 57 void sctp_inq_free(struct sctp_inq *queue) argument 61 /* Empty the queue. */ 62 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) { 70 if (queue->in_progress) { 71 sctp_chunk_free(queue 99 sctp_inq_peek(struct sctp_inq *queue) argument 122 sctp_inq_pop(struct sctp_inq *queue) argument [all...] |
H A D | ulpqueue.c | 55 /* Initialize a ULP queue from a block of memory. */ 196 struct sk_buff_head *queue, *skb_list; local 221 queue = &sk->sk_receive_queue; 232 queue = &sctp_sk(sk)->pd_lobby; 235 queue = &sk->sk_receive_queue; 240 * can queue this to the receive queue instead 244 queue = &sk->sk_receive_queue; 246 queue = &sctp_sk(sk)->pd_lobby; 254 sctp_skb_list_tail(skb_list, queue); 326 sctp_make_reassembled_event(struct net *net, struct sk_buff_head *queue, struct sk_buff *f_frag, struct sk_buff *l_frag) argument [all...] |
/net/ipv4/ |
H A D | tcp_fastopen.c | 135 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; local 146 spin_lock(&queue->fastopenq->lock); 147 queue->fastopenq->qlen++; 148 spin_unlock(&queue->fastopenq->lock); 171 * because it's been added to the accept queue directly. 176 /* Add the child socket directly into the accept queue */ 190 * (any reason not to?) but no need to queue the skb since
|
H A D | tcp_yeah.c | 19 #define TCP_YEAH_GAMMA 1 /* fraction of queue to be removed per rtt */ 131 u32 rtt, queue; local 153 queue = bw; 155 if (queue > TCP_YEAH_ALPHA || 157 if (queue > TCP_YEAH_ALPHA && 159 u32 reduction = min(queue / TCP_YEAH_GAMMA , 188 yeah->lastQ = queue;
|
H A D | inet_connection_sock.c | 261 * beginning of the wait-queue. As such, it's ok to "drop" 263 * having to remove and re-insert us on the wait queue. 295 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local 310 if (reqsk_queue_empty(queue)) { 322 req = reqsk_queue_remove(queue); 326 if (sk->sk_protocol == IPPROTO_TCP && queue->fastopenq != NULL) { 327 spin_lock_bh(&queue->fastopenq->lock); 338 spin_unlock_bh(&queue->fastopenq->lock); 567 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local 568 struct listen_sock *lopt = queue 785 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local [all...] |
H A D | ip_output.c | 479 * single device frame, and queue such a frame for sending. 731 * Put this fragment into the sending queue. 784 struct sk_buff_head *queue, 797 if ((skb = skb_peek_tail(queue)) == NULL) { 820 __skb_queue_tail(queue, skb); 837 struct sk_buff_head *queue, 860 skb = skb_peek_tail(queue); 894 err = ip_ufo_append_data(sk, queue, getfrag, from, length, 1019 * Put the packet on the pending queue. 1021 __skb_queue_tail(queue, sk 783 ip_ufo_append_data(struct sock *sk, struct sk_buff_head *queue, int getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb), void *from, int length, int hh_len, int fragheaderlen, int transhdrlen, int maxfraglen, unsigned int flags) argument 835 __ip_append_data(struct sock *sk, struct flowi4 *fl4, struct sk_buff_head *queue, struct inet_cork *cork, struct page_frag *pfrag, int getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb), void *from, int length, int transhdrlen, unsigned int flags) argument 1320 __ip_make_skb(struct sock *sk, struct flowi4 *fl4, struct sk_buff_head *queue, struct inet_cork *cork) argument 1441 __ip_flush_pending_frames(struct sock *sk, struct sk_buff_head *queue, struct inet_cork *cork) argument 1467 struct sk_buff_head queue; local [all...] |
/net/irda/ |
H A D | irlmp_frame.c | 418 * Function irlmp_find_handle (self, dlsap_sel, slsap_sel, status, queue) 436 hashbin_t *queue) 455 spin_lock_irqsave(&queue->hb_spinlock, flags); 457 lsap = (struct lsap_cb *) hashbin_get_first(queue); 480 lsap = (struct lsap_cb *) hashbin_get_next(queue); 486 spin_unlock_irqrestore(&queue->hb_spinlock, flags); 434 irlmp_find_lsap(struct lap_cb *self, __u8 dlsap_sel, __u8 slsap_sel, int status, hashbin_t *queue) argument
|
H A D | irqueue.c | 5 * Description: General queue implementation 228 * Function enqueue_first (queue, proc) 230 * Insert item first in queue. 233 static void enqueue_first(irda_queue_t **queue, irda_queue_t* element) argument 239 * Check if queue is empty. 241 if ( *queue == NULL ) { 243 * Queue is empty. Insert one element into the queue. 245 element->q_next = element->q_prev = *queue = element; 249 * Queue is not empty. Insert element into front of queue. 251 element->q_next = (*queue); 266 dequeue_first(irda_queue_t **queue) argument 307 dequeue_general(irda_queue_t **queue, irda_queue_t* element) argument 393 irda_queue_t* queue; local [all...] |
/net/decnet/ |
H A D | dn_nsp_in.c | 585 static __inline__ int dn_queue_skb(struct sock *sk, struct sk_buff *skb, int sig, struct sk_buff_head *queue) argument 603 skb_queue_tail(queue, skb);
|
H A D | af_decnet.c | 553 * By checking the other_data transmit queue is empty 1129 * the newly created socket on the wrong hash queue. 1677 struct sk_buff_head *queue = &sk->sk_receive_queue; local 1708 queue = &scp->other_receive_queue; 1741 if (dn_data_ready(sk, queue, flags, target)) 1751 sk_wait_event(sk, &timeo, dn_data_ready(sk, queue, flags, target)); 1756 skb_queue_walk_safe(queue, skb, n) { 1775 skb_unlink(skb, queue); 1823 static inline int dn_queue_too_long(struct dn_scp *scp, struct sk_buff_head *queue, int flags) argument 1826 if (skb_queue_len(queue) > 1914 struct sk_buff_head *queue = &scp->data_xmit_queue; local [all...] |
/net/mac80211/ |
H A D | agg-tx.c | 160 * refcount the global queue stop caused by that in order 171 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local 173 /* we do refcounting here, so don't use the queue reason refcounting */ 175 if (atomic_inc_return(&sdata->local->agg_queue_stop[queue]) == 1) 177 &sdata->local->hw, queue, 186 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local 188 if (atomic_dec_return(&sdata->local->agg_queue_stop[queue]) == 0) 190 &sdata->local->hw, queue, 205 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local 211 "TID %d gone but expected when splicing aggregates from the pending queue\ [all...] |
H A D | tkip.c | 245 u8 *ra, int only_iv, int queue, 268 if (key->u.tkip.rx[queue].state != TKIP_STATE_NOT_INIT && 269 (iv32 < key->u.tkip.rx[queue].iv32 || 270 (iv32 == key->u.tkip.rx[queue].iv32 && 271 iv16 <= key->u.tkip.rx[queue].iv16))) 276 key->u.tkip.rx[queue].state = TKIP_STATE_PHASE1_HW_UPLOADED; 280 if (key->u.tkip.rx[queue].state == TKIP_STATE_NOT_INIT || 281 key->u.tkip.rx[queue].iv32 != iv32) { 283 tkip_mixing_phase1(tk, &key->u.tkip.rx[queue], ta, iv32); 287 key->u.tkip.rx[queue] 242 ieee80211_tkip_decrypt_data(struct crypto_cipher *tfm, struct ieee80211_key *key, u8 *payload, size_t payload_len, u8 *ta, u8 *ra, int only_iv, int queue, u32 *out_iv32, u16 *out_iv16) argument [all...] |
H A D | wpa.c | 498 int queue; local 521 queue = rx->security_idx; 523 if (memcmp(pn, key->u.ccmp.rx_pn[queue], IEEE80211_CCMP_PN_LEN) <= 0) { 542 memcpy(key->u.ccmp.rx_pn[queue], pn, IEEE80211_CCMP_PN_LEN);
|
H A D | rx.c | 1708 int queue = rx->security_idx; local 1713 rx->key->u.ccmp.rx_pn[queue], 1734 int queue; local 1743 queue = rx->security_idx; 1744 rpn = rx->key->u.ccmp.rx_pn[queue]; 2537 goto queue; 2601 goto queue; 2644 goto queue; 2675 goto queue; 2693 goto queue; [all...] |
/net/atm/ |
H A D | common.c | 231 struct sk_buff_head queue, *rq; local 235 __skb_queue_head_init(&queue); 239 skb_queue_splice_init(rq, &queue); 242 skb_queue_walk_safe(&queue, skb, tmp) { 243 __skb_unlink(skb, &queue);
|
/net/nfc/ |
H A D | digital_core.c | 33 struct list_head queue; member in struct:digital_cmd 125 queue); 131 list_del(&cmd->queue); 169 queue); 220 list_del(&cmd->queue); 247 INIT_LIST_HEAD(&cmd->queue); 250 list_add_tail(&cmd->queue, &ddev->cmd_queue); 837 list_for_each_entry_safe(cmd, n, &ddev->cmd_queue, queue) { 838 list_del(&cmd->queue);
|
/net/phonet/ |
H A D | pep.c | 349 struct sk_buff_head *queue; local 395 queue = &pn->ctrlreq_queue; 396 goto queue; 417 queue = &sk->sk_receive_queue; 418 goto queue; 462 queue: 465 skb_queue_tail(queue, skb); 664 * We then queue the skb to the right connected sock (if any).
|
/net/sunrpc/ |
H A D | sched.c | 61 * queue->lock and bh_disabled in order to avoid races within 65 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) argument 72 if (list_empty(&queue->timer_list.list)) 73 del_timer(&queue->timer_list.timer); 77 rpc_set_queue_timer(struct rpc_wait_queue *queue, unsigned long expires) argument 79 queue->timer_list.expires = expires; 80 mod_timer(&queue->timer_list.timer, expires); 87 __rpc_add_timer(struct rpc_wait_queue *queue, struct rpc_task *task) argument 96 if (list_empty(&queue->timer_list.list) || time_before(task->u.tk_wait.expires, queue 101 rpc_rotate_queue_owner(struct rpc_wait_queue *queue) argument 113 rpc_set_waitqueue_priority(struct rpc_wait_queue *queue, int priority) argument 122 rpc_set_waitqueue_owner(struct rpc_wait_queue *queue, pid_t pid) argument 128 rpc_reset_waitqueue_priority(struct rpc_wait_queue *queue) argument 137 __rpc_add_wait_queue_priority(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument 167 __rpc_add_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument 209 __rpc_remove_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task) argument 220 __rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname, unsigned char nr_queues) argument 235 rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument 241 rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument 247 rpc_destroy_wait_queue(struct rpc_wait_queue *queue) argument 416 __rpc_do_wake_up_task(struct rpc_wait_queue *queue, struct rpc_task *task) argument 439 rpc_wake_up_task_queue_locked(struct rpc_wait_queue *queue, struct rpc_task *task) argument 451 rpc_wake_up_queued_task(struct rpc_wait_queue *queue, struct rpc_task *task) argument 462 __rpc_find_next_queued_priority(struct rpc_wait_queue *queue) argument 509 __rpc_find_next_queued(struct rpc_wait_queue *queue) argument 521 rpc_wake_up_first(struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument 550 rpc_wake_up_next(struct rpc_wait_queue *queue) argument 562 rpc_wake_up(struct rpc_wait_queue *queue) argument 591 rpc_wake_up_status(struct rpc_wait_queue *queue, int status) argument 616 struct rpc_wait_queue *queue = (struct rpc_wait_queue *)ptr; local 720 struct rpc_wait_queue *queue; local [all...] |