Searched defs:queue (Results 1 - 25 of 31) sorted by relevance

12

/net/core/
H A Drequest_sock.c24 * Maximum number of SYN_RECV sockets in queue per LISTEN socket.
40 int reqsk_queue_alloc(struct request_sock_queue *queue, argument
61 rwlock_init(&queue->syn_wait_lock);
62 queue->rskq_accept_head = NULL;
66 write_lock_bh(&queue->syn_wait_lock);
67 queue->listen_opt = lopt;
68 write_unlock_bh(&queue->syn_wait_lock);
73 void __reqsk_queue_destroy(struct request_sock_queue *queue) argument
76 kvfree(queue->listen_opt);
80 struct request_sock_queue *queue)
79 reqsk_queue_yank_listen_sk( struct request_sock_queue *queue) argument
92 reqsk_queue_destroy(struct request_sock_queue *queue) argument
[all...]
H A Ddatagram.c187 struct sk_buff_head *queue = &sk->sk_receive_queue; local
190 last = (struct sk_buff *)queue;
191 spin_lock_irqsave(&queue->lock, cpu_flags);
192 skb_queue_walk(queue, skb) {
204 __skb_unlink(skb, queue);
206 spin_unlock_irqrestore(&queue->lock, cpu_flags);
210 spin_unlock_irqrestore(&queue->lock, cpu_flags);
278 * receive queue of the socket, it will be taken off the queue
837 * sequenced packet sockets providing the socket receive queue
[all...]
H A Dnet-sysfs.c544 struct netdev_rx_queue *queue = to_rx_queue(kobj); local
549 return attribute->show(queue, attribute, buf);
556 struct netdev_rx_queue *queue = to_rx_queue(kobj); local
561 return attribute->store(queue, attribute, buf, count);
570 static ssize_t show_rps_map(struct netdev_rx_queue *queue, argument
582 map = rcu_dereference(queue->rps_map);
600 static ssize_t store_rps_map(struct netdev_rx_queue *queue, argument
641 old_map = rcu_dereference_protected(queue->rps_map,
643 rcu_assign_pointer(queue->rps_map, map);
656 static ssize_t show_rps_dev_flow_table_cnt(struct netdev_rx_queue *queue, argument
679 store_rps_dev_flow_table_cnt(struct netdev_rx_queue *queue, struct rx_queue_attribute *attr, const char *buf, size_t len) argument
757 struct netdev_rx_queue *queue = to_rx_queue(kobj); local
782 struct netdev_rx_queue *queue = to_rx_queue(kobj); local
801 struct netdev_rx_queue *queue = dev->_rx + index; local
879 struct netdev_queue *queue = to_netdev_queue(kobj); local
892 struct netdev_queue *queue = to_netdev_queue(kobj); local
905 show_trans_timeout(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, char *buf) argument
951 bql_show_hold_time(struct netdev_queue *queue, struct netdev_queue_attribute *attr, char *buf) argument
960 bql_set_hold_time(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, const char *buf, size_t len) argument
981 bql_show_inflight(struct netdev_queue *queue, struct netdev_queue_attribute *attr, char *buf) argument
1032 get_netdev_queue_index(struct netdev_queue *queue) argument
1044 show_xps_map(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, char *buf) argument
1089 store_xps_map(struct netdev_queue *queue, struct netdev_queue_attribute *attribute, const char *buf, size_t len) argument
1133 struct netdev_queue *queue = to_netdev_queue(kobj); local
1141 struct netdev_queue *queue = to_netdev_queue(kobj); local
1160 struct netdev_queue *queue = dev->_tx + index; local
1202 struct netdev_queue *queue = dev->_tx + i; local
[all...]
H A Dpktgen.c201 #define F_QUEUE_MAP_RND (1<<13) /* queue map Random */
202 #define F_QUEUE_MAP_CPU (1<<14) /* queue map mirrors smp_processor_id() */
245 struct list_head list; /* chaining in the thread's run-queue */
433 wait_queue_head_t queue; member in struct:pktgen_thread
2069 /* make sure that we don't pick a non-existing transmit queue */
3425 init_waitqueue_head(&t->queue);
3440 wait_event_interruptible_timeout(t->queue,
/net/netfilter/
H A Dxt_NFQUEUE.c1 /* iptables module for using new netfilter netlink queue
42 u32 queue = info->queuenum; local
45 queue = nfqueue_hash(skb, queue, info->queues_total,
48 return NF_QUEUE_NR(queue);
91 u32 queue = info->queuenum; local
98 queue = info->queuenum + cpu % info->queues_total;
100 queue = nfqueue_hash(skb, queue, info->queues_total,
105 ret = NF_QUEUE_NR(queue);
[all...]
H A Dnft_queue.c35 u32 queue = priv->queuenum; local
42 queue = priv->queuenum + cpu % priv->queues_total;
44 queue = nfqueue_hash(pkt->skb, queue,
50 ret = NF_QUEUE_NR(queue);
110 .name = "queue",
132 MODULE_ALIAS_NFT_EXPR("queue");
H A Dnfnetlink_queue_core.c64 u_int16_t queue_num; /* number of this queue */
74 struct list_head queue_list; /* packets in queue */
158 static void nfqnl_flush(struct nfqnl_instance *queue, nfqnl_cmpfn cmpfn,
188 __enqueue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument
190 list_add_tail(&entry->list, &queue->queue_list);
191 queue->queue_total++;
195 __dequeue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument
198 queue->queue_total--;
202 find_dequeue_entry(struct nfqnl_instance *queue, unsigned int id) argument
206 spin_lock_bh(&queue
224 nfqnl_flush(struct nfqnl_instance *queue, nfqnl_cmpfn cmpfn, unsigned long data) argument
282 nfqnl_build_packet_message(struct net *net, struct nfqnl_instance *queue, struct nf_queue_entry *entry, __be32 **packet_id_ptr) argument
509 __nfqnl_enqueue_packet(struct net *net, struct nfqnl_instance *queue, struct nf_queue_entry *entry) argument
600 __nfqnl_enqueue_packet_gso(struct net *net, struct nfqnl_instance *queue, struct sk_buff *skb, struct nf_queue_entry *entry) argument
633 struct nfqnl_instance *queue; local
727 nfqnl_set_mode(struct nfqnl_instance *queue, unsigned char mode, unsigned int range) argument
865 struct nfqnl_instance *queue; local
907 struct nfqnl_instance *queue; local
957 struct nfqnl_instance *queue; local
1035 struct nfqnl_instance *queue; local
[all...]
/net/sctp/
H A Dinqueue.c10 * An SCTP inqueue is a queue into which you push SCTP packets
47 void sctp_inq_init(struct sctp_inq *queue) argument
49 INIT_LIST_HEAD(&queue->in_chunk_list);
50 queue->in_progress = NULL;
53 INIT_WORK(&queue->immediate, NULL);
57 void sctp_inq_free(struct sctp_inq *queue) argument
61 /* Empty the queue. */
62 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) {
70 if (queue->in_progress) {
71 sctp_chunk_free(queue
99 sctp_inq_peek(struct sctp_inq *queue) argument
122 sctp_inq_pop(struct sctp_inq *queue) argument
[all...]
H A Dulpqueue.c55 /* Initialize a ULP queue from a block of memory. */
196 struct sk_buff_head *queue, *skb_list; local
221 queue = &sk->sk_receive_queue;
232 queue = &sctp_sk(sk)->pd_lobby;
235 queue = &sk->sk_receive_queue;
240 * can queue this to the receive queue instead
244 queue = &sk->sk_receive_queue;
246 queue = &sctp_sk(sk)->pd_lobby;
254 sctp_skb_list_tail(skb_list, queue);
326 sctp_make_reassembled_event(struct net *net, struct sk_buff_head *queue, struct sk_buff *f_frag, struct sk_buff *l_frag) argument
[all...]
/net/ipv4/
H A Dtcp_fastopen.c135 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; local
146 spin_lock(&queue->fastopenq->lock);
147 queue->fastopenq->qlen++;
148 spin_unlock(&queue->fastopenq->lock);
171 * because it's been added to the accept queue directly.
176 /* Add the child socket directly into the accept queue */
190 * (any reason not to?) but no need to queue the skb since
H A Dtcp_yeah.c19 #define TCP_YEAH_GAMMA 1 /* fraction of queue to be removed per rtt */
131 u32 rtt, queue; local
153 queue = bw;
155 if (queue > TCP_YEAH_ALPHA ||
157 if (queue > TCP_YEAH_ALPHA &&
159 u32 reduction = min(queue / TCP_YEAH_GAMMA ,
188 yeah->lastQ = queue;
H A Dinet_connection_sock.c261 * beginning of the wait-queue. As such, it's ok to "drop"
263 * having to remove and re-insert us on the wait queue.
295 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local
310 if (reqsk_queue_empty(queue)) {
322 req = reqsk_queue_remove(queue);
326 if (sk->sk_protocol == IPPROTO_TCP && queue->fastopenq != NULL) {
327 spin_lock_bh(&queue->fastopenq->lock);
338 spin_unlock_bh(&queue->fastopenq->lock);
567 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local
568 struct listen_sock *lopt = queue
785 struct request_sock_queue *queue = &icsk->icsk_accept_queue; local
[all...]
H A Dip_output.c479 * single device frame, and queue such a frame for sending.
731 * Put this fragment into the sending queue.
784 struct sk_buff_head *queue,
797 if ((skb = skb_peek_tail(queue)) == NULL) {
820 __skb_queue_tail(queue, skb);
837 struct sk_buff_head *queue,
860 skb = skb_peek_tail(queue);
894 err = ip_ufo_append_data(sk, queue, getfrag, from, length,
1019 * Put the packet on the pending queue.
1021 __skb_queue_tail(queue, sk
783 ip_ufo_append_data(struct sock *sk, struct sk_buff_head *queue, int getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb), void *from, int length, int hh_len, int fragheaderlen, int transhdrlen, int maxfraglen, unsigned int flags) argument
835 __ip_append_data(struct sock *sk, struct flowi4 *fl4, struct sk_buff_head *queue, struct inet_cork *cork, struct page_frag *pfrag, int getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb), void *from, int length, int transhdrlen, unsigned int flags) argument
1320 __ip_make_skb(struct sock *sk, struct flowi4 *fl4, struct sk_buff_head *queue, struct inet_cork *cork) argument
1441 __ip_flush_pending_frames(struct sock *sk, struct sk_buff_head *queue, struct inet_cork *cork) argument
1467 struct sk_buff_head queue; local
[all...]
/net/irda/
H A Dirlmp_frame.c418 * Function irlmp_find_handle (self, dlsap_sel, slsap_sel, status, queue)
436 hashbin_t *queue)
455 spin_lock_irqsave(&queue->hb_spinlock, flags);
457 lsap = (struct lsap_cb *) hashbin_get_first(queue);
480 lsap = (struct lsap_cb *) hashbin_get_next(queue);
486 spin_unlock_irqrestore(&queue->hb_spinlock, flags);
434 irlmp_find_lsap(struct lap_cb *self, __u8 dlsap_sel, __u8 slsap_sel, int status, hashbin_t *queue) argument
H A Dirqueue.c5 * Description: General queue implementation
228 * Function enqueue_first (queue, proc)
230 * Insert item first in queue.
233 static void enqueue_first(irda_queue_t **queue, irda_queue_t* element) argument
239 * Check if queue is empty.
241 if ( *queue == NULL ) {
243 * Queue is empty. Insert one element into the queue.
245 element->q_next = element->q_prev = *queue = element;
249 * Queue is not empty. Insert element into front of queue.
251 element->q_next = (*queue);
266 dequeue_first(irda_queue_t **queue) argument
307 dequeue_general(irda_queue_t **queue, irda_queue_t* element) argument
393 irda_queue_t* queue; local
[all...]
/net/decnet/
H A Ddn_nsp_in.c585 static __inline__ int dn_queue_skb(struct sock *sk, struct sk_buff *skb, int sig, struct sk_buff_head *queue) argument
603 skb_queue_tail(queue, skb);
H A Daf_decnet.c553 * By checking the other_data transmit queue is empty
1129 * the newly created socket on the wrong hash queue.
1677 struct sk_buff_head *queue = &sk->sk_receive_queue; local
1708 queue = &scp->other_receive_queue;
1741 if (dn_data_ready(sk, queue, flags, target))
1751 sk_wait_event(sk, &timeo, dn_data_ready(sk, queue, flags, target));
1756 skb_queue_walk_safe(queue, skb, n) {
1775 skb_unlink(skb, queue);
1823 static inline int dn_queue_too_long(struct dn_scp *scp, struct sk_buff_head *queue, int flags) argument
1826 if (skb_queue_len(queue) >
1914 struct sk_buff_head *queue = &scp->data_xmit_queue; local
[all...]
/net/mac80211/
H A Dagg-tx.c160 * refcount the global queue stop caused by that in order
171 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local
173 /* we do refcounting here, so don't use the queue reason refcounting */
175 if (atomic_inc_return(&sdata->local->agg_queue_stop[queue]) == 1)
177 &sdata->local->hw, queue,
186 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local
188 if (atomic_dec_return(&sdata->local->agg_queue_stop[queue]) == 0)
190 &sdata->local->hw, queue,
205 int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)]; local
211 "TID %d gone but expected when splicing aggregates from the pending queue\
[all...]
H A Dtkip.c245 u8 *ra, int only_iv, int queue,
268 if (key->u.tkip.rx[queue].state != TKIP_STATE_NOT_INIT &&
269 (iv32 < key->u.tkip.rx[queue].iv32 ||
270 (iv32 == key->u.tkip.rx[queue].iv32 &&
271 iv16 <= key->u.tkip.rx[queue].iv16)))
276 key->u.tkip.rx[queue].state = TKIP_STATE_PHASE1_HW_UPLOADED;
280 if (key->u.tkip.rx[queue].state == TKIP_STATE_NOT_INIT ||
281 key->u.tkip.rx[queue].iv32 != iv32) {
283 tkip_mixing_phase1(tk, &key->u.tkip.rx[queue], ta, iv32);
287 key->u.tkip.rx[queue]
242 ieee80211_tkip_decrypt_data(struct crypto_cipher *tfm, struct ieee80211_key *key, u8 *payload, size_t payload_len, u8 *ta, u8 *ra, int only_iv, int queue, u32 *out_iv32, u16 *out_iv16) argument
[all...]
H A Dwpa.c498 int queue; local
521 queue = rx->security_idx;
523 if (memcmp(pn, key->u.ccmp.rx_pn[queue], IEEE80211_CCMP_PN_LEN) <= 0) {
542 memcpy(key->u.ccmp.rx_pn[queue], pn, IEEE80211_CCMP_PN_LEN);
H A Drx.c1708 int queue = rx->security_idx; local
1713 rx->key->u.ccmp.rx_pn[queue],
1734 int queue; local
1743 queue = rx->security_idx;
1744 rpn = rx->key->u.ccmp.rx_pn[queue];
2537 goto queue;
2601 goto queue;
2644 goto queue;
2675 goto queue;
2693 goto queue;
[all...]
/net/atm/
H A Dcommon.c231 struct sk_buff_head queue, *rq; local
235 __skb_queue_head_init(&queue);
239 skb_queue_splice_init(rq, &queue);
242 skb_queue_walk_safe(&queue, skb, tmp) {
243 __skb_unlink(skb, &queue);
/net/nfc/
H A Ddigital_core.c33 struct list_head queue; member in struct:digital_cmd
125 queue);
131 list_del(&cmd->queue);
169 queue);
220 list_del(&cmd->queue);
247 INIT_LIST_HEAD(&cmd->queue);
250 list_add_tail(&cmd->queue, &ddev->cmd_queue);
837 list_for_each_entry_safe(cmd, n, &ddev->cmd_queue, queue) {
838 list_del(&cmd->queue);
/net/phonet/
H A Dpep.c349 struct sk_buff_head *queue; local
395 queue = &pn->ctrlreq_queue;
396 goto queue;
417 queue = &sk->sk_receive_queue;
418 goto queue;
462 queue:
465 skb_queue_tail(queue, skb);
664 * We then queue the skb to the right connected sock (if any).
/net/sunrpc/
H A Dsched.c61 * queue->lock and bh_disabled in order to avoid races within
65 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) argument
72 if (list_empty(&queue->timer_list.list))
73 del_timer(&queue->timer_list.timer);
77 rpc_set_queue_timer(struct rpc_wait_queue *queue, unsigned long expires) argument
79 queue->timer_list.expires = expires;
80 mod_timer(&queue->timer_list.timer, expires);
87 __rpc_add_timer(struct rpc_wait_queue *queue, struct rpc_task *task) argument
96 if (list_empty(&queue->timer_list.list) || time_before(task->u.tk_wait.expires, queue
101 rpc_rotate_queue_owner(struct rpc_wait_queue *queue) argument
113 rpc_set_waitqueue_priority(struct rpc_wait_queue *queue, int priority) argument
122 rpc_set_waitqueue_owner(struct rpc_wait_queue *queue, pid_t pid) argument
128 rpc_reset_waitqueue_priority(struct rpc_wait_queue *queue) argument
137 __rpc_add_wait_queue_priority(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument
167 __rpc_add_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument
209 __rpc_remove_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task) argument
220 __rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname, unsigned char nr_queues) argument
235 rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument
241 rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument
247 rpc_destroy_wait_queue(struct rpc_wait_queue *queue) argument
416 __rpc_do_wake_up_task(struct rpc_wait_queue *queue, struct rpc_task *task) argument
439 rpc_wake_up_task_queue_locked(struct rpc_wait_queue *queue, struct rpc_task *task) argument
451 rpc_wake_up_queued_task(struct rpc_wait_queue *queue, struct rpc_task *task) argument
462 __rpc_find_next_queued_priority(struct rpc_wait_queue *queue) argument
509 __rpc_find_next_queued(struct rpc_wait_queue *queue) argument
521 rpc_wake_up_first(struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument
550 rpc_wake_up_next(struct rpc_wait_queue *queue) argument
562 rpc_wake_up(struct rpc_wait_queue *queue) argument
591 rpc_wake_up_status(struct rpc_wait_queue *queue, int status) argument
616 struct rpc_wait_queue *queue = (struct rpc_wait_queue *)ptr; local
720 struct rpc_wait_queue *queue; local
[all...]

Completed in 260 milliseconds

12