/net/ax25/ |
H A D | ax25_ds_in.c | 150 int queued = 0; local 243 queued = ax25_rx_iframe(ax25, skb); 276 return queued; 284 int queued = 0, frametype, ns, nr, pf; local 290 queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type); 293 queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type); 296 queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type); 300 return queued;
|
H A D | ax25_std_in.c | 146 int queued = 0; local 228 queued = ax25_rx_iframe(ax25, skb); 261 return queued; 271 int queued = 0; local 383 queued = ax25_rx_iframe(ax25, skb); 416 return queued; 424 int queued = 0, frametype, ns, nr, pf; local 430 queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type); 433 queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type); 436 queued [all...] |
H A D | ax25_in.c | 107 int queued = 0; local 149 queued = 1; 155 return queued; 163 int queued = 0; local 171 queued = ax25_std_frame_in(ax25, skb, type); 177 queued = ax25_ds_frame_in(ax25, skb, type); 179 queued = ax25_std_frame_in(ax25, skb, type); 184 return queued; 309 * Process the frame. If it is queued up internally it
|
/net/lapb/ |
H A D | lapb_in.c | 254 int queued = 0; local 408 queued = 1; 470 if (!queued)
|
/net/rose/ |
H A D | rose_in.c | 107 int queued = 0; local 169 queued = 1; 206 return queued; 266 int queued = 0, frametype, ns, nr, q, d, m; local 275 queued = rose_state1_machine(sk, skb, frametype); 278 queued = rose_state2_machine(sk, skb, frametype); 281 queued = rose_state3_machine(sk, skb, frametype, ns, nr, q, d, m); 284 queued = rose_state4_machine(sk, skb, frametype); 287 queued = rose_state5_machine(sk, skb, frametype); 293 return queued; [all...] |
/net/netrom/ |
H A D | nr_in.c | 156 int queued = 0; local 229 queued = 1; 276 return queued; 283 int queued = 0, frametype; local 292 queued = nr_state1_machine(sk, skb, frametype); 295 queued = nr_state2_machine(sk, skb, frametype); 298 queued = nr_state3_machine(sk, skb, frametype); 304 return queued;
|
/net/x25/ |
H A D | x25_dev.c | 56 int queued = 1; local 61 queued = x25_process_rx_frame(sk, skb); 63 queued = !sk_add_backlog(sk, skb, sk->sk_rcvbuf); 67 return queued;
|
H A D | x25_in.c | 206 int queued = 0; local 273 queued = 1; 311 queued = !sock_queue_rcv_skb(sk, skb); 315 queued = 1; 326 return queued; 384 int queued = 0, frametype, ns, nr, q, d, m; local 393 queued = x25_state1_machine(sk, skb, frametype); 396 queued = x25_state2_machine(sk, skb, frametype); 399 queued = x25_state3_machine(sk, skb, frametype, ns, nr, q, d, m); 402 queued 413 int queued = x25_process_rx_frame(sk, skb); local [all...] |
/net/dccp/ |
H A D | input.c | 49 int queued = 0; local 80 queued = 1; 90 return queued; 95 int queued = 0; local 105 return queued; 117 queued = 1; 124 return queued; 528 int queued = 0; local 565 queued = 1; /* packet was queued 580 int queued = 0; local [all...] |
/net/decnet/ |
H A D | dn_nsp_in.c | 222 * the incoming data is in the correct format before it is queued to 583 * also allows data and other data to be queued to a socket. 616 int queued = 0; local 629 queued = 1; 635 if (!queued) 641 int queued = 0; local 655 queued = 1; 666 if (!queued) 824 * sock_release() when there is a backlog queued up.
|
/net/rds/ |
H A D | send.c | 130 * - queued acks can be delayed behind large messages 132 * - small message latency is higher behind queued large messages 147 * sendmsg calls here after having queued its message on the send 376 * not try and send their newly queued message. We need to check the 638 * Transports call here when they've determined that the receiver queued 750 * we only want this to fire once so we use the callers 'queued'. It's 756 __be16 dport, int *queued) 761 if (*queued) 806 rdsdebug("queued msg %p len %d, rs %p bytes %d seq %llu\n", 810 *queued 754 rds_send_queue_rm(struct rds_sock *rs, struct rds_connection *conn, struct rds_message *rm, __be16 sport, __be16 dport, int *queued) argument 934 int queued = 0, allocated_mr = 0; local [all...] |
/net/netfilter/ |
H A D | nfnetlink_queue_core.c | 68 * Following fields are dirtied for each queued packet, 632 unsigned int queued; local 670 queued = 0; 678 queued++; 684 if (queued) { 685 if (err) /* some segments are already queued */
|
/net/mac80211/ |
H A D | cfg.c | 2431 bool queued = false; local 2522 queued = true; 2546 queued = true; 2551 queued = true; 2558 * queued/pending and thus we queue the new ROC 2564 queued = true; 2580 if (!queued)
|
H A D | tx.c | 498 * been queued to pending queue. No reordering can happen, go 530 * We queued up some frames, so the TIM bit might 1052 bool queued = false; local 1092 queued = true; 1110 return queued; 1165 bool queued; local 1167 queued = ieee80211_tx_prep_agg(tx, skb, info, 1170 if (unlikely(queued)) 1272 * Returns false if the frame couldn't be transmitted but was queued instead. 1339 * frame was dropped or queued [all...] |
/net/ipv4/ |
H A D | tcp_input.c | 5588 int queued = 0; local 5634 queued = tcp_rcv_synsent_state_process(sk, skb, th, len); 5635 if (queued >= 0) 5636 return queued; 5834 queued = 1; 5844 if (!queued) { 5927 /* Accept backlog is full. If we have already queued enough
|