Searched defs:lock (Results 51 - 58 of 58) sorted by relevance
123
/net/ipv4/ |
H A D | tcp.c | 446 * Note that we don't need to lock the socket, as the upper poll layers 1494 * splitted a fat GRO packet, while we released socket lock 1548 /* If recv_actor drops the lock (e.g. TCP splice 2140 /* Now socket is owned by kernel and we acquire BH lock 3111 spin_lock_init(&tcp_hashinfo.bhash[i].lock); 3187 spinlock_t *lock = inet_ehash_lockp(&tcp_hashinfo, bucket); local 3190 spin_lock_bh(lock); 3237 spin_unlock_bh(lock); 3251 spin_unlock_bh(lock);
|
H A D | tcp_ipv4.c | 217 * lock select source port, enter ourselves into the hash tables and 903 /* caller either holds rcu_read_lock() or socket lock */ 1422 * We have a potential double-lock case here, so even when 1872 spin_lock_bh(&ilb->lock); 1930 spin_unlock_bh(&ilb->lock); 1934 spin_lock_bh(&ilb->lock); 1978 spinlock_t *lock = inet_ehash_lockp(&tcp_hashinfo, st->bucket); local 1984 spin_lock_bh(lock); 1993 spin_unlock_bh(lock); 2151 spin_unlock_bh(&tcp_hashinfo.listening_hash[st->bucket].lock); [all...] |
/net/mac80211/ |
H A D | sta_info.h | 131 * The TX path can access it under RCU lock-free if, and 215 * tid_start_tx is also protected by sta->lock. 275 * @lock: used for locking all fields that require locking, see comments 354 spinlock_t lock; member in struct:sta_info 369 /* STA powersave lock and frame queues */ 412 * Aggregation information, locked with lock. 533 lockdep_is_held(&sta->lock) || 553 * Get a STA info, must be under RCU read lock.
|
/net/sched/ |
H A D | sch_hfsc.c | 1017 spinlock_t *lock = qdisc_root_sleeping_lock(sch); local 1021 lock,
|
H A D | sch_htb.c | 969 /* always caled under BH & queue lock */ 1069 /* Its safe to not acquire qdisc lock. As we hold RTNL, 1100 /* Its safe to not acquire qdisc lock. As we hold RTNL, 1469 spinlock_t *lock = qdisc_root_sleeping_lock(sch); local 1473 lock, 1548 * another way to "lock" the class - unlike "get" this lock can
|
/net/batman-adv/ |
H A D | types.h | 131 * @lock: lock to protect the list of fragments 138 spinlock_t lock; /* protects head */ member in struct:batadv_frag_table_entry 187 * @ogm_cnt_lock: lock protecting bcast_own, bcast_own_sum, 219 * @tt_buff_lock: lock that protects tt_buff and tt_buff_len 229 * @neigh_list_lock: lock protecting neigh_list and router 232 * @bcast_seqno_lock: lock protecting bcast_bits & last_bcast_seqno 242 * @vlan_list_lock: lock protecting vlan_list 333 * @ifinfo_lock: lock protecting private ifinfo members and list 502 * @changes_list_lock: lock protectin 575 spinlock_t lock; /* protects log_buff, log_start and log_end */ member in struct:batadv_priv_debug_log 841 spinlock_t lock; /* protects queue_list, queue_len & index */ member in struct:batadv_socket_client [all...] |
/net/rxrpc/ |
H A D | ar-internal.h | 68 rwlock_t call_lock; /* lock for calls */ 160 spinlock_t lock; /* access lock */ member in struct:rxrpc_local 161 rwlock_t services_lock; /* lock for services list */ 177 spinlock_t lock; /* access lock */ member in struct:rxrpc_peer 210 spinlock_t client_lock; /* client connection allocation lock */ 211 rwlock_t conn_lock; /* lock for active/dead connections */ 258 rwlock_t lock; /* access lock */ member in struct:rxrpc_connection 346 spinlock_t lock; member in struct:rxrpc_call [all...] |
/net/core/ |
H A D | dev.c | 30 * Alan Cox : Fixed double lock. 48 * Alan Cox : Device lock protection. 53 * Dave Miller : 32bit quantity for the device lock to 206 spin_lock(&sd->input_pkt_queue.lock); 213 spin_unlock(&sd->input_pkt_queue.lock); 318 static inline void netdev_set_xmit_lockdep_class(spinlock_t *lock, argument 324 lockdep_set_class_and_name(lock, &netdev_xmit_lock_key[i], 338 static inline void netdev_set_xmit_lockdep_class(spinlock_t *lock, argument 693 * careful with locks. The caller must hold RCU lock. 767 * about locking. The caller must hold RCU lock [all...] |
Completed in 1502 milliseconds
123