Searched defs:lock (Results 51 - 58 of 58) sorted by relevance

123

/net/ipv4/
H A Dtcp.c446 * Note that we don't need to lock the socket, as the upper poll layers
1494 * splitted a fat GRO packet, while we released socket lock
1548 /* If recv_actor drops the lock (e.g. TCP splice
2140 /* Now socket is owned by kernel and we acquire BH lock
3111 spin_lock_init(&tcp_hashinfo.bhash[i].lock);
3187 spinlock_t *lock = inet_ehash_lockp(&tcp_hashinfo, bucket); local
3190 spin_lock_bh(lock);
3237 spin_unlock_bh(lock);
3251 spin_unlock_bh(lock);
H A Dtcp_ipv4.c217 * lock select source port, enter ourselves into the hash tables and
903 /* caller either holds rcu_read_lock() or socket lock */
1422 * We have a potential double-lock case here, so even when
1872 spin_lock_bh(&ilb->lock);
1930 spin_unlock_bh(&ilb->lock);
1934 spin_lock_bh(&ilb->lock);
1978 spinlock_t *lock = inet_ehash_lockp(&tcp_hashinfo, st->bucket); local
1984 spin_lock_bh(lock);
1993 spin_unlock_bh(lock);
2151 spin_unlock_bh(&tcp_hashinfo.listening_hash[st->bucket].lock);
[all...]
/net/mac80211/
H A Dsta_info.h131 * The TX path can access it under RCU lock-free if, and
215 * tid_start_tx is also protected by sta->lock.
275 * @lock: used for locking all fields that require locking, see comments
354 spinlock_t lock; member in struct:sta_info
369 /* STA powersave lock and frame queues */
412 * Aggregation information, locked with lock.
533 lockdep_is_held(&sta->lock) ||
553 * Get a STA info, must be under RCU read lock.
/net/sched/
H A Dsch_hfsc.c1017 spinlock_t *lock = qdisc_root_sleeping_lock(sch); local
1021 lock,
H A Dsch_htb.c969 /* always caled under BH & queue lock */
1069 /* Its safe to not acquire qdisc lock. As we hold RTNL,
1100 /* Its safe to not acquire qdisc lock. As we hold RTNL,
1469 spinlock_t *lock = qdisc_root_sleeping_lock(sch); local
1473 lock,
1548 * another way to "lock" the class - unlike "get" this lock can
/net/batman-adv/
H A Dtypes.h131 * @lock: lock to protect the list of fragments
138 spinlock_t lock; /* protects head */ member in struct:batadv_frag_table_entry
187 * @ogm_cnt_lock: lock protecting bcast_own, bcast_own_sum,
219 * @tt_buff_lock: lock that protects tt_buff and tt_buff_len
229 * @neigh_list_lock: lock protecting neigh_list and router
232 * @bcast_seqno_lock: lock protecting bcast_bits & last_bcast_seqno
242 * @vlan_list_lock: lock protecting vlan_list
333 * @ifinfo_lock: lock protecting private ifinfo members and list
502 * @changes_list_lock: lock protectin
575 spinlock_t lock; /* protects log_buff, log_start and log_end */ member in struct:batadv_priv_debug_log
841 spinlock_t lock; /* protects queue_list, queue_len & index */ member in struct:batadv_socket_client
[all...]
/net/rxrpc/
H A Dar-internal.h68 rwlock_t call_lock; /* lock for calls */
160 spinlock_t lock; /* access lock */ member in struct:rxrpc_local
161 rwlock_t services_lock; /* lock for services list */
177 spinlock_t lock; /* access lock */ member in struct:rxrpc_peer
210 spinlock_t client_lock; /* client connection allocation lock */
211 rwlock_t conn_lock; /* lock for active/dead connections */
258 rwlock_t lock; /* access lock */ member in struct:rxrpc_connection
346 spinlock_t lock; member in struct:rxrpc_call
[all...]
/net/core/
H A Ddev.c30 * Alan Cox : Fixed double lock.
48 * Alan Cox : Device lock protection.
53 * Dave Miller : 32bit quantity for the device lock to
206 spin_lock(&sd->input_pkt_queue.lock);
213 spin_unlock(&sd->input_pkt_queue.lock);
318 static inline void netdev_set_xmit_lockdep_class(spinlock_t *lock, argument
324 lockdep_set_class_and_name(lock, &netdev_xmit_lock_key[i],
338 static inline void netdev_set_xmit_lockdep_class(spinlock_t *lock, argument
693 * careful with locks. The caller must hold RCU lock.
767 * about locking. The caller must hold RCU lock
[all...]

Completed in 1502 milliseconds

123