/net/9p/ |
H A D | util.c | 38 * @lock: protects the pool 44 spinlock_t lock; member in struct:p9_idpool 61 spin_lock_init(&p->lock); 85 * the lock included in struct idr? 94 spin_lock_irqsave(&p->lock, flags); 99 spin_unlock_irqrestore(&p->lock, flags); 115 * the lock included in struct idr? 124 spin_lock_irqsave(&p->lock, flags); 126 spin_unlock_irqrestore(&p->lock, flags);
|
H A D | trans_virtio.c | 65 * @lock: protects multiple elements within this structure 80 spinlock_t lock; member in struct:virtio_chan 154 spin_lock_irqsave(&chan->lock, flags); 157 spin_unlock_irqrestore(&chan->lock, flags); 161 spin_unlock_irqrestore(&chan->lock, flags); 275 spin_lock_irqsave(&chan->lock, flags); 294 spin_unlock_irqrestore(&chan->lock, flags); 303 spin_unlock_irqrestore(&chan->lock, flags); 310 spin_unlock_irqrestore(&chan->lock, flags); 415 spin_lock_irqsave(&chan->lock, flag [all...] |
/net/netfilter/ |
H A D | xt_quota.c | 15 spinlock_t lock; member in struct:xt_quota_priv 32 spin_lock_bh(&priv->lock); 40 spin_unlock_bh(&priv->lock); 56 spin_lock_init(&q->master->lock);
|
H A D | nft_counter.c | 21 seqlock_t lock; member in struct:nft_counter 32 write_seqlock_bh(&priv->lock); 35 write_sequnlock_bh(&priv->lock); 46 seq = read_seqbegin(&priv->lock); 49 } while (read_seqretry(&priv->lock, seq)); 77 seqlock_init(&priv->lock);
|
H A D | xt_quota2.c | 29 * @lock: lock to protect quota writers from each other 33 spinlock_t lock; member in struct:xt_quota_counter 129 spin_lock_bh(&e->lock); 131 spin_unlock_bh(&e->lock); 147 spin_lock_bh(&e->lock); 149 spin_unlock_bh(&e->lock); 172 spin_lock_init(&e->lock); 195 /* No need to hold a lock while getting a new counter */ 218 * So release the lock [all...] |
/net/mac802154/ |
H A D | llsec.h | 51 spinlock_t lock; member in struct:mac802154_llsec_device 70 rwlock_t lock; member in struct:mac802154_llsec
|
/net/nfc/ |
H A D | nfc.h | 46 rwlock_t lock; member in struct:nfc_sock_list
|
/net/core/ |
H A D | gen_stats.c | 35 spin_unlock_bh(d->lock); 45 * @lock: statistics lock 48 * Initializes the dumping handle, grabs the statistic lock and appends 59 int xstats_type, spinlock_t *lock, struct gnet_dump *d) 60 __acquires(lock) 64 spin_lock_bh(lock); 65 d->lock = lock; 83 * @lock 93 gnet_stats_start_copy(struct sk_buff *skb, int type, spinlock_t *lock, struct gnet_dump *d) argument [all...] |
H A D | drop_monitor.c | 48 spinlock_t lock; member in struct:per_cpu_dm_data 101 spin_lock_irqsave(&data->lock, flags); 103 spin_unlock_irqrestore(&data->lock, flags); 150 spin_lock(&data->lock); 182 spin_unlock_irqrestore(&data->lock, flags); 391 spin_lock_init(&data->lock);
|
H A D | dst.c | 39 * We want to keep lock & list close together 44 spinlock_t lock; member in struct:__anon28 49 .lock = __SPIN_LOCK_UNLOCKED(dst_garbage.lock), 106 spin_lock_bh(&dst_garbage.lock); 110 spin_unlock_bh(&dst_garbage.lock); 141 spin_unlock_bh(&dst_garbage.lock); 221 spin_lock_bh(&dst_garbage.lock); 231 spin_unlock_bh(&dst_garbage.lock); 396 spin_lock_bh(&dst_garbage.lock); [all...] |
/net/dccp/ |
H A D | probe.c | 48 spinlock_t lock; member in struct:__anon36 71 kfifo_in_locked(&dccpw.fifo, tbuf, len, &dccpw.lock); 139 cnt = kfifo_out_locked(&dccpw.fifo, tbuf, len, &dccpw.lock); 160 spin_lock_init(&dccpw.lock);
|
/net/sctp/ |
H A D | probe.c | 66 spinlock_t lock; member in struct:__anon162 81 kfifo_in_locked(&sctpw.fifo, tbuf, len, &sctpw.lock); 114 cnt = kfifo_out_locked(&sctpw.fifo, tbuf, len, &sctpw.lock); 212 spin_lock_init(&sctpw.lock);
|
/net/tipc/ |
H A D | subscr.c | 44 * @lock: control access to subscriber 49 spinlock_t lock; member in struct:tipc_subscriber 130 * Protected by nameseq.lock in name_table.c 148 /* The spin lock per subscriber is used to protect its members */ 149 spin_lock_bh(&subscriber->lock); 153 spin_unlock_bh(&subscriber->lock); 163 spin_unlock_bh(&subscriber->lock); 178 * Called with subscriber lock held. 203 spin_lock_bh(&subscriber->lock); 209 spin_unlock_bh(&subscriber->lock); [all...] |
H A D | discover.c | 53 * @lock: spinlock for controlling access to requests 63 spinlock_t lock; member in struct:tipc_link_req 279 spin_lock_bh(&req->lock); 281 spin_unlock_bh(&req->lock); 290 spin_lock_bh(&req->lock); 293 spin_unlock_bh(&req->lock); 306 spin_lock_bh(&req->lock); 334 spin_unlock_bh(&req->lock); 365 spin_lock_init(&req->lock); 394 spin_lock_bh(&req->lock); [all...] |
H A D | node.h | 96 * @lock: spinlock governing access to structure 112 spinlock_t lock; member in struct:tipc_node 150 spin_lock_bh(&node->lock);
|
/net/wireless/ |
H A D | lib80211.c | 50 spinlock_t *lock) 55 info->lock = lock; 93 spin_lock_irqsave(info->lock, flags); 106 spin_unlock_irqrestore(info->lock, flags); 114 spin_lock_irqsave(info->lock, flags); 116 spin_unlock_irqrestore(info->lock, flags); 126 spin_lock_irqsave(info->lock, flags); 133 spin_unlock_irqrestore(info->lock, flags); 152 spin_lock_irqsave(info->lock, flag 49 lib80211_crypt_info_init(struct lib80211_crypt_info *info, char *name, spinlock_t *lock) argument [all...] |
/net/caif/ |
H A D | caif_dev.c | 47 struct mutex lock; member in struct:caif_device_entry_list 199 /* Hold lock while accessing xoff */ 334 mutex_lock(&caifdevs->lock); 348 mutex_unlock(&caifdevs->lock); 448 mutex_lock(&caifdevs->lock); 452 mutex_unlock(&caifdevs->lock); 475 mutex_unlock(&caifdevs->lock); 484 mutex_unlock(&caifdevs->lock); 500 mutex_init(&caifn->caifdevs.lock); 517 mutex_lock(&caifdevs->lock); [all...] |
H A D | cfcnfg.c | 60 struct mutex lock; member in struct:cfcnfg 105 mutex_init(&this->lock); 470 mutex_lock(&cnfg->lock); 518 mutex_unlock(&cnfg->lock); 523 mutex_unlock(&cnfg->lock); 567 mutex_lock(&cnfg->lock); 573 mutex_unlock(&cnfg->lock); 588 mutex_unlock(&cnfg->lock); 607 mutex_unlock(&cnfg->lock);
|
H A D | cfpkt_skbuff.c | 29 spinlock_t lock; member in struct:cfpktq
|
/net/ipv4/ |
H A D | inet_timewait_sock.c | 25 * ehash lock must be held by caller. 26 * Returns 1 if caller should call inet_twsk_put() after lock release. 36 * We cannot call inet_twsk_put() ourself under lock, 48 * bind hash lock must be held by caller. 49 * Returns 1 if caller should call inet_twsk_put() after lock release. 63 * We cannot call inet_twsk_put() ourself under lock, 76 spinlock_t *lock = inet_ehash_lockp(hashinfo, tw->tw_hash); local 78 spin_lock(lock); 80 spin_unlock(lock); 86 spin_lock(&bhead->lock); 136 spinlock_t *lock = inet_ehash_lockp(hashinfo, sk->sk_hash); local [all...] |
H A D | tcp_probe.c | 77 spinlock_t lock; member in struct:__anon55 120 spin_lock(&tcp_probe.lock); 160 spin_unlock(&tcp_probe.lock); 178 spin_lock_bh(&tcp_probe.lock); 181 spin_unlock_bh(&tcp_probe.lock); 220 spin_lock_bh(&tcp_probe.lock); 223 spin_unlock_bh(&tcp_probe.lock); 232 spin_unlock_bh(&tcp_probe.lock); 266 spin_lock_init(&tcp_probe.lock);
|
H A D | inet_hashtables.c | 76 * Caller must hold hashbucket lock for this tb with local BH disabled 113 spin_lock(&head->lock); 120 spin_unlock(&head->lock); 140 spin_lock(&head->lock); 157 spin_unlock(&head->lock); 163 spin_unlock(&head->lock); 336 spinlock_t *lock = inet_ehash_lockp(hinfo, hash); local 342 spin_lock(lock); 371 spin_unlock(lock); 387 spin_unlock(lock); 403 spinlock_t *lock; local 458 spinlock_t *lock; local [all...] |
/net/ipv6/ |
H A D | inet6_hashtables.c | 70 spin_lock(&ilb->lock); 72 spin_unlock(&ilb->lock); 76 spinlock_t *lock; local 80 lock = inet_ehash_lockp(hashinfo, hash); 81 spin_lock(lock); 87 spin_unlock(lock); 99 * The sockhash lock must be held as a reader here. 256 spinlock_t *lock = inet_ehash_lockp(hinfo, hash); local 262 spin_lock(lock); 290 spin_unlock(lock); [all...] |
H A D | mip6.c | 119 spinlock_t lock; member in struct:mip6_report_rate_limiter 127 .lock = __SPIN_LOCK_UNLOCKED(mip6_report_rl.lock) 136 spin_lock(&x->lock); 140 spin_unlock(&x->lock); 176 spin_lock_bh(&x->lock); 178 spin_unlock_bh(&x->lock); 192 spin_lock_bh(&mip6_report_rl.lock); 205 spin_unlock_bh(&mip6_report_rl.lock); 358 spin_lock(&x->lock); [all...] |
/net/phonet/ |
H A D | pn_dev.c | 38 struct mutex lock; member in struct:phonet_routes 72 BUG_ON(!mutex_is_locked(&pndevs->lock)); 82 BUG_ON(!mutex_is_locked(&pndevs->lock)); 109 mutex_lock(&pndevs->lock); 113 mutex_unlock(&pndevs->lock); 152 mutex_lock(&pndevs->lock); 161 mutex_unlock(&pndevs->lock); 171 mutex_lock(&pndevs->lock); 180 mutex_unlock(&pndevs->lock); 276 mutex_lock(&pnn->routes.lock); [all...] |