Searched defs:lock (Results 1 - 25 of 187) sorted by relevance

12345678

/include/net/
H A Dgen_stats.h10 spinlock_t * lock; member in struct:gnet_dump
23 spinlock_t *lock, struct gnet_dump *d);
27 spinlock_t *lock, struct gnet_dump *d);
H A Dping.h33 rwlock_t lock; member in struct:ping_table
H A Draw.h34 rwlock_t lock; member in struct:raw_hashinfo
/include/net/netfilter/
H A Dxt_rateest.h5 /* keep lock and bstats on same cache line to speedup xt_rateest_tg() */
7 spinlock_t lock; member in struct:xt_rateest
8 /* keep rstats and lock on same cache line to speedup xt_rateest_mt() */
/include/net/phonet/
H A Dpn_dev.h28 struct mutex lock; member in struct:phonet_device_list
/include/drm/ttm/
H A Dttm_lock.h33 * of the DRM heavyweight hardware lock.
34 * The lock is a read-write lock. Taking it in read mode and write mode
39 * It's allowed to leave kernel space with the vt lock held.
40 * If a user-space process dies while having the vt-lock,
41 * it will be released during the file descriptor release. The vt lock
42 * excludes write lock and read lock.
44 * The suspend mode is used to lock out all TTM users when preparing for
59 * @base: ttm base object used solely to release the lock i
72 spinlock_t lock; member in struct:ttm_lock
239 ttm_lock_set_kill(struct ttm_lock *lock, bool val, int signal) argument
[all...]
H A Dttm_memory.h64 * @lock: Lock to protect the @shrink - and the memory accounting members,
84 spinlock_t lock; member in struct:ttm_mem_global
121 spin_lock(&glob->lock);
123 spin_unlock(&glob->lock);
127 spin_unlock(&glob->lock);
142 spin_lock(&glob->lock);
145 spin_unlock(&glob->lock);
/include/linux/
H A Dblockgroup_lock.h37 spinlock_t lock; member in struct:bgl_lock
49 spin_lock_init(&bgl->locks[i].lock);
59 return &bgl->locks[(block_group) & (NR_BG_LOCKS-1)].lock;
H A Dfs_struct.h10 spinlock_t lock; member in struct:fs_struct
29 spin_lock(&fs->lock);
32 spin_unlock(&fs->lock);
37 spin_lock(&fs->lock);
40 spin_unlock(&fs->lock);
46 spin_lock(&fs->lock);
51 spin_unlock(&fs->lock);
H A Dsemaphore.h17 raw_spinlock_t lock; member in struct:semaphore
24 .lock = __RAW_SPIN_LOCK_UNLOCKED((name).lock), \
36 lockdep_init_map(&sem->lock.dep_map, "semaphore->lock", &__key, 0);
H A Dwakelock.h36 static inline void wake_lock_init(struct wake_lock *lock, int type, argument
39 wakeup_source_init(&lock->ws, name);
42 static inline void wake_lock_destroy(struct wake_lock *lock) argument
44 wakeup_source_trash(&lock->ws);
47 static inline void wake_lock(struct wake_lock *lock) argument
49 __pm_stay_awake(&lock->ws);
52 static inline void wake_lock_timeout(struct wake_lock *lock, long timeout) argument
54 __pm_wakeup_event(&lock->ws, jiffies_to_msecs(timeout));
57 static inline void wake_unlock(struct wake_lock *lock) argument
59 __pm_relax(&lock
62 wake_lock_active(struct wake_lock *lock) argument
[all...]
H A Dbasic_mmio_gpio.h49 * Used to lock bgpio_chip->data. Also, this is needed to keep
52 spinlock_t lock; member in struct:bgpio_chip
H A Dgenalloc.h15 * available. If new memory is added to the pool a lock has to be
36 spinlock_t lock; member in struct:gen_pool
H A Dmempool.h15 spinlock_t lock; member in struct:mempool_s
H A Dratelimit.h11 raw_spinlock_t lock; /* protect the state */ member in struct:ratelimit_state
23 .lock = __RAW_SPIN_LOCK_UNLOCKED(name.lock), \
31 raw_spin_lock_init(&rs->lock);
H A Dspinlock_up.h25 static inline void arch_spin_lock(arch_spinlock_t *lock) argument
27 lock->slock = 0;
31 arch_spin_lock_flags(arch_spinlock_t *lock, unsigned long flags) argument
34 lock->slock = 0;
37 static inline int arch_spin_trylock(arch_spinlock_t *lock) argument
39 char oldval = lock->slock;
41 lock->slock = 0;
46 static inline void arch_spin_unlock(arch_spinlock_t *lock) argument
48 lock->slock = 1;
54 #define arch_read_lock(lock) d
[all...]
H A Dconcap.h34 spinlock_t lock; member in struct:concap_proto
H A Dcyclomx.h48 spinlock_t lock; member in struct:cycx_device
63 spinlock_t lock; member in struct:cycx_device::__anon287::__anon288
H A Diocontext.h50 * queue lock but the returned icq is valid only until the queue lock is
57 * - ioc lock nests inside q lock.
59 * - ioc->icq_list and icq->ioc_node are protected by ioc lock.
60 * q->icq_list and icq->q_node by q lock.
62 * - ioc->icq_tree and ioc->icq_hint are protected by ioc lock, while icq
63 * itself is protected by q lock. However, both the indexes and icq
65 * the q lock.
72 * locks. Due to the lock orderin
106 spinlock_t lock; member in struct:io_context
[all...]
H A Dkthread.h57 spinlock_t lock; member in struct:kthread_worker
71 .lock = __SPIN_LOCK_UNLOCKED((worker).lock), \
88 * kthread_worker.lock and kthread_work.done need their own lockdep class
112 __init_kthread_worker((worker), "("#worker")->lock", &__key); \
/include/linux/spi/
H A Dmax7301.h11 struct mutex lock; member in struct:max7301
H A Dspi_bitbang.h10 spinlock_t lock; member in struct:spi_bitbang
/include/media/
H A Dmedia-device.h46 * @lock: Entities list lock
47 * @graph_mutex: Entities graph operation lock
74 spinlock_t lock; member in struct:media_device
/include/sound/
H A Dseq_midi_event.h38 spinlock_t lock; member in struct:snd_midi_event
/include/drm/
H A Ddrm_sarea.h70 struct drm_hw_lock lock; member in struct:drm_sarea
71 /** \todo Use readers/writer lock for drm_sarea::drawable_lock */

Completed in 1113 milliseconds

12345678