/include/net/ |
H A D | gen_stats.h | 10 spinlock_t * lock; member in struct:gnet_dump 23 spinlock_t *lock, struct gnet_dump *d); 27 spinlock_t *lock, struct gnet_dump *d);
|
H A D | ping.h | 33 rwlock_t lock; member in struct:ping_table
|
H A D | raw.h | 34 rwlock_t lock; member in struct:raw_hashinfo
|
/include/net/netfilter/ |
H A D | xt_rateest.h | 5 /* keep lock and bstats on same cache line to speedup xt_rateest_tg() */ 7 spinlock_t lock; member in struct:xt_rateest 8 /* keep rstats and lock on same cache line to speedup xt_rateest_mt() */
|
/include/net/phonet/ |
H A D | pn_dev.h | 28 struct mutex lock; member in struct:phonet_device_list
|
/include/drm/ttm/ |
H A D | ttm_lock.h | 33 * of the DRM heavyweight hardware lock. 34 * The lock is a read-write lock. Taking it in read mode and write mode 39 * It's allowed to leave kernel space with the vt lock held. 40 * If a user-space process dies while having the vt-lock, 41 * it will be released during the file descriptor release. The vt lock 42 * excludes write lock and read lock. 44 * The suspend mode is used to lock out all TTM users when preparing for 59 * @base: ttm base object used solely to release the lock i 72 spinlock_t lock; member in struct:ttm_lock 239 ttm_lock_set_kill(struct ttm_lock *lock, bool val, int signal) argument [all...] |
H A D | ttm_memory.h | 64 * @lock: Lock to protect the @shrink - and the memory accounting members, 84 spinlock_t lock; member in struct:ttm_mem_global 121 spin_lock(&glob->lock); 123 spin_unlock(&glob->lock); 127 spin_unlock(&glob->lock); 142 spin_lock(&glob->lock); 145 spin_unlock(&glob->lock);
|
/include/linux/ |
H A D | blockgroup_lock.h | 37 spinlock_t lock; member in struct:bgl_lock 49 spin_lock_init(&bgl->locks[i].lock); 59 return &bgl->locks[(block_group) & (NR_BG_LOCKS-1)].lock;
|
H A D | fs_struct.h | 10 spinlock_t lock; member in struct:fs_struct 29 spin_lock(&fs->lock); 32 spin_unlock(&fs->lock); 37 spin_lock(&fs->lock); 40 spin_unlock(&fs->lock); 46 spin_lock(&fs->lock); 51 spin_unlock(&fs->lock);
|
H A D | semaphore.h | 17 raw_spinlock_t lock; member in struct:semaphore 24 .lock = __RAW_SPIN_LOCK_UNLOCKED((name).lock), \ 36 lockdep_init_map(&sem->lock.dep_map, "semaphore->lock", &__key, 0);
|
H A D | wakelock.h | 36 static inline void wake_lock_init(struct wake_lock *lock, int type, argument 39 wakeup_source_init(&lock->ws, name); 42 static inline void wake_lock_destroy(struct wake_lock *lock) argument 44 wakeup_source_trash(&lock->ws); 47 static inline void wake_lock(struct wake_lock *lock) argument 49 __pm_stay_awake(&lock->ws); 52 static inline void wake_lock_timeout(struct wake_lock *lock, long timeout) argument 54 __pm_wakeup_event(&lock->ws, jiffies_to_msecs(timeout)); 57 static inline void wake_unlock(struct wake_lock *lock) argument 59 __pm_relax(&lock 62 wake_lock_active(struct wake_lock *lock) argument [all...] |
H A D | basic_mmio_gpio.h | 49 * Used to lock bgpio_chip->data. Also, this is needed to keep 52 spinlock_t lock; member in struct:bgpio_chip
|
H A D | genalloc.h | 15 * available. If new memory is added to the pool a lock has to be 36 spinlock_t lock; member in struct:gen_pool
|
H A D | mempool.h | 15 spinlock_t lock; member in struct:mempool_s
|
H A D | ratelimit.h | 11 raw_spinlock_t lock; /* protect the state */ member in struct:ratelimit_state 23 .lock = __RAW_SPIN_LOCK_UNLOCKED(name.lock), \ 31 raw_spin_lock_init(&rs->lock);
|
H A D | spinlock_up.h | 25 static inline void arch_spin_lock(arch_spinlock_t *lock) argument 27 lock->slock = 0; 31 arch_spin_lock_flags(arch_spinlock_t *lock, unsigned long flags) argument 34 lock->slock = 0; 37 static inline int arch_spin_trylock(arch_spinlock_t *lock) argument 39 char oldval = lock->slock; 41 lock->slock = 0; 46 static inline void arch_spin_unlock(arch_spinlock_t *lock) argument 48 lock->slock = 1; 54 #define arch_read_lock(lock) d [all...] |
H A D | concap.h | 34 spinlock_t lock; member in struct:concap_proto
|
H A D | cyclomx.h | 48 spinlock_t lock; member in struct:cycx_device 63 spinlock_t lock; member in struct:cycx_device::__anon287::__anon288
|
H A D | iocontext.h | 50 * queue lock but the returned icq is valid only until the queue lock is 57 * - ioc lock nests inside q lock. 59 * - ioc->icq_list and icq->ioc_node are protected by ioc lock. 60 * q->icq_list and icq->q_node by q lock. 62 * - ioc->icq_tree and ioc->icq_hint are protected by ioc lock, while icq 63 * itself is protected by q lock. However, both the indexes and icq 65 * the q lock. 72 * locks. Due to the lock orderin 106 spinlock_t lock; member in struct:io_context [all...] |
H A D | kthread.h | 57 spinlock_t lock; member in struct:kthread_worker 71 .lock = __SPIN_LOCK_UNLOCKED((worker).lock), \ 88 * kthread_worker.lock and kthread_work.done need their own lockdep class 112 __init_kthread_worker((worker), "("#worker")->lock", &__key); \
|
/include/linux/spi/ |
H A D | max7301.h | 11 struct mutex lock; member in struct:max7301
|
H A D | spi_bitbang.h | 10 spinlock_t lock; member in struct:spi_bitbang
|
/include/media/ |
H A D | media-device.h | 46 * @lock: Entities list lock 47 * @graph_mutex: Entities graph operation lock 74 spinlock_t lock; member in struct:media_device
|
/include/sound/ |
H A D | seq_midi_event.h | 38 spinlock_t lock; member in struct:snd_midi_event
|
/include/drm/ |
H A D | drm_sarea.h | 70 struct drm_hw_lock lock; member in struct:drm_sarea 71 /** \todo Use readers/writer lock for drm_sarea::drawable_lock */
|