Searched refs:ring (Results 1 - 25 of 249) sorted by relevance

12345678910

/drivers/gpu/drm/i915/
H A Dintel_ringbuffer.c47 static inline int ring_space(struct intel_ring_buffer *ring) argument
49 int space = (ring->head & HEAD_ADDR) - (ring->tail + 8);
51 space += ring->size;
56 render_ring_flush(struct intel_ring_buffer *ring, argument
60 struct drm_device *dev = ring->dev;
111 ret = intel_ring_begin(ring, 2);
115 intel_ring_emit(ring, cmd);
116 intel_ring_emit(ring, MI_NOOP);
117 intel_ring_advance(ring);
160 intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring) argument
196 gen6_render_ring_flush(struct intel_ring_buffer *ring, u32 invalidate_domains, u32 flush_domains) argument
234 ring_write_tail(struct intel_ring_buffer *ring, u32 value) argument
241 intel_ring_get_active_head(struct intel_ring_buffer *ring) argument
250 init_ring_common(struct intel_ring_buffer *ring) argument
319 init_pipe_control(struct intel_ring_buffer *ring) argument
364 cleanup_pipe_control(struct intel_ring_buffer *ring) argument
381 init_render_ring(struct intel_ring_buffer *ring) argument
421 render_ring_cleanup(struct intel_ring_buffer *ring) argument
430 update_mboxes(struct intel_ring_buffer *ring, u32 seqno, u32 mmio_offset) argument
452 gen6_add_request(struct intel_ring_buffer *ring, u32 *seqno) argument
487 intel_ring_sync(struct intel_ring_buffer *waiter, struct intel_ring_buffer *signaller, int ring, u32 seqno) argument
561 pc_render_add_request(struct intel_ring_buffer *ring, u32 *result) argument
613 render_ring_add_request(struct intel_ring_buffer *ring, u32 *result) argument
634 gen6_ring_get_seqno(struct intel_ring_buffer *ring) argument
647 ring_get_seqno(struct intel_ring_buffer *ring) argument
653 pc_render_get_seqno(struct intel_ring_buffer *ring) argument
692 render_ring_get_irq(struct intel_ring_buffer *ring) argument
714 render_ring_put_irq(struct intel_ring_buffer *ring) argument
731 intel_ring_setup_status_page(struct intel_ring_buffer *ring) argument
763 bsd_ring_flush(struct intel_ring_buffer *ring, u32 invalidate_domains, u32 flush_domains) argument
780 ring_add_request(struct intel_ring_buffer *ring, u32 *result) argument
803 gen6_ring_get_irq(struct intel_ring_buffer *ring, u32 gflag, u32 rflag) argument
828 gen6_ring_put_irq(struct intel_ring_buffer *ring, u32 gflag, u32 rflag) argument
845 bsd_ring_get_irq(struct intel_ring_buffer *ring) argument
865 bsd_ring_put_irq(struct intel_ring_buffer *ring) argument
881 ring_dispatch_execbuffer(struct intel_ring_buffer *ring, u32 offset, u32 length) argument
899 render_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, u32 offset, u32 len) argument
935 cleanup_status_page(struct intel_ring_buffer *ring) argument
952 init_status_page(struct intel_ring_buffer *ring) argument
996 intel_init_ring_buffer(struct drm_device *dev, struct intel_ring_buffer *ring) argument
1074 intel_cleanup_ring_buffer(struct intel_ring_buffer *ring) argument
1103 intel_wrap_ring_buffer(struct intel_ring_buffer *ring) argument
1127 intel_ring_wait_seqno(struct intel_ring_buffer *ring, u32 seqno) argument
1147 intel_ring_wait_request(struct intel_ring_buffer *ring, int n) argument
1204 intel_wait_ring_buffer(struct intel_ring_buffer *ring, int n) argument
1248 intel_ring_begin(struct intel_ring_buffer *ring, int num_dwords) argument
1274 intel_ring_advance(struct intel_ring_buffer *ring) argument
1319 gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring, u32 value) argument
1341 gen6_ring_flush(struct intel_ring_buffer *ring, u32 invalidate, u32 flush) argument
1363 gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, u32 offset, u32 len) argument
1381 gen6_render_ring_get_irq(struct intel_ring_buffer *ring) argument
1389 gen6_render_ring_put_irq(struct intel_ring_buffer *ring) argument
1397 gen6_bsd_ring_get_irq(struct intel_ring_buffer *ring) argument
1405 gen6_bsd_ring_put_irq(struct intel_ring_buffer *ring) argument
1436 blt_ring_get_irq(struct intel_ring_buffer *ring) argument
1444 blt_ring_put_irq(struct intel_ring_buffer *ring) argument
1451 blt_ring_flush(struct intel_ring_buffer *ring, u32 invalidate, u32 flush) argument
1495 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; local
1520 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; local
1565 struct intel_ring_buffer *ring = &dev_priv->ring[VCS]; local
1578 struct intel_ring_buffer *ring = &dev_priv->ring[BCS]; local
[all...]
H A Dintel_ringbuffer.h10 #define I915_READ_TAIL(ring) I915_READ(RING_TAIL((ring)->mmio_base))
11 #define I915_WRITE_TAIL(ring, val) I915_WRITE(RING_TAIL((ring)->mmio_base), val)
13 #define I915_READ_START(ring) I915_READ(RING_START((ring)->mmio_base))
14 #define I915_WRITE_START(ring, val) I915_WRITE(RING_START((ring)->mmio_base), val)
16 #define I915_READ_HEAD(ring) I915_READ(RING_HEAD((ring)
127 intel_ring_flag(struct intel_ring_buffer *ring) argument
133 intel_ring_sync_index(struct intel_ring_buffer *ring, struct intel_ring_buffer *other) argument
152 intel_read_status_page(struct intel_ring_buffer *ring, int reg) argument
181 intel_wait_ring_idle(struct intel_ring_buffer *ring) argument
188 intel_ring_emit(struct intel_ring_buffer *ring, u32 data) argument
206 intel_ring_get_tail(struct intel_ring_buffer *ring) argument
211 i915_trace_irq_get(struct intel_ring_buffer *ring, u32 seqno) argument
[all...]
H A Di915_trace.h236 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno),
237 TP_ARGS(ring, seqno),
241 __field(u32, ring)
246 __entry->dev = ring->dev->primary->index;
247 __entry->ring = ring->id;
249 i915_trace_irq_get(ring, seqno);
252 TP_printk("dev=%u, ring=%u, seqno=%u",
253 __entry->dev, __entry->ring, __entry->seqno)
257 TP_PROTO(struct intel_ring_buffer *ring, u3
[all...]
H A Di915_irq.c354 struct intel_ring_buffer *ring)
359 if (ring->obj == NULL)
362 seqno = ring->get_seqno(ring);
363 trace_i915_gem_request_complete(ring, seqno);
365 ring->irq_seqno = seqno;
366 wake_up_all(&ring->irq_queue);
524 notify_ring(dev, &dev_priv->ring[RCS]);
526 notify_ring(dev, &dev_priv->ring[VCS]);
528 notify_ring(dev, &dev_priv->ring[BC
353 notify_ring(struct drm_device *dev, struct intel_ring_buffer *ring) argument
864 i915_error_first_batchbuffer(struct drm_i915_private *dev_priv, struct intel_ring_buffer *ring) argument
893 i915_record_ring_state(struct drm_device *dev, struct drm_i915_error_state *error, struct intel_ring_buffer *ring) argument
941 struct intel_ring_buffer *ring = &dev_priv->ring[i]; local
1449 struct intel_ring_buffer *ring = LP_RING(dev_priv); local
1677 ring_last_seqno(struct intel_ring_buffer *ring) argument
1683 i915_hangcheck_ring_idle(struct intel_ring_buffer *ring, bool *err) argument
1701 kick_ring(struct intel_ring_buffer *ring) argument
[all...]
/drivers/staging/iio/
H A Dring_sw.c1 /* The industrial I/O simple minimally locked ring buffer.
21 * struct iio_sw_ring_buffer - software ring buffer
22 * @buf: generic ring buffer elements
23 * @data: the ring buffer memory
29 * Note that the first element of all ring buffers must be a
44 static inline int __iio_allocate_sw_ring_buffer(struct iio_sw_ring_buffer *ring, argument
49 __iio_update_buffer(&ring->buf, bytes_per_datum, length);
50 ring->data = kmalloc(length*ring->buf.bytes_per_datum, GFP_ATOMIC);
51 ring
57 __iio_free_sw_ring_buffer(struct iio_sw_ring_buffer *ring) argument
67 iio_store_to_sw_ring(struct iio_sw_ring_buffer *ring, unsigned char *data, s64 timestamp) argument
142 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local
262 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local
269 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local
283 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local
289 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local
345 struct iio_sw_ring_buffer *ring; local
[all...]
H A Dring_sw.h1 /* The industrial I/O simple minimally locked ring buffer.
10 * as it is intended that in the future a number of different software ring
16 * process. To this end the ring is not locked during a read.
29 void iio_sw_rb_free(struct iio_buffer *ring);
/drivers/net/wireless/b43legacy/
H A Ddma.c45 struct b43legacy_dmadesc32 *op32_idx2desc(struct b43legacy_dmaring *ring, argument
51 *meta = &(ring->meta[slot]);
52 desc = ring->descbase;
58 static void op32_fill_descriptor(struct b43legacy_dmaring *ring, argument
63 struct b43legacy_dmadesc32 *descbase = ring->descbase;
70 B43legacy_WARN_ON(!(slot >= 0 && slot < ring->nr_slots));
75 addr |= ring->dev->dma.translation;
76 ctl = (bufsize - ring->frameoffset)
78 if (slot == ring->nr_slots - 1)
93 static void op32_poke_tx(struct b43legacy_dmaring *ring, in argument
99 op32_tx_suspend(struct b43legacy_dmaring *ring) argument
106 op32_tx_resume(struct b43legacy_dmaring *ring) argument
113 op32_get_current_rxslot(struct b43legacy_dmaring *ring) argument
123 op32_set_current_rxslot(struct b43legacy_dmaring *ring, int slot) argument
130 free_slots(struct b43legacy_dmaring *ring) argument
135 next_slot(struct b43legacy_dmaring *ring, int slot) argument
143 prev_slot(struct b43legacy_dmaring *ring, int slot) argument
152 update_max_used_slots(struct b43legacy_dmaring *ring, int current_used_slots) argument
167 update_max_used_slots(struct b43legacy_dmaring *ring, int current_used_slots) argument
174 request_slot(struct b43legacy_dmaring *ring) argument
196 struct b43legacy_dmaring *ring; local
230 txring_to_priority(struct b43legacy_dmaring *ring) argument
260 map_descbuffer(struct b43legacy_dmaring *ring, unsigned char *buf, size_t len, int tx) argument
280 unmap_descbuffer(struct b43legacy_dmaring *ring, dma_addr_t addr, size_t len, int tx) argument
296 sync_descbuffer_for_cpu(struct b43legacy_dmaring *ring, dma_addr_t addr, size_t len) argument
307 sync_descbuffer_for_device(struct b43legacy_dmaring *ring, dma_addr_t addr, size_t len) argument
318 free_descriptor_buffer(struct b43legacy_dmaring *ring, struct b43legacy_dmadesc_meta *meta, int irq_context) argument
331 alloc_ringmemory(struct b43legacy_dmaring *ring) argument
348 free_ringmemory(struct b43legacy_dmaring *ring) argument
429 b43legacy_dma_mapping_error(struct b43legacy_dmaring *ring, dma_addr_t addr, size_t buffersize, bool dma_to_device) argument
458 setup_rx_descbuffer(struct b43legacy_dmaring *ring, struct b43legacy_dmadesc32 *desc, struct b43legacy_dmadesc_meta *meta, gfp_t gfp_flags) argument
508 alloc_initial_descbuffers(struct b43legacy_dmaring *ring) argument
545 dmacontroller_setup(struct b43legacy_dmaring *ring) argument
587 dmacontroller_cleanup(struct b43legacy_dmaring *ring) argument
600 free_all_descbuffers(struct b43legacy_dmaring *ring) argument
658 struct b43legacy_dmaring *ring; local
757 b43legacy_destroy_dmaring(struct b43legacy_dmaring *ring) argument
848 struct b43legacy_dmaring *ring; local
945 generate_cookie(struct b43legacy_dmaring *ring, int slot) argument
988 struct b43legacy_dmaring *ring = NULL; local
1018 dma_tx_fragment(struct b43legacy_dmaring *ring, struct sk_buff **in_skb) argument
1119 should_inject_overflow(struct b43legacy_dmaring *ring) argument
1144 struct b43legacy_dmaring *ring; local
1197 struct b43legacy_dmaring *ring; local
1313 dma_rx(struct b43legacy_dmaring *ring, int *slot) argument
1411 b43legacy_dma_rx(struct b43legacy_dmaring *ring) argument
1431 b43legacy_dma_tx_suspend_ring(struct b43legacy_dmaring *ring) argument
1437 b43legacy_dma_tx_resume_ring(struct b43legacy_dmaring *ring) argument
[all...]
/drivers/net/wireless/b43/
H A Ddma.c85 struct b43_dmadesc_generic *op32_idx2desc(struct b43_dmaring *ring, argument
91 *meta = &(ring->meta[slot]);
92 desc = ring->descbase;
98 static void op32_fill_descriptor(struct b43_dmaring *ring, argument
103 struct b43_dmadesc32 *descbase = ring->descbase;
110 B43_WARN_ON(!(slot >= 0 && slot < ring->nr_slots));
112 addr = b43_dma_address(&ring->dev->dma, dmaaddr, B43_DMA_ADDR_LOW);
113 addrext = b43_dma_address(&ring->dev->dma, dmaaddr, B43_DMA_ADDR_EXT);
116 if (slot == ring->nr_slots - 1)
131 static void op32_poke_tx(struct b43_dmaring *ring, in argument
137 op32_tx_suspend(struct b43_dmaring *ring) argument
143 op32_tx_resume(struct b43_dmaring *ring) argument
149 op32_get_current_rxslot(struct b43_dmaring *ring) argument
159 op32_set_current_rxslot(struct b43_dmaring *ring, int slot) argument
177 op64_idx2desc(struct b43_dmaring *ring, int slot, struct b43_dmadesc_meta **meta) argument
190 op64_fill_descriptor(struct b43_dmaring *ring, struct b43_dmadesc_generic *desc, dma_addr_t dmaaddr, u16 bufsize, int start, int end, int irq) argument
226 op64_poke_tx(struct b43_dmaring *ring, int slot) argument
232 op64_tx_suspend(struct b43_dmaring *ring) argument
238 op64_tx_resume(struct b43_dmaring *ring) argument
244 op64_get_current_rxslot(struct b43_dmaring *ring) argument
254 op64_set_current_rxslot(struct b43_dmaring *ring, int slot) argument
270 free_slots(struct b43_dmaring *ring) argument
275 next_slot(struct b43_dmaring *ring, int slot) argument
283 prev_slot(struct b43_dmaring *ring, int slot) argument
292 update_max_used_slots(struct b43_dmaring *ring, int current_used_slots) argument
307 update_max_used_slots(struct b43_dmaring *ring, int current_used_slots) argument
313 request_slot(struct b43_dmaring *ring) argument
360 map_descbuffer(struct b43_dmaring *ring, unsigned char *buf, size_t len, int tx) argument
377 unmap_descbuffer(struct b43_dmaring *ring, dma_addr_t addr, size_t len, int tx) argument
390 sync_descbuffer_for_cpu(struct b43_dmaring *ring, dma_addr_t addr, size_t len) argument
399 sync_descbuffer_for_device(struct b43_dmaring *ring, dma_addr_t addr, size_t len) argument
408 free_descriptor_buffer(struct b43_dmaring *ring, struct b43_dmadesc_meta *meta) argument
417 alloc_ringmemory(struct b43_dmaring *ring) argument
445 free_ringmemory(struct b43_dmaring *ring) argument
553 b43_dma_mapping_error(struct b43_dmaring *ring, dma_addr_t addr, size_t buffersize, bool dma_to_device) argument
585 b43_rx_buffer_is_poisoned(struct b43_dmaring *ring, struct sk_buff *skb) argument
592 b43_poison_rx_buffer(struct b43_dmaring *ring, struct sk_buff *skb) argument
607 setup_rx_descbuffer(struct b43_dmaring *ring, struct b43_dmadesc_generic *desc, struct b43_dmadesc_meta *meta, gfp_t gfp_flags) argument
651 alloc_initial_descbuffers(struct b43_dmaring *ring) argument
687 dmacontroller_setup(struct b43_dmaring *ring) argument
768 dmacontroller_cleanup(struct b43_dmaring *ring) argument
789 free_all_descbuffers(struct b43_dmaring *ring) argument
865 struct b43_dmaring *ring; local
994 b43_destroy_dmaring(struct b43_dmaring *ring, const char *ringname) argument
1203 generate_cookie(struct b43_dmaring *ring, int slot) argument
1227 struct b43_dmaring *ring = NULL; local
1256 dma_tx_fragment(struct b43_dmaring *ring, struct sk_buff *skb) argument
1355 should_inject_overflow(struct b43_dmaring *ring) argument
1380 struct b43_dmaring *ring; local
1409 struct b43_dmaring *ring; local
1485 struct b43_dmaring *ring; local
1606 dma_rx(struct b43_dmaring *ring, int *slot) argument
1692 b43_dma_rx(struct b43_dmaring *ring) argument
1712 b43_dma_tx_suspend_ring(struct b43_dmaring *ring) argument
1718 b43_dma_tx_resume_ring(struct b43_dmaring *ring) argument
[all...]
/drivers/gpu/drm/radeon/
H A Devergreen_blit_kms.c42 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
56 radeon_ring_write(ring, PACKET3(PACKET3_SET_CONTEXT_REG, 15));
57 radeon_ring_write(ring, (CB_COLOR0_BASE - PACKET3_SET_CONTEXT_REG_START) >> 2);
58 radeon_ring_write(ring, gpu_addr >> 8);
59 radeon_ring_write(ring, pitch);
60 radeon_ring_write(ring, slice);
61 radeon_ring_write(ring, 0);
62 radeon_ring_write(ring, cb_color_info);
63 radeon_ring_write(ring,
81 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
109 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
137 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
182 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
223 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
254 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
279 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
[all...]
H A Dradeon_ring.c64 void radeon_ring_write(struct radeon_ring *ring, uint32_t v) argument
67 if (ring->count_dw <= 0) {
68 DRM_ERROR("radeon: writting more dword to ring than expected !\n");
71 ring->ring[ring->wptr++] = v;
72 ring->wptr &= ring->ptr_mask;
73 ring->count_dw--;
74 ring
95 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib **ib, unsigned size) argument
185 struct radeon_ring *ring = &rdev->ring[ib->fence->ring]; local
276 radeon_ring_index(struct radeon_device *rdev, struct radeon_ring *ring) argument
291 radeon_ring_free_size(struct radeon_device *rdev, struct radeon_ring *ring) argument
310 radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) argument
331 radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) argument
344 radeon_ring_commit(struct radeon_device *rdev, struct radeon_ring *ring) argument
360 radeon_ring_unlock_commit(struct radeon_device *rdev, struct radeon_ring *ring) argument
366 radeon_ring_unlock_undo(struct radeon_device *rdev, struct radeon_ring *ring) argument
372 radeon_ring_init(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ring_size, unsigned rptr_offs, unsigned rptr_reg, unsigned wptr_reg, u32 ptr_reg_shift, u32 ptr_reg_mask, u32 nop) argument
417 radeon_ring_fini(struct radeon_device *rdev, struct radeon_ring *ring) argument
450 struct radeon_ring *ring = &rdev->ring[ridx]; local
[all...]
H A Dradeon_fence.c43 static void radeon_fence_write(struct radeon_device *rdev, u32 seq, int ring) argument
46 *rdev->fence_drv[ring].cpu_addr = cpu_to_le32(seq);
48 WREG32(rdev->fence_drv[ring].scratch_reg, seq);
52 static u32 radeon_fence_read(struct radeon_device *rdev, int ring) argument
57 seq = le32_to_cpu(*rdev->fence_drv[ring].cpu_addr);
59 seq = RREG32(rdev->fence_drv[ring].scratch_reg);
73 fence->seq = atomic_add_return(1, &rdev->fence_drv[fence->ring].seq);
74 if (!rdev->ring[fence->ring].ready)
78 radeon_fence_write(rdev, fence->seq, fence->ring);
89 radeon_fence_poll_locked(struct radeon_device *rdev, int ring) argument
161 radeon_fence_create(struct radeon_device *rdev, struct radeon_fence **fence, int ring) argument
286 radeon_fence_wait_next(struct radeon_device *rdev, int ring) argument
309 radeon_fence_wait_last(struct radeon_device *rdev, int ring) argument
348 radeon_fence_process(struct radeon_device *rdev, int ring) argument
361 radeon_fence_count_emitted(struct radeon_device *rdev, int ring) argument
384 radeon_fence_driver_start_ring(struct radeon_device *rdev, int ring) argument
416 radeon_fence_driver_init_ring(struct radeon_device *rdev, int ring) argument
432 int ring; local
448 int ring; local
[all...]
H A Dr600_blit_kms.c40 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
54 radeon_ring_write(ring, PACKET3(PACKET3_SET_CONTEXT_REG, 1));
55 radeon_ring_write(ring, (CB_COLOR0_BASE - PACKET3_SET_CONTEXT_REG_OFFSET) >> 2);
56 radeon_ring_write(ring, gpu_addr >> 8);
59 radeon_ring_write(ring, PACKET3(PACKET3_SURFACE_BASE_UPDATE, 0));
60 radeon_ring_write(ring, 2 << 0);
63 radeon_ring_write(ring, PACKET3(PACKET3_SET_CONTEXT_REG, 1));
64 radeon_ring_write(ring, (CB_COLOR0_SIZE - PACKET3_SET_CONTEXT_REG_OFFSET) >> 2);
65 radeon_ring_write(ring, (pitc
94 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
113 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
160 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
197 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
236 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
257 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
282 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
693 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local
[all...]
/drivers/net/ethernet/mellanox/mlx4/
H A Den_tx.c57 struct mlx4_en_tx_ring *ring, int qpn, u32 size,
64 ring->size = size;
65 ring->size_mask = size - 1;
66 ring->stride = stride;
70 spin_lock_init(&ring->comp_lock);
73 ring->tx_info = vmalloc(tmp);
74 if (!ring->tx_info)
77 en_dbg(DRV, priv, "Allocated tx_info ring at addr:%p size:%d\n",
78 ring->tx_info, tmp);
80 ring
56 mlx4_en_create_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int qpn, u32 size, u16 stride) argument
138 mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring) argument
157 mlx4_en_activate_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int cq) argument
187 mlx4_en_deactivate_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring) argument
197 mlx4_en_free_tx_desc(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int index, u8 owner) argument
277 mlx4_en_free_tx_buf(struct net_device *dev, struct mlx4_en_tx_ring *ring) argument
311 struct mlx4_en_tx_ring *ring = &priv->tx_ring[cq->ring]; local
380 struct mlx4_en_tx_ring *ring = &priv->tx_ring[cq->ring]; local
394 struct mlx4_en_tx_ring *ring = &priv->tx_ring[cq->ring]; local
415 mlx4_en_bounce_to_desc(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, u32 index, unsigned int desc_size) argument
446 struct mlx4_en_tx_ring *ring = &priv->tx_ring[tx_ind]; local
596 struct mlx4_en_tx_ring *ring; local
[all...]
H A Den_rx.c82 struct mlx4_en_rx_ring *ring)
88 page_alloc = &ring->page_alloc[i];
102 page_alloc = &ring->page_alloc[i];
110 struct mlx4_en_rx_ring *ring)
116 page_alloc = &ring->page_alloc[i];
127 struct mlx4_en_rx_ring *ring, int index)
129 struct mlx4_en_rx_desc *rx_desc = ring->buf + ring->stride * index;
130 struct skb_frag_struct *skb_frags = ring->rx_info +
143 /* If the number of used fragments does not fill up the ring strid
81 mlx4_en_init_allocator(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument
109 mlx4_en_destroy_allocator(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument
126 mlx4_en_init_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument
155 mlx4_en_prepare_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument
179 mlx4_en_update_rx_prod_db(struct mlx4_en_rx_ring *ring) argument
184 mlx4_en_free_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument
207 struct mlx4_en_rx_ring *ring; local
249 mlx4_en_free_rx_buf(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument
267 mlx4_en_create_rx_ring(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, u32 size, u16 stride) argument
316 struct mlx4_en_rx_ring *ring; local
383 mlx4_en_destroy_rx_ring(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, u32 size, u16 stride) argument
394 mlx4_en_deactivate_rx_ring(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument
533 struct mlx4_en_rx_ring *ring = &priv->rx_ring[cq->ring]; local
802 mlx4_en_config_rss_qp(struct mlx4_en_priv *priv, int qpn, struct mlx4_en_rx_ring *ring, enum mlx4_qp_state *state, struct mlx4_qp *qp) argument
[all...]
/drivers/scsi/fnic/
H A Dvnic_cq_copy.h35 desc = (struct fcpio_fw_req *)((u8 *)cq->ring.descs +
36 cq->ring.desc_size * cq->to_clean);
45 if (cq->to_clean == cq->ring.desc_count) {
50 desc = (struct fcpio_fw_req *)((u8 *)cq->ring.descs +
51 cq->ring.desc_size * cq->to_clean);
H A Dvnic_wq_copy.h31 struct vnic_dev_ring ring; member in struct:vnic_wq_copy
38 return wq->ring.desc_avail;
43 return wq->ring.desc_count - 1 - wq->ring.desc_avail;
48 struct fcpio_host_req *desc = wq->ring.descs;
55 ((wq->to_use_index + 1) == wq->ring.desc_count) ?
57 wq->ring.desc_avail--;
76 cnt = wq->ring.desc_count - wq->to_clean_index + index + 1;
78 wq->to_clean_index = ((index + 1) % wq->ring.desc_count);
79 wq->ring
[all...]
/drivers/staging/comedi/drivers/
H A Dmite.h60 struct mite_dma_descriptor_ring *ring; member in struct:mite_channel
84 struct mite_dma_descriptor_ring *ring = local
86 if (ring == NULL)
87 return ring;
88 ring->hw_dev = get_device(&mite->pcidev->dev);
89 if (ring->hw_dev == NULL) {
90 kfree(ring);
93 ring->n_links = 0;
94 ring->descriptors = NULL;
95 ring
99 mite_free_ring(struct mite_dma_descriptor_ring *ring) argument
137 mite_request_channel(struct mite_struct *mite, struct mite_dma_descriptor_ring *ring) argument
[all...]
/drivers/staging/iio/accel/
H A Dadis16201_ring.c15 * adis16201_read_ring_data() read data registers which will be placed into ring
65 struct iio_buffer *ring = indio_dev->buffer; local
69 size_t datasize = ring->access->get_bytes_per_datum(ring);
73 dev_err(&st->us->dev, "memory alloc failed in ring bh");
84 if (ring->scan_timestamp)
87 ring->access->store_to(ring, (u8 *)data, pf->timestamp);
110 struct iio_buffer *ring; local
112 ring
[all...]
H A Dadis16203_ring.c14 * adis16203_read_ring_data() read data registers which will be placed into ring
65 struct iio_buffer *ring = indio_dev->buffer; local
69 size_t datasize = ring->access->get_bytes_per_datum(ring);
73 dev_err(&st->us->dev, "memory alloc failed in ring bh");
84 if (ring->scan_timestamp)
87 ring->access->store_to(ring,
112 struct iio_buffer *ring; local
114 ring
[all...]
H A Dadis16204_ring.c14 * adis16204_read_ring_data() read data registers which will be placed into ring
63 struct iio_buffer *ring = indio_dev->buffer; local
66 size_t datasize = ring->access->get_bytes_per_datum(ring);
70 dev_err(&st->us->dev, "memory alloc failed in ring bh");
81 if (ring->scan_timestamp)
84 ring->access->store_to(ring, (u8 *)data, pf->timestamp);
107 struct iio_buffer *ring; local
109 ring
[all...]
H A Dadis16209_ring.c14 * adis16209_read_ring_data() read data registers which will be placed into ring
63 struct iio_buffer *ring = indio_dev->buffer; local
67 size_t datasize = ring->access->get_bytes_per_datum(ring);
71 dev_err(&st->us->dev, "memory alloc failed in ring bh");
82 if (ring->scan_timestamp)
85 ring->access->store_to(ring, (u8 *)data, pf->timestamp);
108 struct iio_buffer *ring; local
110 ring
[all...]
H A Dadis16240_ring.c14 * adis16240_read_ring_data() read data registers which will be placed into ring
60 struct iio_buffer *ring = indio_dev->buffer; local
64 size_t datasize = ring->access->get_bytes_per_datum(ring);
68 dev_err(&st->us->dev, "memory alloc failed in ring bh");
79 if (ring->scan_timestamp)
82 ring->access->store_to(ring, (u8 *)data, pf->timestamp);
105 struct iio_buffer *ring; local
107 ring
[all...]
/drivers/staging/iio/gyro/
H A Dadis16260_ring.c14 * adis16260_read_ring_data() read data registers which will be placed into ring
66 struct iio_buffer *ring = indio_dev->buffer; local
69 size_t datasize = ring->access->get_bytes_per_datum(ring);
73 dev_err(&st->us->dev, "memory alloc failed in ring bh");
84 if (ring->scan_timestamp)
87 ring->access->store_to(ring, (u8 *)data, pf->timestamp);
110 struct iio_buffer *ring; local
112 ring
[all...]
/drivers/gpu/drm/r128/
H A Dr128_drv.h79 drm_r128_ring_buffer_t ring; member in struct:drm_r128_private
415 drm_r128_ring_buffer_t *ring = &dev_priv->ring; local
416 ring->space = (GET_RING_HEAD(dev_priv) - ring->tail) * sizeof(u32);
417 if (ring->space <= 0)
418 ring->space += ring->size;
435 drm_r128_ring_buffer_t *ring = &dev_priv->ring; in
[all...]
/drivers/scsi/
H A Dscsi_tgt_if.c56 static inline void tgt_ring_idx_inc(struct tgt_ring *ring) argument
58 if (ring->tr_idx == TGT_MAX_EVENTS - 1)
59 ring->tr_idx = 0;
61 ring->tr_idx++;
64 static struct tgt_event *tgt_head_event(struct tgt_ring *ring, u32 idx) argument
72 (ring->tr_pages[pidx] + sizeof(struct tgt_event) * off);
78 struct tgt_ring *ring = &tx_ring; local
82 spin_lock_irqsave(&ring->tr_lock, flags);
84 ev = tgt_head_event(ring, ring
243 struct tgt_ring *ring = &rx_ring; local
264 struct tgt_ring *ring = &tx_ring; local
283 uspace_ring_map(struct vm_area_struct *vma, unsigned long addr, struct tgt_ring *ring) argument
344 tgt_ring_exit(struct tgt_ring *ring) argument
352 tgt_ring_init(struct tgt_ring *ring) argument
[all...]

Completed in 462 milliseconds

12345678910