/drivers/gpu/drm/i915/ |
H A D | intel_ringbuffer.c | 47 static inline int ring_space(struct intel_ring_buffer *ring) argument 49 int space = (ring->head & HEAD_ADDR) - (ring->tail + 8); 51 space += ring->size; 56 render_ring_flush(struct intel_ring_buffer *ring, argument 60 struct drm_device *dev = ring->dev; 111 ret = intel_ring_begin(ring, 2); 115 intel_ring_emit(ring, cmd); 116 intel_ring_emit(ring, MI_NOOP); 117 intel_ring_advance(ring); 160 intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring) argument 196 gen6_render_ring_flush(struct intel_ring_buffer *ring, u32 invalidate_domains, u32 flush_domains) argument 234 ring_write_tail(struct intel_ring_buffer *ring, u32 value) argument 241 intel_ring_get_active_head(struct intel_ring_buffer *ring) argument 250 init_ring_common(struct intel_ring_buffer *ring) argument 319 init_pipe_control(struct intel_ring_buffer *ring) argument 364 cleanup_pipe_control(struct intel_ring_buffer *ring) argument 381 init_render_ring(struct intel_ring_buffer *ring) argument 421 render_ring_cleanup(struct intel_ring_buffer *ring) argument 430 update_mboxes(struct intel_ring_buffer *ring, u32 seqno, u32 mmio_offset) argument 452 gen6_add_request(struct intel_ring_buffer *ring, u32 *seqno) argument 487 intel_ring_sync(struct intel_ring_buffer *waiter, struct intel_ring_buffer *signaller, int ring, u32 seqno) argument 561 pc_render_add_request(struct intel_ring_buffer *ring, u32 *result) argument 613 render_ring_add_request(struct intel_ring_buffer *ring, u32 *result) argument 634 gen6_ring_get_seqno(struct intel_ring_buffer *ring) argument 647 ring_get_seqno(struct intel_ring_buffer *ring) argument 653 pc_render_get_seqno(struct intel_ring_buffer *ring) argument 692 render_ring_get_irq(struct intel_ring_buffer *ring) argument 714 render_ring_put_irq(struct intel_ring_buffer *ring) argument 731 intel_ring_setup_status_page(struct intel_ring_buffer *ring) argument 763 bsd_ring_flush(struct intel_ring_buffer *ring, u32 invalidate_domains, u32 flush_domains) argument 780 ring_add_request(struct intel_ring_buffer *ring, u32 *result) argument 803 gen6_ring_get_irq(struct intel_ring_buffer *ring, u32 gflag, u32 rflag) argument 828 gen6_ring_put_irq(struct intel_ring_buffer *ring, u32 gflag, u32 rflag) argument 845 bsd_ring_get_irq(struct intel_ring_buffer *ring) argument 865 bsd_ring_put_irq(struct intel_ring_buffer *ring) argument 881 ring_dispatch_execbuffer(struct intel_ring_buffer *ring, u32 offset, u32 length) argument 899 render_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, u32 offset, u32 len) argument 935 cleanup_status_page(struct intel_ring_buffer *ring) argument 952 init_status_page(struct intel_ring_buffer *ring) argument 996 intel_init_ring_buffer(struct drm_device *dev, struct intel_ring_buffer *ring) argument 1074 intel_cleanup_ring_buffer(struct intel_ring_buffer *ring) argument 1103 intel_wrap_ring_buffer(struct intel_ring_buffer *ring) argument 1127 intel_ring_wait_seqno(struct intel_ring_buffer *ring, u32 seqno) argument 1147 intel_ring_wait_request(struct intel_ring_buffer *ring, int n) argument 1204 intel_wait_ring_buffer(struct intel_ring_buffer *ring, int n) argument 1248 intel_ring_begin(struct intel_ring_buffer *ring, int num_dwords) argument 1274 intel_ring_advance(struct intel_ring_buffer *ring) argument 1319 gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring, u32 value) argument 1341 gen6_ring_flush(struct intel_ring_buffer *ring, u32 invalidate, u32 flush) argument 1363 gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, u32 offset, u32 len) argument 1381 gen6_render_ring_get_irq(struct intel_ring_buffer *ring) argument 1389 gen6_render_ring_put_irq(struct intel_ring_buffer *ring) argument 1397 gen6_bsd_ring_get_irq(struct intel_ring_buffer *ring) argument 1405 gen6_bsd_ring_put_irq(struct intel_ring_buffer *ring) argument 1436 blt_ring_get_irq(struct intel_ring_buffer *ring) argument 1444 blt_ring_put_irq(struct intel_ring_buffer *ring) argument 1451 blt_ring_flush(struct intel_ring_buffer *ring, u32 invalidate, u32 flush) argument 1495 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; local 1520 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; local 1565 struct intel_ring_buffer *ring = &dev_priv->ring[VCS]; local 1578 struct intel_ring_buffer *ring = &dev_priv->ring[BCS]; local [all...] |
H A D | intel_ringbuffer.h | 10 #define I915_READ_TAIL(ring) I915_READ(RING_TAIL((ring)->mmio_base)) 11 #define I915_WRITE_TAIL(ring, val) I915_WRITE(RING_TAIL((ring)->mmio_base), val) 13 #define I915_READ_START(ring) I915_READ(RING_START((ring)->mmio_base)) 14 #define I915_WRITE_START(ring, val) I915_WRITE(RING_START((ring)->mmio_base), val) 16 #define I915_READ_HEAD(ring) I915_READ(RING_HEAD((ring) 127 intel_ring_flag(struct intel_ring_buffer *ring) argument 133 intel_ring_sync_index(struct intel_ring_buffer *ring, struct intel_ring_buffer *other) argument 152 intel_read_status_page(struct intel_ring_buffer *ring, int reg) argument 181 intel_wait_ring_idle(struct intel_ring_buffer *ring) argument 188 intel_ring_emit(struct intel_ring_buffer *ring, u32 data) argument 206 intel_ring_get_tail(struct intel_ring_buffer *ring) argument 211 i915_trace_irq_get(struct intel_ring_buffer *ring, u32 seqno) argument [all...] |
H A D | i915_trace.h | 236 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), 237 TP_ARGS(ring, seqno), 241 __field(u32, ring) 246 __entry->dev = ring->dev->primary->index; 247 __entry->ring = ring->id; 249 i915_trace_irq_get(ring, seqno); 252 TP_printk("dev=%u, ring=%u, seqno=%u", 253 __entry->dev, __entry->ring, __entry->seqno) 257 TP_PROTO(struct intel_ring_buffer *ring, u3 [all...] |
H A D | i915_irq.c | 354 struct intel_ring_buffer *ring) 359 if (ring->obj == NULL) 362 seqno = ring->get_seqno(ring); 363 trace_i915_gem_request_complete(ring, seqno); 365 ring->irq_seqno = seqno; 366 wake_up_all(&ring->irq_queue); 524 notify_ring(dev, &dev_priv->ring[RCS]); 526 notify_ring(dev, &dev_priv->ring[VCS]); 528 notify_ring(dev, &dev_priv->ring[BC 353 notify_ring(struct drm_device *dev, struct intel_ring_buffer *ring) argument 864 i915_error_first_batchbuffer(struct drm_i915_private *dev_priv, struct intel_ring_buffer *ring) argument 893 i915_record_ring_state(struct drm_device *dev, struct drm_i915_error_state *error, struct intel_ring_buffer *ring) argument 941 struct intel_ring_buffer *ring = &dev_priv->ring[i]; local 1449 struct intel_ring_buffer *ring = LP_RING(dev_priv); local 1677 ring_last_seqno(struct intel_ring_buffer *ring) argument 1683 i915_hangcheck_ring_idle(struct intel_ring_buffer *ring, bool *err) argument 1701 kick_ring(struct intel_ring_buffer *ring) argument [all...] |
/drivers/staging/iio/ |
H A D | ring_sw.c | 1 /* The industrial I/O simple minimally locked ring buffer. 21 * struct iio_sw_ring_buffer - software ring buffer 22 * @buf: generic ring buffer elements 23 * @data: the ring buffer memory 29 * Note that the first element of all ring buffers must be a 44 static inline int __iio_allocate_sw_ring_buffer(struct iio_sw_ring_buffer *ring, argument 49 __iio_update_buffer(&ring->buf, bytes_per_datum, length); 50 ring->data = kmalloc(length*ring->buf.bytes_per_datum, GFP_ATOMIC); 51 ring 57 __iio_free_sw_ring_buffer(struct iio_sw_ring_buffer *ring) argument 67 iio_store_to_sw_ring(struct iio_sw_ring_buffer *ring, unsigned char *data, s64 timestamp) argument 142 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local 262 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local 269 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local 283 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local 289 struct iio_sw_ring_buffer *ring = iio_to_sw_ring(r); local 345 struct iio_sw_ring_buffer *ring; local [all...] |
H A D | ring_sw.h | 1 /* The industrial I/O simple minimally locked ring buffer. 10 * as it is intended that in the future a number of different software ring 16 * process. To this end the ring is not locked during a read. 29 void iio_sw_rb_free(struct iio_buffer *ring);
|
/drivers/net/wireless/b43legacy/ |
H A D | dma.c | 45 struct b43legacy_dmadesc32 *op32_idx2desc(struct b43legacy_dmaring *ring, argument 51 *meta = &(ring->meta[slot]); 52 desc = ring->descbase; 58 static void op32_fill_descriptor(struct b43legacy_dmaring *ring, argument 63 struct b43legacy_dmadesc32 *descbase = ring->descbase; 70 B43legacy_WARN_ON(!(slot >= 0 && slot < ring->nr_slots)); 75 addr |= ring->dev->dma.translation; 76 ctl = (bufsize - ring->frameoffset) 78 if (slot == ring->nr_slots - 1) 93 static void op32_poke_tx(struct b43legacy_dmaring *ring, in argument 99 op32_tx_suspend(struct b43legacy_dmaring *ring) argument 106 op32_tx_resume(struct b43legacy_dmaring *ring) argument 113 op32_get_current_rxslot(struct b43legacy_dmaring *ring) argument 123 op32_set_current_rxslot(struct b43legacy_dmaring *ring, int slot) argument 130 free_slots(struct b43legacy_dmaring *ring) argument 135 next_slot(struct b43legacy_dmaring *ring, int slot) argument 143 prev_slot(struct b43legacy_dmaring *ring, int slot) argument 152 update_max_used_slots(struct b43legacy_dmaring *ring, int current_used_slots) argument 167 update_max_used_slots(struct b43legacy_dmaring *ring, int current_used_slots) argument 174 request_slot(struct b43legacy_dmaring *ring) argument 196 struct b43legacy_dmaring *ring; local 230 txring_to_priority(struct b43legacy_dmaring *ring) argument 260 map_descbuffer(struct b43legacy_dmaring *ring, unsigned char *buf, size_t len, int tx) argument 280 unmap_descbuffer(struct b43legacy_dmaring *ring, dma_addr_t addr, size_t len, int tx) argument 296 sync_descbuffer_for_cpu(struct b43legacy_dmaring *ring, dma_addr_t addr, size_t len) argument 307 sync_descbuffer_for_device(struct b43legacy_dmaring *ring, dma_addr_t addr, size_t len) argument 318 free_descriptor_buffer(struct b43legacy_dmaring *ring, struct b43legacy_dmadesc_meta *meta, int irq_context) argument 331 alloc_ringmemory(struct b43legacy_dmaring *ring) argument 348 free_ringmemory(struct b43legacy_dmaring *ring) argument 429 b43legacy_dma_mapping_error(struct b43legacy_dmaring *ring, dma_addr_t addr, size_t buffersize, bool dma_to_device) argument 458 setup_rx_descbuffer(struct b43legacy_dmaring *ring, struct b43legacy_dmadesc32 *desc, struct b43legacy_dmadesc_meta *meta, gfp_t gfp_flags) argument 508 alloc_initial_descbuffers(struct b43legacy_dmaring *ring) argument 545 dmacontroller_setup(struct b43legacy_dmaring *ring) argument 587 dmacontroller_cleanup(struct b43legacy_dmaring *ring) argument 600 free_all_descbuffers(struct b43legacy_dmaring *ring) argument 658 struct b43legacy_dmaring *ring; local 757 b43legacy_destroy_dmaring(struct b43legacy_dmaring *ring) argument 848 struct b43legacy_dmaring *ring; local 945 generate_cookie(struct b43legacy_dmaring *ring, int slot) argument 988 struct b43legacy_dmaring *ring = NULL; local 1018 dma_tx_fragment(struct b43legacy_dmaring *ring, struct sk_buff **in_skb) argument 1119 should_inject_overflow(struct b43legacy_dmaring *ring) argument 1144 struct b43legacy_dmaring *ring; local 1197 struct b43legacy_dmaring *ring; local 1313 dma_rx(struct b43legacy_dmaring *ring, int *slot) argument 1411 b43legacy_dma_rx(struct b43legacy_dmaring *ring) argument 1431 b43legacy_dma_tx_suspend_ring(struct b43legacy_dmaring *ring) argument 1437 b43legacy_dma_tx_resume_ring(struct b43legacy_dmaring *ring) argument [all...] |
/drivers/net/wireless/b43/ |
H A D | dma.c | 85 struct b43_dmadesc_generic *op32_idx2desc(struct b43_dmaring *ring, argument 91 *meta = &(ring->meta[slot]); 92 desc = ring->descbase; 98 static void op32_fill_descriptor(struct b43_dmaring *ring, argument 103 struct b43_dmadesc32 *descbase = ring->descbase; 110 B43_WARN_ON(!(slot >= 0 && slot < ring->nr_slots)); 112 addr = b43_dma_address(&ring->dev->dma, dmaaddr, B43_DMA_ADDR_LOW); 113 addrext = b43_dma_address(&ring->dev->dma, dmaaddr, B43_DMA_ADDR_EXT); 116 if (slot == ring->nr_slots - 1) 131 static void op32_poke_tx(struct b43_dmaring *ring, in argument 137 op32_tx_suspend(struct b43_dmaring *ring) argument 143 op32_tx_resume(struct b43_dmaring *ring) argument 149 op32_get_current_rxslot(struct b43_dmaring *ring) argument 159 op32_set_current_rxslot(struct b43_dmaring *ring, int slot) argument 177 op64_idx2desc(struct b43_dmaring *ring, int slot, struct b43_dmadesc_meta **meta) argument 190 op64_fill_descriptor(struct b43_dmaring *ring, struct b43_dmadesc_generic *desc, dma_addr_t dmaaddr, u16 bufsize, int start, int end, int irq) argument 226 op64_poke_tx(struct b43_dmaring *ring, int slot) argument 232 op64_tx_suspend(struct b43_dmaring *ring) argument 238 op64_tx_resume(struct b43_dmaring *ring) argument 244 op64_get_current_rxslot(struct b43_dmaring *ring) argument 254 op64_set_current_rxslot(struct b43_dmaring *ring, int slot) argument 270 free_slots(struct b43_dmaring *ring) argument 275 next_slot(struct b43_dmaring *ring, int slot) argument 283 prev_slot(struct b43_dmaring *ring, int slot) argument 292 update_max_used_slots(struct b43_dmaring *ring, int current_used_slots) argument 307 update_max_used_slots(struct b43_dmaring *ring, int current_used_slots) argument 313 request_slot(struct b43_dmaring *ring) argument 360 map_descbuffer(struct b43_dmaring *ring, unsigned char *buf, size_t len, int tx) argument 377 unmap_descbuffer(struct b43_dmaring *ring, dma_addr_t addr, size_t len, int tx) argument 390 sync_descbuffer_for_cpu(struct b43_dmaring *ring, dma_addr_t addr, size_t len) argument 399 sync_descbuffer_for_device(struct b43_dmaring *ring, dma_addr_t addr, size_t len) argument 408 free_descriptor_buffer(struct b43_dmaring *ring, struct b43_dmadesc_meta *meta) argument 417 alloc_ringmemory(struct b43_dmaring *ring) argument 445 free_ringmemory(struct b43_dmaring *ring) argument 553 b43_dma_mapping_error(struct b43_dmaring *ring, dma_addr_t addr, size_t buffersize, bool dma_to_device) argument 585 b43_rx_buffer_is_poisoned(struct b43_dmaring *ring, struct sk_buff *skb) argument 592 b43_poison_rx_buffer(struct b43_dmaring *ring, struct sk_buff *skb) argument 607 setup_rx_descbuffer(struct b43_dmaring *ring, struct b43_dmadesc_generic *desc, struct b43_dmadesc_meta *meta, gfp_t gfp_flags) argument 651 alloc_initial_descbuffers(struct b43_dmaring *ring) argument 687 dmacontroller_setup(struct b43_dmaring *ring) argument 768 dmacontroller_cleanup(struct b43_dmaring *ring) argument 789 free_all_descbuffers(struct b43_dmaring *ring) argument 865 struct b43_dmaring *ring; local 994 b43_destroy_dmaring(struct b43_dmaring *ring, const char *ringname) argument 1203 generate_cookie(struct b43_dmaring *ring, int slot) argument 1227 struct b43_dmaring *ring = NULL; local 1256 dma_tx_fragment(struct b43_dmaring *ring, struct sk_buff *skb) argument 1355 should_inject_overflow(struct b43_dmaring *ring) argument 1380 struct b43_dmaring *ring; local 1409 struct b43_dmaring *ring; local 1485 struct b43_dmaring *ring; local 1606 dma_rx(struct b43_dmaring *ring, int *slot) argument 1692 b43_dma_rx(struct b43_dmaring *ring) argument 1712 b43_dma_tx_suspend_ring(struct b43_dmaring *ring) argument 1718 b43_dma_tx_resume_ring(struct b43_dmaring *ring) argument [all...] |
/drivers/gpu/drm/radeon/ |
H A D | evergreen_blit_kms.c | 42 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 56 radeon_ring_write(ring, PACKET3(PACKET3_SET_CONTEXT_REG, 15)); 57 radeon_ring_write(ring, (CB_COLOR0_BASE - PACKET3_SET_CONTEXT_REG_START) >> 2); 58 radeon_ring_write(ring, gpu_addr >> 8); 59 radeon_ring_write(ring, pitch); 60 radeon_ring_write(ring, slice); 61 radeon_ring_write(ring, 0); 62 radeon_ring_write(ring, cb_color_info); 63 radeon_ring_write(ring, 81 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 109 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 137 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 182 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 223 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 254 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 279 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local [all...] |
H A D | radeon_ring.c | 64 void radeon_ring_write(struct radeon_ring *ring, uint32_t v) argument 67 if (ring->count_dw <= 0) { 68 DRM_ERROR("radeon: writting more dword to ring than expected !\n"); 71 ring->ring[ring->wptr++] = v; 72 ring->wptr &= ring->ptr_mask; 73 ring->count_dw--; 74 ring 95 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib **ib, unsigned size) argument 185 struct radeon_ring *ring = &rdev->ring[ib->fence->ring]; local 276 radeon_ring_index(struct radeon_device *rdev, struct radeon_ring *ring) argument 291 radeon_ring_free_size(struct radeon_device *rdev, struct radeon_ring *ring) argument 310 radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) argument 331 radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) argument 344 radeon_ring_commit(struct radeon_device *rdev, struct radeon_ring *ring) argument 360 radeon_ring_unlock_commit(struct radeon_device *rdev, struct radeon_ring *ring) argument 366 radeon_ring_unlock_undo(struct radeon_device *rdev, struct radeon_ring *ring) argument 372 radeon_ring_init(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ring_size, unsigned rptr_offs, unsigned rptr_reg, unsigned wptr_reg, u32 ptr_reg_shift, u32 ptr_reg_mask, u32 nop) argument 417 radeon_ring_fini(struct radeon_device *rdev, struct radeon_ring *ring) argument 450 struct radeon_ring *ring = &rdev->ring[ridx]; local [all...] |
H A D | radeon_fence.c | 43 static void radeon_fence_write(struct radeon_device *rdev, u32 seq, int ring) argument 46 *rdev->fence_drv[ring].cpu_addr = cpu_to_le32(seq); 48 WREG32(rdev->fence_drv[ring].scratch_reg, seq); 52 static u32 radeon_fence_read(struct radeon_device *rdev, int ring) argument 57 seq = le32_to_cpu(*rdev->fence_drv[ring].cpu_addr); 59 seq = RREG32(rdev->fence_drv[ring].scratch_reg); 73 fence->seq = atomic_add_return(1, &rdev->fence_drv[fence->ring].seq); 74 if (!rdev->ring[fence->ring].ready) 78 radeon_fence_write(rdev, fence->seq, fence->ring); 89 radeon_fence_poll_locked(struct radeon_device *rdev, int ring) argument 161 radeon_fence_create(struct radeon_device *rdev, struct radeon_fence **fence, int ring) argument 286 radeon_fence_wait_next(struct radeon_device *rdev, int ring) argument 309 radeon_fence_wait_last(struct radeon_device *rdev, int ring) argument 348 radeon_fence_process(struct radeon_device *rdev, int ring) argument 361 radeon_fence_count_emitted(struct radeon_device *rdev, int ring) argument 384 radeon_fence_driver_start_ring(struct radeon_device *rdev, int ring) argument 416 radeon_fence_driver_init_ring(struct radeon_device *rdev, int ring) argument 432 int ring; local 448 int ring; local [all...] |
H A D | r600_blit_kms.c | 40 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 54 radeon_ring_write(ring, PACKET3(PACKET3_SET_CONTEXT_REG, 1)); 55 radeon_ring_write(ring, (CB_COLOR0_BASE - PACKET3_SET_CONTEXT_REG_OFFSET) >> 2); 56 radeon_ring_write(ring, gpu_addr >> 8); 59 radeon_ring_write(ring, PACKET3(PACKET3_SURFACE_BASE_UPDATE, 0)); 60 radeon_ring_write(ring, 2 << 0); 63 radeon_ring_write(ring, PACKET3(PACKET3_SET_CONTEXT_REG, 1)); 64 radeon_ring_write(ring, (CB_COLOR0_SIZE - PACKET3_SET_CONTEXT_REG_OFFSET) >> 2); 65 radeon_ring_write(ring, (pitc 94 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 113 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 160 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 197 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 236 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 257 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 282 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local 693 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; local [all...] |
/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_tx.c | 57 struct mlx4_en_tx_ring *ring, int qpn, u32 size, 64 ring->size = size; 65 ring->size_mask = size - 1; 66 ring->stride = stride; 70 spin_lock_init(&ring->comp_lock); 73 ring->tx_info = vmalloc(tmp); 74 if (!ring->tx_info) 77 en_dbg(DRV, priv, "Allocated tx_info ring at addr:%p size:%d\n", 78 ring->tx_info, tmp); 80 ring 56 mlx4_en_create_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int qpn, u32 size, u16 stride) argument 138 mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring) argument 157 mlx4_en_activate_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int cq) argument 187 mlx4_en_deactivate_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring) argument 197 mlx4_en_free_tx_desc(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int index, u8 owner) argument 277 mlx4_en_free_tx_buf(struct net_device *dev, struct mlx4_en_tx_ring *ring) argument 311 struct mlx4_en_tx_ring *ring = &priv->tx_ring[cq->ring]; local 380 struct mlx4_en_tx_ring *ring = &priv->tx_ring[cq->ring]; local 394 struct mlx4_en_tx_ring *ring = &priv->tx_ring[cq->ring]; local 415 mlx4_en_bounce_to_desc(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, u32 index, unsigned int desc_size) argument 446 struct mlx4_en_tx_ring *ring = &priv->tx_ring[tx_ind]; local 596 struct mlx4_en_tx_ring *ring; local [all...] |
H A D | en_rx.c | 82 struct mlx4_en_rx_ring *ring) 88 page_alloc = &ring->page_alloc[i]; 102 page_alloc = &ring->page_alloc[i]; 110 struct mlx4_en_rx_ring *ring) 116 page_alloc = &ring->page_alloc[i]; 127 struct mlx4_en_rx_ring *ring, int index) 129 struct mlx4_en_rx_desc *rx_desc = ring->buf + ring->stride * index; 130 struct skb_frag_struct *skb_frags = ring->rx_info + 143 /* If the number of used fragments does not fill up the ring strid 81 mlx4_en_init_allocator(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument 109 mlx4_en_destroy_allocator(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument 126 mlx4_en_init_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument 155 mlx4_en_prepare_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument 179 mlx4_en_update_rx_prod_db(struct mlx4_en_rx_ring *ring) argument 184 mlx4_en_free_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, int index) argument 207 struct mlx4_en_rx_ring *ring; local 249 mlx4_en_free_rx_buf(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument 267 mlx4_en_create_rx_ring(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, u32 size, u16 stride) argument 316 struct mlx4_en_rx_ring *ring; local 383 mlx4_en_destroy_rx_ring(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring, u32 size, u16 stride) argument 394 mlx4_en_deactivate_rx_ring(struct mlx4_en_priv *priv, struct mlx4_en_rx_ring *ring) argument 533 struct mlx4_en_rx_ring *ring = &priv->rx_ring[cq->ring]; local 802 mlx4_en_config_rss_qp(struct mlx4_en_priv *priv, int qpn, struct mlx4_en_rx_ring *ring, enum mlx4_qp_state *state, struct mlx4_qp *qp) argument [all...] |
/drivers/scsi/fnic/ |
H A D | vnic_cq_copy.h | 35 desc = (struct fcpio_fw_req *)((u8 *)cq->ring.descs + 36 cq->ring.desc_size * cq->to_clean); 45 if (cq->to_clean == cq->ring.desc_count) { 50 desc = (struct fcpio_fw_req *)((u8 *)cq->ring.descs + 51 cq->ring.desc_size * cq->to_clean);
|
H A D | vnic_wq_copy.h | 31 struct vnic_dev_ring ring; member in struct:vnic_wq_copy 38 return wq->ring.desc_avail; 43 return wq->ring.desc_count - 1 - wq->ring.desc_avail; 48 struct fcpio_host_req *desc = wq->ring.descs; 55 ((wq->to_use_index + 1) == wq->ring.desc_count) ? 57 wq->ring.desc_avail--; 76 cnt = wq->ring.desc_count - wq->to_clean_index + index + 1; 78 wq->to_clean_index = ((index + 1) % wq->ring.desc_count); 79 wq->ring [all...] |
/drivers/staging/comedi/drivers/ |
H A D | mite.h | 60 struct mite_dma_descriptor_ring *ring; member in struct:mite_channel 84 struct mite_dma_descriptor_ring *ring = local 86 if (ring == NULL) 87 return ring; 88 ring->hw_dev = get_device(&mite->pcidev->dev); 89 if (ring->hw_dev == NULL) { 90 kfree(ring); 93 ring->n_links = 0; 94 ring->descriptors = NULL; 95 ring 99 mite_free_ring(struct mite_dma_descriptor_ring *ring) argument 137 mite_request_channel(struct mite_struct *mite, struct mite_dma_descriptor_ring *ring) argument [all...] |
/drivers/staging/iio/accel/ |
H A D | adis16201_ring.c | 15 * adis16201_read_ring_data() read data registers which will be placed into ring 65 struct iio_buffer *ring = indio_dev->buffer; local 69 size_t datasize = ring->access->get_bytes_per_datum(ring); 73 dev_err(&st->us->dev, "memory alloc failed in ring bh"); 84 if (ring->scan_timestamp) 87 ring->access->store_to(ring, (u8 *)data, pf->timestamp); 110 struct iio_buffer *ring; local 112 ring [all...] |
H A D | adis16203_ring.c | 14 * adis16203_read_ring_data() read data registers which will be placed into ring 65 struct iio_buffer *ring = indio_dev->buffer; local 69 size_t datasize = ring->access->get_bytes_per_datum(ring); 73 dev_err(&st->us->dev, "memory alloc failed in ring bh"); 84 if (ring->scan_timestamp) 87 ring->access->store_to(ring, 112 struct iio_buffer *ring; local 114 ring [all...] |
H A D | adis16204_ring.c | 14 * adis16204_read_ring_data() read data registers which will be placed into ring 63 struct iio_buffer *ring = indio_dev->buffer; local 66 size_t datasize = ring->access->get_bytes_per_datum(ring); 70 dev_err(&st->us->dev, "memory alloc failed in ring bh"); 81 if (ring->scan_timestamp) 84 ring->access->store_to(ring, (u8 *)data, pf->timestamp); 107 struct iio_buffer *ring; local 109 ring [all...] |
H A D | adis16209_ring.c | 14 * adis16209_read_ring_data() read data registers which will be placed into ring 63 struct iio_buffer *ring = indio_dev->buffer; local 67 size_t datasize = ring->access->get_bytes_per_datum(ring); 71 dev_err(&st->us->dev, "memory alloc failed in ring bh"); 82 if (ring->scan_timestamp) 85 ring->access->store_to(ring, (u8 *)data, pf->timestamp); 108 struct iio_buffer *ring; local 110 ring [all...] |
H A D | adis16240_ring.c | 14 * adis16240_read_ring_data() read data registers which will be placed into ring 60 struct iio_buffer *ring = indio_dev->buffer; local 64 size_t datasize = ring->access->get_bytes_per_datum(ring); 68 dev_err(&st->us->dev, "memory alloc failed in ring bh"); 79 if (ring->scan_timestamp) 82 ring->access->store_to(ring, (u8 *)data, pf->timestamp); 105 struct iio_buffer *ring; local 107 ring [all...] |
/drivers/staging/iio/gyro/ |
H A D | adis16260_ring.c | 14 * adis16260_read_ring_data() read data registers which will be placed into ring 66 struct iio_buffer *ring = indio_dev->buffer; local 69 size_t datasize = ring->access->get_bytes_per_datum(ring); 73 dev_err(&st->us->dev, "memory alloc failed in ring bh"); 84 if (ring->scan_timestamp) 87 ring->access->store_to(ring, (u8 *)data, pf->timestamp); 110 struct iio_buffer *ring; local 112 ring [all...] |
/drivers/gpu/drm/r128/ |
H A D | r128_drv.h | 79 drm_r128_ring_buffer_t ring; member in struct:drm_r128_private 415 drm_r128_ring_buffer_t *ring = &dev_priv->ring; local 416 ring->space = (GET_RING_HEAD(dev_priv) - ring->tail) * sizeof(u32); 417 if (ring->space <= 0) 418 ring->space += ring->size; 435 drm_r128_ring_buffer_t *ring = &dev_priv->ring; in [all...] |
/drivers/scsi/ |
H A D | scsi_tgt_if.c | 56 static inline void tgt_ring_idx_inc(struct tgt_ring *ring) argument 58 if (ring->tr_idx == TGT_MAX_EVENTS - 1) 59 ring->tr_idx = 0; 61 ring->tr_idx++; 64 static struct tgt_event *tgt_head_event(struct tgt_ring *ring, u32 idx) argument 72 (ring->tr_pages[pidx] + sizeof(struct tgt_event) * off); 78 struct tgt_ring *ring = &tx_ring; local 82 spin_lock_irqsave(&ring->tr_lock, flags); 84 ev = tgt_head_event(ring, ring 243 struct tgt_ring *ring = &rx_ring; local 264 struct tgt_ring *ring = &tx_ring; local 283 uspace_ring_map(struct vm_area_struct *vma, unsigned long addr, struct tgt_ring *ring) argument 344 tgt_ring_exit(struct tgt_ring *ring) argument 352 tgt_ring_init(struct tgt_ring *ring) argument [all...] |