/drivers/gpu/drm/radeon/ |
H A D | r200.c | 149 volatile uint32_t *ib; local 157 ib = p->ib->ptr; 164 DRM_ERROR("No reloc for ib[%d]=0x%04X\n", 181 DRM_ERROR("No reloc for ib[%d]=0x%04X\n", 189 ib[idx] = idx_value + ((u32)reloc->lobj.gpu_offset); 194 DRM_ERROR("No reloc for ib[%d]=0x%04X\n", 202 ib[idx] = idx_value + ((u32)reloc->lobj.gpu_offset); 213 DRM_ERROR("No reloc for ib[%d]=0x%04X\n", 226 ib[id [all...] |
H A D | r300.c | 181 * for enough space (today caller are ib schedule and buffer move) */ 624 volatile uint32_t *ib; local 630 ib = p->ib->ptr; 639 DRM_ERROR("No reloc for ib[%d]=0x%04X\n", 658 DRM_ERROR("No reloc for ib[%d]=0x%04X\n", 666 ib[idx] = idx_value + ((u32)reloc->lobj.gpu_offset); 671 DRM_ERROR("No reloc for ib[%d]=0x%04X\n", 679 ib[idx] = idx_value + ((u32)reloc->lobj.gpu_offset); 700 DRM_ERROR("No reloc for ib[ 1168 volatile uint32_t *ib; local [all...] |
H A D | radeon_ring.c | 80 bool radeon_ib_try_free(struct radeon_device *rdev, struct radeon_ib *ib) argument 84 /* only free ib which have been emited */ 85 if (ib->fence && ib->fence->emitted) { 86 if (radeon_fence_signaled(ib->fence)) { 87 radeon_fence_unref(&ib->fence); 88 radeon_sa_bo_free(rdev, &ib->sa_bo); 96 struct radeon_ib **ib, unsigned size) 102 *ib = NULL; 116 dev_err(rdev->dev, "failed to get an ib afte 95 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib **ib, unsigned size) argument 167 radeon_ib_free(struct radeon_device *rdev, struct radeon_ib **ib) argument 183 radeon_ib_schedule(struct radeon_device *rdev, struct radeon_ib *ib) argument 484 struct radeon_ib *ib = &rdev->ib_pool.ibs[*((unsigned*)node->info_ent->data)]; local [all...] |
H A D | evergreen_cs.c | 447 volatile u32 *ib = p->ib->ptr; local 469 ib[track->cb_color_slice_idx[id]] = slice; 982 * evergreen_cs_packet_parse() - parse cp packet and point ib index to next packet 987 * if packet is bigger than remaining ib size. or if packets is unknown. 1097 volatile uint32_t *ib; local 1099 ib = p->ib->ptr; 1157 ib[h_idx + 2] = PACKET2(0); 1158 ib[h_id 1236 u32 m, i, tmp, *ib; local 1932 volatile u32 *ib; local 2782 evergreen_vm_packet3_check(struct radeon_device *rdev, u32 *ib, struct radeon_cs_packet *pkt) argument 2868 evergreen_ib_parse(struct radeon_device *rdev, struct radeon_ib *ib) argument [all...] |
H A D | ni.c | 1144 void cayman_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) argument 1146 struct radeon_ring *ring = &rdev->ring[ib->fence->ring]; 1156 (ib->gpu_addr & 0xFFFFFFFC)); 1157 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFF); 1158 radeon_ring_write(ring, ib->length_dw | (ib->vm_id << 24)); 1163 radeon_ring_write(ring, ib->vm_id);
|
H A D | evergreen.c | 1372 void evergreen_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) argument 1374 struct radeon_ring *ring = &rdev->ring[ib->fence->ring]; 1385 (ib->gpu_addr & 0xFFFFFFFC)); 1386 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFF); 1387 radeon_ring_write(ring, ib->length_dw);
|
H A D | r600.c | 2691 void r600_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) argument 2693 struct radeon_ring *ring = &rdev->ring[ib->fence->ring]; 2701 (ib->gpu_addr & 0xFFFFFFFC)); 2702 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFF); 2703 radeon_ring_write(ring, ib->length_dw); 2708 struct radeon_ib *ib; local 2721 r = radeon_ib_get(rdev, ring_index, &ib, 256); 2723 DRM_ERROR("radeon: failed to get ib (%d).\n", r); 2726 ib->ptr[0] = PACKET3(PACKET3_SET_CONFIG_REG, 1); 2727 ib [all...] |
H A D | r600_cp.c | 65 unsigned family, u32 *ib, int *l); 2623 u32 *ib, cs_id = 0; local 2635 /* get ib */ 2641 ib = dev->agp_buffer_map->handle + buf->offset; 2643 r = r600_cs_legacy(dev, data, fpriv, family, ib, &l);
|
H A D | r600_cs.c | 348 volatile u32 *ib = p->ib->ptr; local 460 ib[track->cb_color_size_idx[i]] = tmp; 474 volatile u32 *ib = p->ib->ptr; local 512 ib[track->db_depth_size_idx] = S_028000_SLICE_TILE_MAX(tmp - 1) | (track->db_depth_size & 0x3FF); 778 * r600_cs_packet_parse() - parse cp packet and point ib index to next packet 783 * if packet is bigger than remaining ib size. or if packets is unknown. 962 volatile uint32_t *ib; local 964 ib 1102 u32 m, i, tmp, *ib; local 1709 volatile u32 *ib; local 2296 r600_cs_legacy(struct drm_device *dev, void *data, struct drm_file *filp, unsigned family, u32 *ib, int *l) argument [all...] |
H A D | r100.c | 119 DRM_ERROR("No reloc for ib[%d]=0x%04X\n", 142 p->ib->ptr[idx] = (value & 0x3fc00000) | tmp; 144 p->ib->ptr[idx] = (value & 0xffc00000) | tmp; 156 volatile uint32_t *ib; local 159 ib = p->ib->ptr; 178 ib[idx+1] = radeon_get_ib_value(p, idx + 1) + ((u32)reloc->lobj.gpu_offset); 190 ib[idx+2] = radeon_get_ib_value(p, idx + 2) + ((u32)reloc->lobj.gpu_offset); 204 ib[idx+1] = radeon_get_ib_value(p, idx + 1) + ((u32)reloc->lobj.gpu_offset); 846 * for enough space (today caller are ib schedul 1270 volatile uint32_t *ib; local 1351 volatile uint32_t *ib; local 1525 volatile uint32_t *ib; local 1885 volatile uint32_t *ib; local 3734 r100_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) argument 3745 struct radeon_ib *ib; local [all...] |
H A D | si.c | 1916 void si_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) argument 1918 struct radeon_ring *ring = &rdev->ring[ib->fence->ring]; 1921 if (ib->is_const_ib) 1931 (ib->gpu_addr & 0xFFFFFFFC)); 1932 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFFFF); 1933 radeon_ring_write(ring, ib->length_dw | (ib->vm_id << 24)); 1938 radeon_ring_write(ring, ib->vm_id); 2628 u32 *ib, struct radeon_cs_packet *pkt) 2650 u32 *ib, struc 2627 si_vm_packet3_ce_check(struct radeon_device *rdev, u32 *ib, struct radeon_cs_packet *pkt) argument 2649 si_vm_packet3_gfx_check(struct radeon_device *rdev, u32 *ib, struct radeon_cs_packet *pkt) argument 2761 si_vm_packet3_compute_check(struct radeon_device *rdev, u32 *ib, struct radeon_cs_packet *pkt) argument 2843 si_ib_parse(struct radeon_device *rdev, struct radeon_ib *ib) argument [all...] |
/drivers/isdn/divert/ |
H A D | divert_procfs.c | 43 struct divert_info *ib; local 52 if (!(ib = kmalloc(sizeof(struct divert_info) + strlen(cp), GFP_ATOMIC))) 54 strcpy(ib->info_start, cp); /* set output string */ 55 ib->next = NULL; 57 ib->usage_cnt = if_used; 59 divert_info_head = ib; /* new head */ 61 divert_info_tail->next = ib; /* follows existing messages */ 62 divert_info_tail = ib; /* new tail */ 68 ib = divert_info_head; 70 kfree(ib); [all...] |
/drivers/isdn/hysdn/ |
H A D | hysdn_proclog.c | 103 struct log_data *ib; local 117 if (!(ib = kmalloc(sizeof(struct log_data) + strlen(cp), GFP_ATOMIC))) 119 strcpy(ib->log_start, cp); /* set output string */ 120 ib->next = NULL; 121 ib->proc_ctrl = pd; /* point to own control structure */ 123 ib->usage_cnt = pd->if_used; 125 pd->log_head = ib; /* new head */ 127 pd->log_tail->next = ib; /* follows existing messages */ 128 pd->log_tail = ib; /* new tail */ 137 ib [all...] |
/drivers/net/ethernet/amd/ |
H A D | 7990.c | 100 t, ib->brx_ring[t].rmd1_hadr, ib->brx_ring[t].rmd0,\ 101 ib->brx_ring[t].length,\ 102 ib->brx_ring[t].mblength, ib->brx_ring[t].rmd1_bits);\ 106 t, ib->btx_ring[t].tmd1_hadr, ib->btx_ring[t].tmd0,\ 107 ib->btx_ring[t].length,\ 108 ib->btx_ring[t].misc, ib 140 volatile struct lance_init_block *ib = lp->init_block; local 275 volatile struct lance_init_block *ib = lp->init_block; local 352 volatile struct lance_init_block *ib = lp->init_block; local 536 volatile struct lance_init_block *ib = lp->init_block; local 593 volatile struct lance_init_block *ib = lp->init_block; local 620 volatile struct lance_init_block *ib = lp->init_block; local [all...] |
H A D | a2065.c | 151 volatile struct lance_init_block *ib = lp->init_block; local 162 ib->mode = 0; 167 ib->phys_addr[0] = dev->dev_addr[1]; 168 ib->phys_addr[1] = dev->dev_addr[0]; 169 ib->phys_addr[2] = dev->dev_addr[3]; 170 ib->phys_addr[3] = dev->dev_addr[2]; 171 ib->phys_addr[4] = dev->dev_addr[5]; 172 ib->phys_addr[5] = dev->dev_addr[4]; 178 ib->btx_ring[i].tmd0 = leptr; 179 ib 246 volatile struct lance_init_block *ib = lp->init_block; local 326 volatile struct lance_init_block *ib = lp->init_block; local 539 volatile struct lance_init_block *ib = lp->init_block; local 587 volatile struct lance_init_block *ib = lp->init_block; local 613 volatile struct lance_init_block *ib = lp->init_block; local [all...] |
H A D | ni65.c | 226 struct init_block ib; member in struct:priv 580 p->ib.eaddr[i] = daddr[i]; 583 p->ib.filter[i] = filter; 584 p->ib.mode = mode; 586 p->ib.trp = (u32) isa_virt_to_bus(p->tmdhead) | TMDNUMMASK; 587 p->ib.rrp = (u32) isa_virt_to_bus(p->rmdhead) | RMDNUMMASK; 589 pib = (u32) isa_virt_to_bus(&p->ib);
|
H A D | declance.c | 234 #define lib_ptr(ib, rt, type) \ 235 ((volatile u16 *)((u8 *)(ib) + lib_off(rt, type))) 451 volatile u16 *ib = (volatile u16 *)dev->mem_start; local 463 *lib_ptr(ib, phys_addr[0], lp->type) = (dev->dev_addr[1] << 8) | 465 *lib_ptr(ib, phys_addr[1], lp->type) = (dev->dev_addr[3] << 8) | 467 *lib_ptr(ib, phys_addr[2], lp->type) = (dev->dev_addr[5] << 8) | 473 *lib_ptr(ib, rx_len, lp->type) = (LANCE_LOG_RX_BUFFERS << 13) | 475 *lib_ptr(ib, rx_ptr, lp->type) = leptr; 482 *lib_ptr(ib, tx_len, lp->type) = (LANCE_LOG_TX_BUFFERS << 13) | 484 *lib_ptr(ib, tx_pt 557 volatile u16 *ib = (volatile u16 *)dev->mem_start; local 646 volatile u16 *ib = (volatile u16 *)dev->mem_start; local 782 volatile u16 *ib = (volatile u16 *)dev->mem_start; local 902 volatile u16 *ib = (volatile u16 *)dev->mem_start; local 946 volatile u16 *ib = (volatile u16 *)dev->mem_start; local 975 volatile u16 *ib = (volatile u16 *)dev->mem_start; local [all...] |
H A D | sunlance.c | 324 struct lance_init_block *ib = lp->init_block_mem; local 337 ib->phys_addr [0] = dev->dev_addr [1]; 338 ib->phys_addr [1] = dev->dev_addr [0]; 339 ib->phys_addr [2] = dev->dev_addr [3]; 340 ib->phys_addr [3] = dev->dev_addr [2]; 341 ib->phys_addr [4] = dev->dev_addr [5]; 342 ib->phys_addr [5] = dev->dev_addr [4]; 347 ib->btx_ring [i].tmd0 = leptr; 348 ib->btx_ring [i].tmd1_hadr = leptr >> 16; 349 ib 381 struct lance_init_block __iomem *ib = lp->init_block_iomem; local 511 struct lance_init_block *ib = lp->init_block_mem; local 572 struct lance_init_block *ib = lp->init_block_mem; local 682 struct lance_init_block __iomem *ib = lp->init_block_iomem; local 742 struct lance_init_block __iomem *ib = lp->init_block_iomem; local 890 struct lance_init_block __iomem *ib = lp->init_block_iomem; local 903 struct lance_init_block *ib = lp->init_block_mem; local 945 struct lance_init_block __iomem *ib = lp->init_block_iomem; local 950 struct lance_init_block *ib = lp->init_block_mem; local 1129 struct lance_init_block __iomem *ib = lp->init_block_iomem; local 1137 struct lance_init_block *ib = lp->init_block_mem; local 1182 struct lance_init_block __iomem *ib = lp->init_block_iomem; local 1186 struct lance_init_block *ib = lp->init_block_mem; local 1199 struct lance_init_block __iomem *ib = lp->init_block_iomem; local 1205 struct lance_init_block *ib = lp->init_block_mem; local [all...] |
/drivers/s390/char/ |
H A D | fs3270.c | 247 struct idal_buffer *ib; local 255 ib = idal_buffer_alloc(count, 0); 256 if (IS_ERR(ib)) 263 raw3270_request_set_idal(rq, ib); 270 if (idal_buffer_to_user(ib, data, count) != 0) 280 idal_buffer_free(ib); 292 struct idal_buffer *ib; local 299 ib = idal_buffer_alloc(count, 0); 300 if (IS_ERR(ib)) 304 if (idal_buffer_from_user(ib, dat 433 struct idal_buffer *ib; local [all...] |
H A D | raw3270.c | 206 raw3270_request_set_idal(struct raw3270_request *rq, struct idal_buffer *ib) argument 208 rq->ccw.cda = __pa(ib->data); 209 rq->ccw.count = ib->size;
|
/drivers/net/ethernet/chelsio/cxgb/ |
H A D | vsc7326.c | 221 static void run_table(adapter_t *adapter, struct init_table *ib, int len) argument 226 if (ib[i].addr == INITBLOCK_SLEEP) { 227 udelay( ib[i].data ); 228 pr_err("sleep %d us\n",ib[i].data); 230 vsc_write( adapter, ib[i].addr, ib[i].data );
|
/drivers/media/dvb/frontends/ |
H A D | bcm3510.c | 204 u8 ob[olen+2],ib[ilen+2]; local 219 (ret = bcm3510_hab_get_response(st, ib, ilen+2)) < 0) 223 dbufout(ib,ilen+2,deb_hab); 226 memcpy(ibuf,&ib[2],ilen);
|
/drivers/video/matrox/ |
H A D | matroxfb_maven.c | 818 unsigned int ib; local 855 ib = ((0x3C0000 * i - 0x8000)/ hdec + 0x05E7) >> 8; 857 } while (ib < ibmin); 858 if (ib >= m->htotal + 2) { 859 ib = ibmin; 865 m->regs[0x9E] = ib; 866 m->regs[0x9F] = ib >> 8;
|
/drivers/infiniband/hw/ehca/ |
H A D | ehca_classes.h | 270 } ib; member in struct:ehca_mr
|
/drivers/infiniband/core/ |
H A D | cma.c | 131 struct ib_cm_id *ib; member in union:rdma_id_private::__anon909 147 struct ib_sa_multicast *ib; member in union:cma_multicast::__anon910 293 kfree(mc->multicast.ib); 649 if (!id_priv->cm_id.ib || (id_priv->id.qp_type == IB_QPT_UD)) 652 ret = ib_cm_init_qp_attr(id_priv->cm_id.ib, qp_attr, 893 ib_sa_free_multicast(mc->multicast.ib); 924 if (id_priv->cm_id.ib) 925 ib_destroy_cm_id(id_priv->cm_id.ib); 962 ret = ib_send_cm_rtu(id_priv->cm_id.ib, NULL, 0); 969 ib_send_cm_rej(id_priv->cm_id.ib, IB_CM_REJ_CONSUMER_DEFINE [all...] |