/drivers/infiniband/hw/mthca/ |
H A D | mthca_allocator.c | 160 int mthca_array_init(struct mthca_array *array, int nent) argument 162 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; 177 void mthca_array_cleanup(struct mthca_array *array, int nent) argument 181 for (i = 0; i < (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; ++i)
|
H A D | mthca_eq.c | 184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), 230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; 466 int nent, 479 eq->nent = roundup_pow_of_two(max(nent, 2)); 480 npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE; 511 for (i = 0; i < eq->nent; ++i) 535 eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24); 560 eq->eqn, eq->nent); 593 int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZ 465 mthca_create_eq(struct mthca_dev *dev, int nent, u8 intr, struct mthca_eq *eq) argument [all...] |
H A D | mthca_cq.c | 353 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent) argument 358 ret = mthca_buf_alloc(dev, nent * MTHCA_CQ_ENTRY_SIZE, 365 for (i = 0; i < nent; ++i) 776 int mthca_init_cq(struct mthca_dev *dev, int nent, argument 784 cq->ibcq.cqe = nent - 1; 820 err = mthca_alloc_cq_buf(dev, &cq->buf, nent); 834 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24);
|
H A D | mthca_dev.h | 422 int mthca_array_init(struct mthca_array *array, int nent); 423 void mthca_array_cleanup(struct mthca_array *array, int nent); 498 int mthca_init_cq(struct mthca_dev *dev, int nent, 509 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent);
|
H A D | mthca_provider.h | 113 int nent; member in struct:mthca_eq
|
H A D | mthca_cmd.c | 662 int nent = 0; local 692 pages[nent * 2] = cpu_to_be64(virt); 696 pages[nent * 2 + 1] = 702 if (++nent == MTHCA_MAILBOX_SIZE / 16) { 703 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, 707 nent = 0; 712 if (nent) 713 err = mthca_cmd(dev, mailbox->dma, nent, 0, op,
|
H A D | mthca_provider.c | 651 int nent; local 686 for (nent = 1; nent <= entries; nent <<= 1) 689 err = mthca_init_cq(to_mdev(ibdev), nent,
|
/drivers/iommu/ |
H A D | omap-iommu.c | 686 int nent = 1; local 697 nent *= 16; 701 bytes *= nent; 702 memset(iopte, 0, nent * sizeof(*iopte)); 703 flush_iopte_range(iopte, iopte + (nent - 1) * sizeof(*iopte)); 714 nent = 1; /* for the next L1 entry */ 718 nent *= 16; 722 bytes *= nent; 724 memset(iopgd, 0, nent * sizeof(*iopgd)); 725 flush_iopgd_range(iopgd, iopgd + (nent [all...] |
/drivers/net/ethernet/mellanox/mlx4/ |
H A D | eq.c | 96 unsigned long off = (entry & (eq->nent - 1)) * MLX4_EQ_ENTRY_SIZE; 103 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; 467 "nent=0x%x, slave=%x, ownership=%s\n", 469 eq->cons_index, eqe->owner, eq->nent, 472 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); 480 "index %u. owner=%x, nent=0x%x, slave=%x, " 483 eq->cons_index, eqe->owner, eq->nent, 486 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); 622 static int mlx4_create_eq(struct mlx4_dev *dev, int nent, argument 636 eq->nent [all...] |
H A D | cq.c | 243 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt, argument 279 cq_context->logsize_usrpage = cpu_to_be32((ilog2(nent) << 24) | uar->index);
|
H A D | fw.c | 694 int nent = 0; local 725 pages[nent * 2] = cpu_to_be64(virt); 729 pages[nent * 2 + 1] = 735 if (++nent == MLX4_MAILBOX_SIZE / 16) { 736 err = mlx4_cmd(dev, mailbox->dma, nent, 0, op, 741 nent = 0; 746 if (nent) 747 err = mlx4_cmd(dev, mailbox->dma, nent, 0, op,
|
H A D | main.c | 1295 int nent; local 1300 nent = dev->caps.max_counters; 1301 return mlx4_bitmap_init(&priv->counters_bitmap, nent, nent - 1, 0, 0);
|
H A D | mlx4.h | 383 int nent; member in struct:mlx4_eq
|
/drivers/tty/serial/ |
H A D | pch_uart.c | 248 int nent; member in struct:eg20t_port 768 for (i = 0; i < priv->nent; i++, sg++) { 774 dma_unmap_sg(port->dev, sg, priv->nent, DMA_TO_DEVICE); 776 priv->nent = 0; 916 int nent; local 991 nent = dma_map_sg(port->dev, sg, num, DMA_TO_DEVICE); 992 if (!nent) { 996 priv->nent = nent; 998 for (i = 0; i < nent; [all...] |
H A D | sh-sci.c | 1606 int nent; local 1632 nent = dma_map_sg(port->dev, &s->sg_tx, 1, DMA_TO_DEVICE); 1633 if (!nent) 1640 s->sg_len_tx = nent;
|
/drivers/infiniband/hw/qib/ |
H A D | qib_pcie.c | 270 int qib_pcie_params(struct qib_devdata *dd, u32 minw, u32 *nent, argument 286 if (nent && *nent && pos) { 287 qib_msix_setup(dd, pos, nent, entry);
|
/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 98 static int mlx4_ib_alloc_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int nent) argument 102 err = mlx4_buf_alloc(dev->dev, nent * sizeof(struct mlx4_cqe), 123 mlx4_buf_free(dev->dev, nent * sizeof(struct mlx4_cqe),
|
/drivers/spi/ |
H A D | spi-topcliff-pch.c | 129 int nent; member in struct:pch_spi_dma_ctrl 863 dma_sync_sg_for_cpu(&data->master->dev, dma->sg_rx_p, dma->nent, 866 dma_sync_sg_for_cpu(&data->master->dev, dma->sg_tx_p, dma->nent, 1113 dma->nent = num; 1172 dma->nent = num;
|