/drivers/infiniband/hw/amso1100/ |
H A D | c2_pd.c | 43 int c2_pd_alloc(struct c2_dev *c2dev, int privileged, struct c2_pd *pd) argument 55 pd->pd_id = obj; 66 void c2_pd_free(struct c2_dev *c2dev, struct c2_pd *pd) argument 69 __clear_bit(pd->pd_id, c2dev->pd_table.table);
|
/drivers/infiniband/hw/mthca/ |
H A D | mthca_pd.c | 39 int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd) argument 43 pd->privileged = privileged; 45 atomic_set(&pd->sqp_count, 0); 46 pd->pd_num = mthca_alloc(&dev->pd_table.alloc); 47 if (pd->pd_num == -1) 51 err = mthca_mr_alloc_notrans(dev, pd->pd_num, 54 &pd->ntmr); 56 mthca_free(&dev->pd_table.alloc, pd->pd_num); 62 void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd) argument 64 if (pd [all...] |
H A D | mthca_allocator.c | 195 union mthca_buf *buf, int *is_direct, struct mthca_pd *pd, 260 err = mthca_mr_alloc_phys(dev, pd->pd_num, 194 mthca_buf_alloc(struct mthca_dev *dev, int size, int max_direct, union mthca_buf *buf, int *is_direct, struct mthca_pd *pd, int hca_write, struct mthca_mr *mr) argument
|
/drivers/staging/ozwpan/ |
H A D | ozusbsvc.h | 18 struct oz_pd *pd; member in struct:oz_usb_ctx 25 int oz_usb_start(struct oz_pd *pd, int resume); 26 void oz_usb_stop(struct oz_pd *pd, int pause); 27 void oz_usb_rx(struct oz_pd *pd, struct oz_elt *elt); 28 int oz_usb_heartbeat(struct oz_pd *pd); 29 void oz_usb_farewell(struct oz_pd *pd, u8 ep_num, u8 *data, u8 len);
|
H A D | ozusbsvc.c | 53 int oz_usb_start(struct oz_pd *pd, int resume) argument 71 usb_ctx->pd = pd; 77 spin_lock_bh(&pd->app_lock[OZ_APPID_USB-1]); 78 old_ctx = pd->app_ctx[OZ_APPID_USB-1]; 80 pd->app_ctx[OZ_APPID_USB-1] = usb_ctx; 81 oz_usb_get(pd->app_ctx[OZ_APPID_USB-1]); 82 spin_unlock_bh(&pd->app_lock[OZ_APPID_USB-1]); 91 oz_pd_get(pd); 103 spin_lock_bh(&pd 117 oz_usb_stop(struct oz_pd *pd, int pause) argument 176 oz_usb_heartbeat(struct oz_pd *pd) argument 202 struct oz_pd *pd = usb_ctx->pd; local 223 struct oz_pd *pd = usb_ctx->pd; local [all...] |
/drivers/infiniband/hw/ehca/ |
H A D | ehca_pd.c | 51 struct ehca_pd *pd; local 54 pd = kmem_cache_zalloc(pd_cache, GFP_KERNEL); 55 if (!pd) { 62 INIT_LIST_HEAD(&pd->free[i]); 63 INIT_LIST_HEAD(&pd->full[i]); 65 mutex_init(&pd->lock); 78 pd->fw_pd.value = shca->pd->fw_pd.value; 80 pd->fw_pd.value = (u64)pd; 85 ehca_dealloc_pd(struct ib_pd *pd) argument [all...] |
H A D | ehca_av.c | 88 struct ib_ah *ehca_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr) argument 92 struct ehca_shca *shca = container_of(pd->device, struct ehca_shca, 97 ehca_err(pd->device, "Out of memory pd=%p ah_attr=%p", 98 pd, ah_attr); 132 rc = ehca_query_port(pd->device, ah_attr->port_num, 136 ehca_err(pd->device, "Invalid port number " 138 "pd=%p ah_attr=%p", rc, pd, ah_attr); 142 rc = ehca_query_gid(pd [all...] |
H A D | ipz_pt_fn.c | 127 static int alloc_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument 133 mutex_lock(&pd->lock); 135 if (!list_empty(&pd->free[order])) 136 page = list_entry(pd->free[order].next, 149 list_add(&page->list, &pd->free[order]); 157 list_move(&page->list, &pd->full[order]); 159 mutex_unlock(&pd->lock); 167 ehca_err(pd->ib_pd.device, "failed to allocate small queue page"); 168 mutex_unlock(&pd->lock); 172 static void free_small_queue_page(struct ipz_queue *queue, struct ehca_pd *pd) argument 204 ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue, const u32 nr_of_pages, const u32 pagesize, const u32 qe_size, const u32 nr_of_sg, int is_small) argument 255 ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue) argument [all...] |
/drivers/leds/ |
H A D | leds-s3c24xx.c | 47 struct s3c24xx_led_platdata *pd = led->pdata; local 52 s3c2410_gpio_setpin(pd->gpio, (value ? 1 : 0) ^ 53 (pd->flags & S3C24XX_LEDF_ACTLOW)); 55 if (pd->flags & S3C24XX_LEDF_TRISTATE) 56 s3c2410_gpio_cfgpin(pd->gpio,
|
/drivers/base/power/ |
H A D | domain_governor.c | 99 * @pd: PM domain to check. 103 static bool default_power_down_ok(struct dev_pm_domain *pd) argument 105 struct generic_pm_domain *genpd = pd_to_genpd(pd);
|
/drivers/i2c/busses/ |
H A D | i2c-simtec.c | 49 struct simtec_i2c_data *pd = pw; local 50 writeb(CMD_SET_SDA | (state ? STATE_SDA : 0), pd->reg); 55 struct simtec_i2c_data *pd = pw; local 56 writeb(CMD_SET_SCL | (state ? STATE_SCL : 0), pd->reg); 61 struct simtec_i2c_data *pd = pw; local 62 return readb(pd->reg) & STATE_SDA ? 1 : 0; 67 struct simtec_i2c_data *pd = pw; local 68 return readb(pd->reg) & STATE_SCL ? 1 : 0; 75 struct simtec_i2c_data *pd; local 80 pd 147 struct simtec_i2c_data *pd = platform_get_drvdata(dev); local [all...] |
H A D | i2c-pca-isa.c | 50 static void pca_isa_writebyte(void *pd, int reg, int val) argument 60 static int pca_isa_readbyte(void *pd, int reg) argument 72 static int pca_isa_waitforcompletion(void *pd) argument 79 pca_isa_readbyte(pd, I2C_PCA_CON) 86 if (pca_isa_readbyte(pd, I2C_PCA_CON) 96 static void pca_isa_resetchip(void *pd) argument
|
H A D | i2c-pca-platform.c | 43 static int i2c_pca_pf_readbyte8(void *pd, int reg) argument 45 struct i2c_pca_pf_data *i2c = pd; 49 static int i2c_pca_pf_readbyte16(void *pd, int reg) argument 51 struct i2c_pca_pf_data *i2c = pd; 55 static int i2c_pca_pf_readbyte32(void *pd, int reg) argument 57 struct i2c_pca_pf_data *i2c = pd; 61 static void i2c_pca_pf_writebyte8(void *pd, int reg, int val) argument 63 struct i2c_pca_pf_data *i2c = pd; 67 static void i2c_pca_pf_writebyte16(void *pd, int reg, int val) argument 69 struct i2c_pca_pf_data *i2c = pd; 73 i2c_pca_pf_writebyte32(void *pd, int reg, int val) argument 80 i2c_pca_pf_waitforcompletion(void *pd) argument 105 i2c_pca_pf_dummyreset(void *pd) argument 112 i2c_pca_pf_resetchip(void *pd) argument [all...] |
/drivers/infiniband/hw/ipath/ |
H A D | ipath_keys.c | 136 struct ipath_pd *pd = to_ipd(qp->ibqp.pd); local 138 if (pd->user) { 151 qp->ibqp.pd != mr->pd)) { 216 struct ipath_pd *pd = to_ipd(qp->ibqp.pd); local 218 if (pd->user) { 234 qp->ibqp.pd != mr->pd)) { [all...] |
H A D | ipath_stats.c | 144 struct ipath_portdata *pd = dd->ipath_pd[0]; local 150 if (pd->port_hdrqfull != dd->ipath_p0_hdrqfull) { 152 pd->port_hdrqfull - 154 dd->ipath_p0_hdrqfull = pd->port_hdrqfull; 186 hdrqtail = ipath_get_hdrqtail(pd); 187 if (pd->port_head != hdrqtail) { 193 pd->port_head, hdrqtail, 197 dd->ipath_rhdrhead_intr_off, pd->port_port); 337 struct ipath_portdata *pd = dd->ipath_pd[i]; local 339 if (pd [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | ah.c | 61 static struct ib_ah *create_ib_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr, argument 64 struct mlx4_dev *dev = to_mdev(pd->device)->dev; 66 ah->av.ib.port_pd = cpu_to_be32(to_mpd(pd)->pdn | (ah_attr->port_num << 24)); 90 static struct ib_ah *create_iboe_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr, argument 93 struct mlx4_ib_dev *ibdev = to_mdev(pd->device); 106 err = ib_get_cached_gid(pd->device, ah_attr->port_num, ah_attr->grh.sgid_index, &sgid); 112 ah->av.eth.port_pd = cpu_to_be32(to_mpd(pd)->pdn | (ah_attr->port_num << 24)); 134 struct ib_ah *mlx4_ib_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr) argument 143 if (rdma_port_get_link_layer(pd->device, ah_attr->port_num) == IB_LINK_LAYER_ETHERNET) { 155 ret = create_iboe_ah(pd, ah_att [all...] |
H A D | srq.c | 71 struct ib_srq *mlx4_ib_create_srq(struct ib_pd *pd, argument 75 struct mlx4_ib_dev *dev = to_mdev(pd->device); 108 if (pd->uobject) { 116 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, 132 err = mlx4_ib_db_map_user(to_mucontext(pd->uobject->context), 184 err = mlx4_srq_alloc(dev->dev, to_mpd(pd)->pdn, cqn, xrcdn, &srq->mtt, 192 if (pd->uobject) 203 if (pd->uobject) 204 mlx4_ib_db_unmap_user(to_mucontext(pd->uobject->context), &srq->db); 212 if (pd [all...] |
/drivers/infiniband/hw/qib/ |
H A D | qib_keys.c | 133 int qib_lkey_ok(struct qib_lkey_table *rkt, struct qib_pd *pd, argument 147 struct qib_ibdev *dev = to_idev(pd->ibpd.device); 149 if (pd->user) 166 mr->pd != &pd->ibpd)) 241 struct qib_pd *pd = to_ipd(qp->ibqp.pd); local 242 struct qib_ibdev *dev = to_idev(pd->ibpd.device); 244 if (pd->user) 261 if (unlikely(mr == NULL || mr->lkey != rkey || qp->ibqp.pd ! 315 struct qib_pd *pd = to_ipd(qp->ibqp.pd); local [all...] |
/drivers/pcmcia/ |
H A D | bcm63xx_pcmcia.h | 25 struct bcm63xx_pcmcia_platform_data *pd; member in struct:bcm63xx_pcmcia_socket
|
/drivers/usb/otg/ |
H A D | nop-usb-xceiv.c | 40 static struct platform_device *pd; variable in typeref:struct:platform_device 44 if (pd) 46 pd = platform_device_register_simple("nop_usb_xceiv", -1, NULL, 0); 47 if (!pd) { 56 platform_device_unregister(pd); 57 pd = NULL;
|
/drivers/isdn/hysdn/ |
H A D | hysdn_procconf.c | 232 struct proc_dir_entry *pd; local 240 pd = card->procconf; 241 if (pd == PDE(ino)) 320 struct proc_dir_entry *pd; local 326 pd = card->procconf; 327 if (pd == PDE(ino))
|
/drivers/spi/ |
H A D | spi-butterfly.c | 72 struct pardevice *pd; member in struct:butterfly 199 struct pardevice *pd; local 236 pd = parport_register_device(p, "spi_butterfly", 239 if (!pd) { 243 pp->pd = pd; 245 status = parport_claim(pd); 301 parport_release(pp->pd); 303 parport_unregister_device(pd); 331 parport_release(pp->pd); [all...] |
/drivers/usb/host/ |
H A D | ehci-orion.c | 196 struct orion_ehci_data *pd = pdev->dev.platform_data; local 270 switch (pd->phy_version) {
|
/drivers/ata/ |
H A D | pata_pxa.c | 63 struct pata_pxa_data *pd = qc->ap->private_data; local 74 pd->dma_desc[pd->dma_desc_id].ddadr = pd->dma_desc_addr + 75 ((pd->dma_desc_id + 1) * sizeof(struct pxa_dma_desc)); 77 pd->dma_desc[pd->dma_desc_id].dcmd = DCMD_BURST32 | 81 pd->dma_desc[pd->dma_desc_id].dsadr = cpu_addr; 82 pd 108 struct pata_pxa_data *pd = qc->ap->private_data; local 147 struct pata_pxa_data *pd = qc->ap->private_data; local 157 struct pata_pxa_data *pd = qc->ap->private_data; local 172 struct pata_pxa_data *pd = ap->private_data; local 223 struct pata_pxa_data *pd = ap->private_data; local [all...] |
/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_hdmi.c | 267 struct exynos_drm_common_hdmi_pd *pd; local 271 pd = pdev->dev.platform_data; 273 if (!pd) { 278 if (!pd->hdmi_dev) { 283 if (!pd->mixer_dev) { 291 to_context(pd->hdmi_dev); 300 to_context(pd->mixer_dev);
|