/drivers/s390/cio/ |
H A D | chp.c | 27 #include "chp.h" 137 struct channel_path *chp; local 141 chp = to_channelpath(device); 142 if (!chp->cmg_chars) 146 chp->cmg_chars, sizeof(struct cmg_chars)); 184 struct channel_path *chp; local 190 chp = to_channelpath(device); 191 css = to_css(chp->dev.parent); 198 chp_measurement_copy_block((struct cmg_entry *)buf, css, chp->chpid); 212 void chp_remove_cmg_attr(struct channel_path *chp) argument 218 chp_add_cmg_attr(struct channel_path *chp) argument 237 struct channel_path *chp = to_channelpath(dev); local 316 struct channel_path *chp = to_channelpath(dev); local 330 struct channel_path *chp = to_channelpath(dev); local 344 struct channel_path *chp = to_channelpath(dev); local 358 struct channel_path *chp = to_channelpath(dev); local 375 struct channel_path *chp = to_channelpath(dev); local 422 chp_update_desc(struct channel_path *chp) argument 444 struct channel_path *chp; local 514 struct channel_path *chp; local [all...] |
H A D | chp.h | 64 void chp_remove_cmg_attr(struct channel_path *chp); 65 int chp_add_cmg_attr(struct channel_path *chp); 66 int chp_update_desc(struct channel_path *chp);
|
H A D | Makefile | 5 obj-y += airq.o blacklist.o chsc.o cio.o css.o chp.o idset.o isc.o \
|
H A D | ccwreq.c | 329 int rc = -ENODEV, chp; local 334 for (chp = 0; chp < 8; chp++) { 335 if ((0x80 >> chp) & sch->schib.pmcw.lpum) 342 sch->schib.pmcw.chpid[chp]);
|
H A D | chsc.c | 29 #include "chp.h" 397 struct channel_path *chp; local 414 chp = chpid_to_chp(chpid); 415 if (!chp) { 419 mutex_lock(&chp->lock); 420 chp_update_desc(chp); 421 mutex_unlock(&chp->lock); 667 struct channel_path *chp = chpid_to_chp(chpid); local 676 chp_update_desc(chp); 886 chsc_initialize_cmg_chars(struct channel_path *chp, u argument 902 chsc_get_channel_measurement_chars(struct channel_path *chp) argument [all...] |
H A D | device_ops.c | 26 #include "chp.h" 716 struct channel_path *chp; local 731 chp = chpid_to_chp(chpid); 732 if (!chp) 735 mutex_lock(&chp->lock); 736 if (!chp->desc_fmt1.f) { 737 mutex_unlock(&chp->lock); 740 if (!chp->desc_fmt1.r) 742 mdc = mdc ? min_t(int, mdc, chp->desc_fmt1.mdc) : 743 chp [all...] |
H A D | device_fsm.c | 24 #include "chp.h" 476 int chp, mask; local 478 for (chp = 0, mask = 0x80; chp < 8; chp++, mask >>= 1) { 479 path_event[chp] = PE_NONE; 481 path_event[chp] |= PE_PATH_GONE; 483 path_event[chp] |= PE_PATH_AVAILABLE; 485 path_event[chp] |= PE_PATHGROUP_ESTABLISHED;
|
H A D | chsc.h | 139 int chsc_get_channel_measurement_chars(struct channel_path *chp);
|
H A D | css.c | 32 #include "chp.h" 1083 struct channel_path *chp; local 1088 chp = chpid_to_chp(chpid); 1089 if (chp) 1090 chp_update_desc(chp);
|
H A D | device.c | 32 #include "chp.h" 223 int chp; local 226 for (chp = 0; chp < 8; chp++) { 227 mask = 0x80 >> chp; 229 ret += sprintf(buf + ret, "%02x ", ssd->chpid[chp].id);
|
H A D | chsc_sch.c | 760 chpid_parm->cssid = ccl->req.chpid.chp.cssid; 761 chpid_parm->chpid = ccl->req.chpid.chp.id;
|
/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 87 static void post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, argument 105 event.device = chp->ibcq.device; 107 event.element.cq = &chp->ibcq; 113 spin_lock_irqsave(&chp->comp_handler_lock, flag); 114 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); 115 spin_unlock_irqrestore(&chp->comp_handler_lock, flag); 120 struct c4iw_cq *chp; local 141 chp 225 struct c4iw_cq *chp; local [all...] |
H A D | cq.c | 235 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq); local 236 struct t4_cq *cq = &chp->cq; 334 void c4iw_flush_hw_cq(struct c4iw_cq *chp) argument 341 PDBG("%s cqid 0x%x\n", __func__, chp->cq.cqid); 342 ret = t4_next_hw_cqe(&chp->cq, &hw_cqe); 350 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); 399 flush_completed_wrs(&qhp->wq, &chp->cq); 401 swcqe = &chp->cq.sw_queue[chp->cq.sw_pidx]; 404 t4_swcq_produce(&chp 677 c4iw_poll_cq_one(struct c4iw_cq *chp, struct ib_wc *wc) argument 823 struct c4iw_cq *chp; local 844 struct c4iw_cq *chp; local 867 struct c4iw_cq *chp; local 997 struct c4iw_cq *chp; local [all...] |
H A D | iw_cxgb4.h | 1011 void c4iw_flush_hw_cq(struct c4iw_cq *chp);
|
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_ev.c | 41 static void post_qp_event(struct iwch_dev *rnicp, struct iwch_cq *chp, argument 89 event.device = chp->ibcq.device; 91 event.element.cq = &chp->ibcq; 98 spin_lock_irqsave(&chp->comp_handler_lock, flag); 99 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); 100 spin_unlock_irqrestore(&chp->comp_handler_lock, flag); 110 struct iwch_cq *chp; local 117 chp [all...] |
H A D | iwch_cq.c | 44 static int iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, argument 55 rd_cqe = cxio_next_cqe(&chp->cq); 67 ret = cxio_poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie, 69 if (t3a_device(chp->rhp) && credit) { 71 credit, chp->cq.cqid); 72 cxio_hal_cq_op(&rhp->rdev, &chp->cq, CQ_CREDIT_UPDATE, credit); 198 struct iwch_cq *chp; local 203 chp = to_iwch_cq(ibcq); 204 rhp = chp->rhp; 206 spin_lock_irqsave(&chp [all...] |
H A D | iwch_provider.c | 127 struct iwch_cq *chp; local 130 chp = to_iwch_cq(ib_cq); 132 remove_handle(chp->rhp, &chp->rhp->cqidr, chp->cq.cqid); 133 atomic_dec(&chp->refcnt); 134 wait_event(chp->wait, !atomic_read(&chp->refcnt)); 136 cxio_destroy_cq(&chp->rhp->rdev, &chp 146 struct iwch_cq *chp; local 248 struct iwch_cq *chp = to_iwch_cq(cq); local 316 struct iwch_cq *chp; local [all...] |
/drivers/media/usb/pvrusb2/ |
H A D | pvrusb2-dvb.c | 400 static void pvr2_dvb_internal_check(struct pvr2_channel *chp) argument 403 adap = container_of(chp, struct pvr2_dvb_adapter, channel);
|
H A D | pvrusb2-sysfs.c | 759 static void pvr2_sysfs_internal_check(struct pvr2_channel *chp) argument 762 sfp = container_of(chp,struct pvr2_sysfs,channel);
|
H A D | pvrusb2-v4l2.c | 905 static void pvr2_v4l2_internal_check(struct pvr2_channel *chp) argument 908 vp = container_of(chp,struct pvr2_v4l2,channel);
|
/drivers/s390/block/ |
H A D | dasd.c | 3586 int chp; local 3593 for (chp = 0; chp < 8; chp++) { 3594 eventlpm = 0x80 >> chp; 3595 if (path_event[chp] & PE_PATH_GONE) { 3611 if (path_event[chp] & PE_PATH_AVAILABLE) { 3618 if (path_event[chp] & PE_PATHGROUP_ESTABLISHED) {
|