Searched refs:chp (Results 1 - 21 of 21) sorted by relevance

/drivers/s390/cio/
H A Dchp.c27 #include "chp.h"
137 struct channel_path *chp; local
141 chp = to_channelpath(device);
142 if (!chp->cmg_chars)
146 chp->cmg_chars, sizeof(struct cmg_chars));
184 struct channel_path *chp; local
190 chp = to_channelpath(device);
191 css = to_css(chp->dev.parent);
198 chp_measurement_copy_block((struct cmg_entry *)buf, css, chp->chpid);
212 void chp_remove_cmg_attr(struct channel_path *chp) argument
218 chp_add_cmg_attr(struct channel_path *chp) argument
237 struct channel_path *chp = to_channelpath(dev); local
316 struct channel_path *chp = to_channelpath(dev); local
330 struct channel_path *chp = to_channelpath(dev); local
344 struct channel_path *chp = to_channelpath(dev); local
358 struct channel_path *chp = to_channelpath(dev); local
375 struct channel_path *chp = to_channelpath(dev); local
422 chp_update_desc(struct channel_path *chp) argument
444 struct channel_path *chp; local
514 struct channel_path *chp; local
[all...]
H A Dchp.h64 void chp_remove_cmg_attr(struct channel_path *chp);
65 int chp_add_cmg_attr(struct channel_path *chp);
66 int chp_update_desc(struct channel_path *chp);
H A DMakefile5 obj-y += airq.o blacklist.o chsc.o cio.o css.o chp.o idset.o isc.o \
H A Dccwreq.c329 int rc = -ENODEV, chp; local
334 for (chp = 0; chp < 8; chp++) {
335 if ((0x80 >> chp) & sch->schib.pmcw.lpum)
342 sch->schib.pmcw.chpid[chp]);
H A Dchsc.c29 #include "chp.h"
397 struct channel_path *chp; local
414 chp = chpid_to_chp(chpid);
415 if (!chp) {
419 mutex_lock(&chp->lock);
420 chp_update_desc(chp);
421 mutex_unlock(&chp->lock);
667 struct channel_path *chp = chpid_to_chp(chpid); local
676 chp_update_desc(chp);
886 chsc_initialize_cmg_chars(struct channel_path *chp, u argument
902 chsc_get_channel_measurement_chars(struct channel_path *chp) argument
[all...]
H A Ddevice_ops.c26 #include "chp.h"
716 struct channel_path *chp; local
731 chp = chpid_to_chp(chpid);
732 if (!chp)
735 mutex_lock(&chp->lock);
736 if (!chp->desc_fmt1.f) {
737 mutex_unlock(&chp->lock);
740 if (!chp->desc_fmt1.r)
742 mdc = mdc ? min_t(int, mdc, chp->desc_fmt1.mdc) :
743 chp
[all...]
H A Ddevice_fsm.c24 #include "chp.h"
476 int chp, mask; local
478 for (chp = 0, mask = 0x80; chp < 8; chp++, mask >>= 1) {
479 path_event[chp] = PE_NONE;
481 path_event[chp] |= PE_PATH_GONE;
483 path_event[chp] |= PE_PATH_AVAILABLE;
485 path_event[chp] |= PE_PATHGROUP_ESTABLISHED;
H A Dchsc.h139 int chsc_get_channel_measurement_chars(struct channel_path *chp);
H A Dcss.c32 #include "chp.h"
1083 struct channel_path *chp; local
1088 chp = chpid_to_chp(chpid);
1089 if (chp)
1090 chp_update_desc(chp);
H A Ddevice.c32 #include "chp.h"
223 int chp; local
226 for (chp = 0; chp < 8; chp++) {
227 mask = 0x80 >> chp;
229 ret += sprintf(buf + ret, "%02x ", ssd->chpid[chp].id);
H A Dchsc_sch.c760 chpid_parm->cssid = ccl->req.chpid.chp.cssid;
761 chpid_parm->chpid = ccl->req.chpid.chp.id;
/drivers/infiniband/hw/cxgb4/
H A Dev.c87 static void post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, argument
105 event.device = chp->ibcq.device;
107 event.element.cq = &chp->ibcq;
113 spin_lock_irqsave(&chp->comp_handler_lock, flag);
114 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context);
115 spin_unlock_irqrestore(&chp->comp_handler_lock, flag);
120 struct c4iw_cq *chp; local
141 chp
225 struct c4iw_cq *chp; local
[all...]
H A Dcq.c235 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq); local
236 struct t4_cq *cq = &chp->cq;
334 void c4iw_flush_hw_cq(struct c4iw_cq *chp) argument
341 PDBG("%s cqid 0x%x\n", __func__, chp->cq.cqid);
342 ret = t4_next_hw_cqe(&chp->cq, &hw_cqe);
350 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe));
399 flush_completed_wrs(&qhp->wq, &chp->cq);
401 swcqe = &chp->cq.sw_queue[chp->cq.sw_pidx];
404 t4_swcq_produce(&chp
677 c4iw_poll_cq_one(struct c4iw_cq *chp, struct ib_wc *wc) argument
823 struct c4iw_cq *chp; local
844 struct c4iw_cq *chp; local
867 struct c4iw_cq *chp; local
997 struct c4iw_cq *chp; local
[all...]
H A Diw_cxgb4.h1011 void c4iw_flush_hw_cq(struct c4iw_cq *chp);
/drivers/infiniband/hw/cxgb3/
H A Diwch_ev.c41 static void post_qp_event(struct iwch_dev *rnicp, struct iwch_cq *chp, argument
89 event.device = chp->ibcq.device;
91 event.element.cq = &chp->ibcq;
98 spin_lock_irqsave(&chp->comp_handler_lock, flag);
99 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context);
100 spin_unlock_irqrestore(&chp->comp_handler_lock, flag);
110 struct iwch_cq *chp; local
117 chp
[all...]
H A Diwch_cq.c44 static int iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, argument
55 rd_cqe = cxio_next_cqe(&chp->cq);
67 ret = cxio_poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie,
69 if (t3a_device(chp->rhp) && credit) {
71 credit, chp->cq.cqid);
72 cxio_hal_cq_op(&rhp->rdev, &chp->cq, CQ_CREDIT_UPDATE, credit);
198 struct iwch_cq *chp; local
203 chp = to_iwch_cq(ibcq);
204 rhp = chp->rhp;
206 spin_lock_irqsave(&chp
[all...]
H A Diwch_provider.c127 struct iwch_cq *chp; local
130 chp = to_iwch_cq(ib_cq);
132 remove_handle(chp->rhp, &chp->rhp->cqidr, chp->cq.cqid);
133 atomic_dec(&chp->refcnt);
134 wait_event(chp->wait, !atomic_read(&chp->refcnt));
136 cxio_destroy_cq(&chp->rhp->rdev, &chp
146 struct iwch_cq *chp; local
248 struct iwch_cq *chp = to_iwch_cq(cq); local
316 struct iwch_cq *chp; local
[all...]
/drivers/media/usb/pvrusb2/
H A Dpvrusb2-dvb.c400 static void pvr2_dvb_internal_check(struct pvr2_channel *chp) argument
403 adap = container_of(chp, struct pvr2_dvb_adapter, channel);
H A Dpvrusb2-sysfs.c759 static void pvr2_sysfs_internal_check(struct pvr2_channel *chp) argument
762 sfp = container_of(chp,struct pvr2_sysfs,channel);
H A Dpvrusb2-v4l2.c905 static void pvr2_v4l2_internal_check(struct pvr2_channel *chp) argument
908 vp = container_of(chp,struct pvr2_v4l2,channel);
/drivers/s390/block/
H A Ddasd.c3586 int chp; local
3593 for (chp = 0; chp < 8; chp++) {
3594 eventlpm = 0x80 >> chp;
3595 if (path_event[chp] & PE_PATH_GONE) {
3611 if (path_event[chp] & PE_PATH_AVAILABLE) {
3618 if (path_event[chp] & PE_PATHGROUP_ESTABLISHED) {

Completed in 202 milliseconds