/drivers/infiniband/hw/ehca/ |
H A D | ehca_eq.c | 54 struct ehca_eq *eq, 64 spin_lock_init(&eq->spinlock); 65 spin_lock_init(&eq->irq_spinlock); 66 eq->is_initialized = 0; 69 ehca_err(ib_dev, "Invalid EQ type %x. eq=%p", type, eq); 73 ehca_err(ib_dev, "EQ length must not be zero. eq=%p", eq); 78 &eq->pf, 81 &eq 53 ehca_create_eq(struct ehca_shca *shca, struct ehca_eq *eq, const enum ehca_eq_type type, const u32 length) argument 155 ehca_poll_eq(struct ehca_shca *shca, struct ehca_eq *eq) argument 167 ehca_destroy_eq(struct ehca_shca *shca, struct ehca_eq *eq) argument [all...] |
H A D | ehca_irq.c | 506 tasklet_hi_schedule(&shca->eq.interrupt_task); 550 struct ehca_eq *eq = &shca->eq; local 551 struct ehca_eqe_cache_entry *eqe_cache = eq->eqe_cache; 556 spin_lock(&eq->irq_spinlock); 563 shca->ipz_hca_handle, eq->ist); 576 eqe_cache[eqe_cnt].eqe = ehca_poll_eq(shca, eq); 603 ret = hipz_h_eoi(eq->ist); 613 if (eq->eqe_cache[i].cq) 614 reset_eq_pending(eq [all...] |
H A D | ehca_main.c | 771 ret = ehca_create_eq(shca, &shca->eq, EHCA_EQ, eq_size); 871 ret = ehca_destroy_eq(shca, &shca->eq); 913 ret = ehca_destroy_eq(shca, &shca->eq); 957 if (shca->eq.is_initialized) { 958 /* call deadman proc only if eq ptr does not change */ 959 struct ehca_eq *eq = &shca->eq; local 963 spin_lock_irqsave(&eq->spinlock, flags); 964 q_ofs = eq->ipz_queue.current_q_offset; 965 spin_unlock_irqrestore(&eq [all...] |
H A D | hcp_if.c | 730 struct ehca_eq *eq) 734 ret = hcp_galpas_dtor(&eq->galpas); 736 ehca_gen_err("Could not destruct eq->galpas"); 742 eq->ipz_eq_handle.handle, /* r5 */ 729 hipz_h_destroy_eq(const struct ipz_adapter_handle adapter_handle, struct ehca_eq *eq) argument
|
H A D | ehca_classes.h | 121 struct ehca_eq eq; member in struct:ehca_shca
|
/drivers/net/ethernet/mellanox/mlx4/ |
H A D | eq.c | 85 static void eq_set_ci(struct mlx4_eq *eq, int req_not) argument 87 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | 89 eq->doorbell); 94 static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry) argument 96 unsigned long off = (entry & (eq->nent - 1)) * MLX4_EQ_ENTRY_SIZE; 97 return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE; 100 static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq) argument 102 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index); 103 return !!(eqe->owner & 0x80) ^ !!(eq 231 mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq) argument 529 struct mlx4_eq *eq = eq_ptr; local 600 mlx4_get_eq_uar(struct mlx4_dev *dev, struct mlx4_eq *eq) argument 622 mlx4_create_eq(struct mlx4_dev *dev, int nent, u8 intr, struct mlx4_eq *eq) argument 733 mlx4_free_eq(struct mlx4_dev *dev, struct mlx4_eq *eq) argument [all...] |
H A D | resource_tracker.c | 829 enum res_eq_states state, struct res_eq **eq) 866 if (eq) 867 *eq = r; 1878 struct res_eq *eq; local 1884 err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_HW, &eq); 1901 eq->mtt = mtt; 1985 struct res_eq *eq; local 1988 err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_RESERVED, &eq); 1992 err = get_res(dev, slave, eq->mtt->com.res_id, RES_MTT, NULL); 2000 atomic_dec(&eq 828 eq_res_start_move_to(struct mlx4_dev *dev, int slave, int index, enum res_eq_states state, struct res_eq **eq) argument 2085 struct res_eq *eq; local 2997 struct res_eq *eq; local [all...] |
H A D | mlx4.h | 466 struct mlx4_slave_eqe eq[MLX4_MFUNC_MAX_EQES]; member in struct:mlx4_slave_state 469 /* event type to eq number lookup */ 576 struct mlx4_eq *eq; member in struct:mlx4_eq_table
|
/drivers/scsi/be2iscsi/ |
H A D | be_cmds.c | 521 struct be_queue_info *eq, int eq_delay) 526 struct be_dma_mem *q_mem = &eq->dma_mem; 545 __ilog2_u32(eq->len / 256)); 554 eq->id = le16_to_cpu(resp->eq_id); 555 eq->created = true; 591 struct be_queue_info *cq, struct be_queue_info *eq, 621 AMAP_SET_BITS(struct amap_cq_context, eqid, ctxt, eq->id); 520 beiscsi_cmd_eq_create(struct be_ctrl_info *ctrl, struct be_queue_info *eq, int eq_delay) argument 590 beiscsi_cmd_cq_create(struct be_ctrl_info *ctrl, struct be_queue_info *cq, struct be_queue_info *eq, bool sol_evts, bool no_delay, int coalesce_wm) argument
|
H A D | be_main.c | 614 struct be_queue_info *eq; local 621 eq = &pbe_eq->q; 624 eqe = queue_tail_node(eq); 640 queue_tail_inc(eq); 641 eqe = queue_tail_node(eq); 647 hwi_ring_eq_db(phba, eq->id, 1, num_eq_processed, 1, 1); 661 struct be_queue_info *eq; local 668 eq = &pbe_eq->q; 670 eqe = queue_tail_node(eq); 683 queue_tail_inc(eq); 723 struct be_queue_info *eq; local 2672 struct be_queue_info *eq; local 2732 struct be_queue_info *cq, *eq; local 3428 struct be_queue_info *eq; local 3653 struct be_queue_info *eq; local [all...] |
/drivers/block/drbd/ |
H A D | drbd_worker.c | 1019 int ok, eq = 0; local 1045 eq = !memcmp(digest, di->digest, digest_size); 1049 if (eq) { 1141 int ok, eq = 0; local 1165 eq = !memcmp(digest, di->digest, digest_size); 1176 if (!eq) 1182 eq ? ID_IN_SYNC : ID_OUT_OF_SYNC);
|
/drivers/infiniband/hw/mthca/ |
H A D | mthca_eq.c | 173 static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), 189 static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 194 dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8); 199 static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 202 arbel_set_eq_ci(dev, eq, ci); 204 tavor_set_eq_ci(dev, eq, ci); 228 static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry) argument 230 unsigned long off = (entry & (eq 234 next_eqe_sw(struct mthca_eq *eq) argument 260 mthca_eq_int(struct mthca_dev *dev, struct mthca_eq *eq) argument 420 struct mthca_eq *eq = eq_ptr; local 454 struct mthca_eq *eq = eq_ptr; local 465 mthca_create_eq(struct mthca_dev *dev, int nent, u8 intr, struct mthca_eq *eq) argument 588 mthca_free_eq(struct mthca_dev *dev, struct mthca_eq *eq) argument [all...] |
H A D | mthca_dev.h | 231 struct mthca_eq eq[MTHCA_NUM_EQ]; member in struct:mthca_eq_table
|
/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.c | 257 struct ehea_eq *eq; local 259 eq = kzalloc(sizeof(*eq), GFP_KERNEL); 260 if (!eq) { 261 pr_err("no mem for eq\n"); 265 eq->adapter = adapter; 266 eq->attr.type = type; 267 eq->attr.max_nr_of_eqes = max_nr_of_eqes; 268 eq->attr.eqe_gen = eqe_gen; 269 spin_lock_init(&eq 326 ehea_poll_eq(struct ehea_eq *eq) argument 338 ehea_destroy_eq_res(struct ehea_eq *eq, u64 force) argument 357 ehea_destroy_eq(struct ehea_eq *eq) argument [all...] |
H A D | ehea.h | 367 struct ehea_eq *eq; member in struct:ehea_port_res
|
/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 2291 struct sge_eth_rxq *eq = adap->sge.ethrxq; local 2296 for (i = 0; i < adap->sge.ethqsets; i++, eq++, etq++) { 2297 if (eq->rspq.desc) 2298 free_rspq_fl(adap, &eq->rspq, &eq->fl);
|
H A D | cxgb4_main.c | 444 struct sge_eth_txq *eq; local 446 eq = container_of(txq, struct sge_eth_txq, q); 447 netif_tx_wake_queue(eq->txq);
|
/drivers/net/ethernet/emulex/benet/ |
H A D | be_cmds.c | 577 struct be_queue_info *eq, int eq_delay) 581 struct be_dma_mem *q_mem = &eq->dma_mem; 599 __ilog2_u32(eq->len/256)); 609 eq->id = le16_to_cpu(resp->eq_id); 610 eq->created = true; 730 struct be_queue_info *eq, bool no_delay, int coalesce_wm) 760 ctxt, eq->id); 770 AMAP_SET_BITS(struct amap_cq_context_be, eqid, ctxt, eq->id); 576 be_cmd_eq_create(struct be_adapter *adapter, struct be_queue_info *eq, int eq_delay) argument 729 be_cmd_cq_create(struct be_adapter *adapter, struct be_queue_info *cq, struct be_queue_info *eq, bool no_delay, int coalesce_wm) argument
|
H A D | be_main.c | 1673 struct be_queue_info *eq; local 1686 eq = &eqo->q; 1687 rc = be_queue_alloc(adapter, eq, EVNT_Q_LEN, 1692 rc = be_cmd_eq_create(adapter, eq, eqo->cur_eqd); 1778 struct be_queue_info *cq, *eq; local 1799 * one txq share an eq 1801 eq = &adapter->eq_obj[i % adapter->num_evt_qs].q; 1802 status = be_cmd_cq_create(adapter, cq, eq, false, 3); 1844 struct be_queue_info *eq, *cq; local 1863 eq [all...] |
/drivers/net/ethernet/tile/ |
H A D | tilepro.c | 204 lepp_queue_t *eq; member in struct:tile_net_priv 205 /* Protects "eq". */ 531 static unsigned int tile_net_lepp_grab_comps(lepp_queue_t *eq, argument 538 unsigned int comp_head = eq->comp_head; 539 unsigned int comp_busy = eq->comp_busy; 542 comps[n++] = eq->comps[comp_head]; 549 eq->comp_head = comp_head; 562 lepp_queue_t *eq = priv->eq; local 572 eq 1816 lepp_queue_t *eq = priv->eq; local 1943 lepp_queue_t *eq = priv->eq; local [all...] |
/drivers/usb/gadget/ |
H A D | f_fs.c | 1089 char *end, *eq, *comma; local 1098 eq = strchr(opts, '='); 1099 if (unlikely(!eq)) { 1103 *eq = 0; 1106 value = simple_strtoul(eq + 1, &end, 0); 1108 pr_err("%s: invalid value: %s\n", opts, eq + 1); 1113 switch (eq - opts) {
|
/drivers/net/wireless/brcm80211/brcmsmac/phy/ |
H A D | phy_lcn.c | 3493 u8 di0, dq0, ei, eq, fi, fq; local 3511 wlc_lcnphy_get_radio_loft(pi, &ei, &eq, &fi, &fq); 3513 cc.im = (u16) eq; 3516 wlc_lcnphy_get_radio_loft(pi, &ei, &eq, &fi, &fq);
|
/drivers/scsi/lpfc/ |
H A D | lpfc_sli.c | 11539 lpfc_sli4_eq_flush(struct lpfc_hba *phba, struct lpfc_queue *eq) argument 11544 while ((eqe = lpfc_sli4_eq_get(eq))) 11548 lpfc_sli4_eq_release(eq, LPFC_QUEUE_REARM); 11865 * @eq: The queue structure to use to create the event queue. 11868 * This function creates an event queue, as detailed in @eq, on a port, 11871 * The @phba struct is used to send mailbox command to HBA. The @eq struct 11883 lpfc_eq_create(struct lpfc_hba *phba, struct lpfc_queue *eq, uint16_t imax) argument 11895 if (!eq) 11910 eq->page_count); 11918 switch (eq 12002 lpfc_cq_create(struct lpfc_hba *phba, struct lpfc_queue *cq, struct lpfc_queue *eq, uint32_t type, uint32_t subtype) argument 12691 lpfc_eq_destroy(struct lpfc_hba *phba, struct lpfc_queue *eq) argument [all...] |