Searched refs:entries (Results 1 - 25 of 251) sorted by relevance

1234567891011

/drivers/misc/vmw_vmci/
H A Dvmci_handle_array.c66 array->entries[array->size] = handle;
80 if (vmci_handle_is_equal(array->entries[i], entry_handle)) {
81 handle = array->entries[i];
83 array->entries[i] = array->entries[array->size];
84 array->entries[array->size] = VMCI_INVALID_HANDLE;
101 handle = array->entries[array->size];
102 array->entries[array->size] = VMCI_INVALID_HANDLE;
117 return array->entries[index];
126 if (vmci_handle_is_equal(array->entries[
[all...]
H A Dvmci_handle_array.h28 struct vmci_handle entries[]; member in struct:vmci_handle_arr
H A Dvmci_resource.c30 struct hlist_head entries[VMCI_RESOURCE_HASH_BUCKETS]; member in struct:vmci_hash_table
53 &vmci_resource_table.entries[idx], node) {
134 hlist_add_head_rcu(&resource->node, &vmci_resource_table.entries[idx]);
152 hlist_for_each_entry(r, &vmci_resource_table.entries[idx], node) {
/drivers/gpu/drm/nouveau/core/subdev/bios/
H A Dbit.c34 u8 entries = nv_ro08(bios, bios->bit_offset + 10); local
36 while (entries--) {
/drivers/crypto/qat/qat_dh895xcc/
H A Dadf_isr.c71 pci_dev_info->msix_entries.entries[i].entry = i;
74 pci_dev_info->msix_entries.entries,
108 struct msix_entry *msixe = pci_dev_info->msix_entries.entries;
150 struct msix_entry *msixe = pci_dev_info->msix_entries.entries;
166 struct msix_entry *entries; local
170 entries = kzalloc_node(msix_num_entries * sizeof(*entries),
172 if (!entries)
177 kfree(entries);
185 accel_dev->accel_pci_dev.msix_entries.entries
[all...]
/drivers/scsi/arm/
H A Dmsgqueue.h31 struct msgqueue_entry entries[NR_MESSAGES]; member in struct:__anon5178
H A Dmsgqueue.c59 msgq->free = &msgq->entries[0];
62 msgq->entries[i].next = &msgq->entries[i + 1];
64 msgq->entries[NR_MESSAGES - 1].next = NULL;
/drivers/net/ethernet/cisco/enic/
H A Dvnic_wq.h67 /* Break the vnic_wq_buf allocations into blocks of 32/64 entries */
70 #define VNIC_WQ_BUF_BLK_ENTRIES(entries) \
71 ((unsigned int)((entries < VNIC_WQ_BUF_DFLT_BLK_ENTRIES) ? \
73 #define VNIC_WQ_BUF_BLK_SZ(entries) \
74 (VNIC_WQ_BUF_BLK_ENTRIES(entries) * sizeof(struct vnic_wq_buf))
75 #define VNIC_WQ_BUF_BLKS_NEEDED(entries) \
76 DIV_ROUND_UP(entries, VNIC_WQ_BUF_BLK_ENTRIES(entries))
H A Dvnic_rq.h55 /* Break the vnic_rq_buf allocations into blocks of 32/64 entries */
58 #define VNIC_RQ_BUF_BLK_ENTRIES(entries) \
59 ((unsigned int)((entries < VNIC_RQ_BUF_DFLT_BLK_ENTRIES) ? \
61 #define VNIC_RQ_BUF_BLK_SZ(entries) \
62 (VNIC_RQ_BUF_BLK_ENTRIES(entries) * sizeof(struct vnic_rq_buf))
63 #define VNIC_RQ_BUF_BLKS_NEEDED(entries) \
64 DIV_ROUND_UP(entries, VNIC_RQ_BUF_BLK_ENTRIES(entries))
/drivers/media/dvb-frontends/
H A Ddvb-pll.c74 } entries[12]; member in struct:dvb_pll_desc
87 .entries = {
110 .entries = {
133 .entries = {
151 .entries = {
166 .entries = {
188 .entries = {
202 .entries = {
228 .entries = {
261 .entries
[all...]
/drivers/gpu/drm/radeon/
H A Dr600_dpm.c824 radeon_table->entries = kzalloc(size, GFP_KERNEL);
825 if (!radeon_table->entries)
828 entry = &atom_table->entries[0];
830 radeon_table->entries[i].clk = le16_to_cpu(entry->usClockLow) |
832 radeon_table->entries[i].v = le16_to_cpu(entry->usVoltage);
926 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries);
937 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries);
938 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries);
949 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries);
950 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries);
[all...]
H A Dpptable.h467 UCHAR ucNumEntries; // Number of entries.
468 ATOM_PPLIB_Clock_Voltage_Dependency_Record entries[1]; // Dynamically allocate entries. member in struct:_ATOM_PPLIB_Clock_Voltage_Dependency_Table
483 UCHAR ucNumEntries; // Number of entries.
484 ATOM_PPLIB_Clock_Voltage_Limit_Record entries[1]; // Dynamically allocate entries. member in struct:_ATOM_PPLIB_Clock_Voltage_Limit_Table
507 UCHAR ucNumEntries; // Number of entries.
508 ATOM_PPLIB_CAC_Leakage_Record entries[1]; // Dynamically allocate entries. member in struct:_ATOM_PPLIB_CAC_Leakage_Table
522 UCHAR ucNumEntries; // Number of entries
523 ATOM_PPLIB_PhaseSheddingLimits_Record entries[1]; // Dynamically allocate entries. member in struct:_ATOM_PPLIB_PhaseSheddingLimits_Table
535 VCEClockInfo entries[1]; member in struct:_VCEClockInfoArray
547 ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record entries[1]; member in struct:_ATOM_PPLIB_VCE_Clock_Voltage_Limit_Table
559 ATOM_PPLIB_VCE_State_Record entries[1]; member in struct:_ATOM_PPLIB_VCE_State_Table
581 UVDClockInfo entries[1]; member in struct:_UVDClockInfoArray
593 ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record entries[1]; member in struct:_ATOM_PPLIB_UVD_Clock_Voltage_Limit_Table
612 ATOM_PPLIB_SAMClk_Voltage_Limit_Record entries[1]; member in struct:_ATOM_PPLIB_SAMClk_Voltage_Limit_Table
630 ATOM_PPLIB_ACPClk_Voltage_Limit_Record entries[1]; member in struct:_ATOM_PPLIB_ACPClk_Voltage_Limit_Table
[all...]
/drivers/gpu/drm/nouveau/core/subdev/mxm/
H A Dmxms.c108 u8 entries = 0; local
125 entries = (ROM32(desc[0]) & 0x01f00000) >> 20;
134 entries = (desc[1] & 0xf0) >> 4;
142 entries = desc[1] & 0x07;
163 for (i = 0; i < entries; i++, dump += recordlen) {
176 desc += headerlen + (entries * recordlen);
/drivers/md/
H A Ddm-bio-prison.c284 struct dm_deferred_entry entries[DEFERRED_SET_SIZE]; member in struct:dm_deferred_set
300 ds->entries[i].ds = ds;
301 ds->entries[i].count = 0;
302 INIT_LIST_HEAD(&ds->entries[i].work_items);
321 entry = ds->entries + ds->current_entry;
337 !ds->entries[ds->sweeper].count) {
338 list_splice_init(&ds->entries[ds->sweeper].work_items, head);
342 if ((ds->sweeper == ds->current_entry) && !ds->entries[ds->sweeper].count)
343 list_splice_init(&ds->entries[ds->sweeper].work_items, head);
369 !ds->entries[d
[all...]
/drivers/net/wireless/p54/
H A Deeprom.c84 size_t entries; member in struct:p54_channel_list
157 if ((!list->entries) || (!list->band_channel_num[band]))
174 (i < list->entries); i++) {
245 for (i = list->entries; i >= 0; i--) {
252 if ((i < 0) && (list->entries < list->max_entries)) {
261 i = list->entries++;
330 if ((priv->iq_autocal_len != priv->curve_data->entries) ||
331 (priv->iq_autocal_len != priv->output_limit->entries))
336 max_channel_num = max_t(unsigned int, priv->output_limit->entries,
339 priv->curve_data->entries);
523 size_t db_len, entries; local
707 size_t payload_len, entries, entry_size, offset; local
[all...]
/drivers/uwb/
H A Dest.c51 u8 entries; member in struct:uwb_est
236 * right position (entries are sorted by type, event_high, vendor and
257 const struct uwb_est_entry *entry, size_t entries)
283 uwb_est[itr].entries = entries;
306 const struct uwb_est_entry *entry, size_t entries)
315 .entries = entries
362 if (event_low >= est->entries) { /* in range? */
365 est->entries, event_lo
256 uwb_est_register(u8 type, u8 event_high, u16 vendor, u16 product, const struct uwb_est_entry *entry, size_t entries) argument
305 uwb_est_unregister(u8 type, u8 event_high, u16 vendor, u16 product, const struct uwb_est_entry *entry, size_t entries) argument
[all...]
/drivers/xen/xen-pciback/
H A Dpciback_ops.c199 struct msix_entry *entries; local
207 entries = kmalloc(op->value * sizeof(*entries), GFP_KERNEL);
208 if (entries == NULL)
212 entries[i].entry = op->msix_entries[i].entry;
213 entries[i].vector = op->msix_entries[i].vector;
216 result = pci_enable_msix_exact(dev, entries, op->value);
219 op->msix_entries[i].entry = entries[i].entry;
220 if (entries[i].vector) {
222 xen_pirq_from_irq(entries[
[all...]
/drivers/acpi/apei/
H A Derst.c381 ERST_TAB_ENTRY(erst_tab), erst_tab->entries);
439 u64 *entries; member in struct:erst_record_id_cache
488 u64 *entries; local
509 entries = erst_record_id_cache.entries;
511 if (entries[i] == id)
529 alloc_size = new_size * sizeof(entries[0]);
536 memcpy(new_entries, entries,
537 erst_record_id_cache.len * sizeof(entries[0]));
539 kfree(entries);
559 u64 *entries; local
604 u64 *entries; local
879 u64 *entries; local
[all...]
/drivers/infiniband/hw/mlx4/
H A Dcq.c169 struct ib_cq *mlx4_ib_create_cq(struct ib_device *ibdev, int entries, int vector, argument
178 if (entries < 1 || entries > dev->dev->caps.max_cqes)
185 entries = roundup_pow_of_two(entries + 1);
186 cq->ibcq.cqe = entries - 1;
201 ucmd.buf_addr, entries);
221 err = mlx4_ib_alloc_cq_buf(dev, &cq->buf, entries);
231 err = mlx4_cq_alloc(dev->dev, entries, &cq->buf.mtt, uar,
270 int entries)
269 mlx4_alloc_resize_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq *cq, int entries) argument
293 mlx4_alloc_resize_umem(struct mlx4_ib_dev *dev, struct mlx4_ib_cq *cq, int entries, struct ib_udata *udata) argument
358 mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) argument
[all...]
/drivers/iio/trigger/
H A DKconfig4 # When adding new entries keep the list in alphabetical order
22 Provides support for using SYSFS entries as IIO triggers.
/drivers/net/wireless/ath/ath6kl/
H A Dtrace.h121 unsigned int entries, struct hif_scatter_item *list),
123 TP_ARGS(addr, flags, total_len, entries, list),
129 __field(unsigned int, entries)
131 __dynamic_array(unsigned int, len_array, entries)
142 __entry->entries = entries;
152 for (i = 0; i < entries; i++) {
164 "%s addr 0x%x flags 0x%x entries %d total_len %zd\n",
168 __entry->entries,
/drivers/net/wireless/rt2x00/
H A Drt2x00queue.c559 * 2) Rule 1 can be broken when the available entries
824 if (fn(&queue->entries[i], data))
829 if (fn(&queue->entries[i], data))
834 if (fn(&queue->entries[i], data))
857 entry = &queue->entries[queue->index[index]];
1094 rt2x00dev->ops->lib->clear_entry(&queue->entries[i]);
1100 struct queue_entry *entries; local
1107 * Allocate all queue entries.
1109 entry_size = sizeof(*entries) + queue->priv_size;
1110 entries
[all...]
/drivers/misc/genwqe/
H A Dcard_debugfs.c50 int entries)
55 for (i = 0; i < entries; i++) {
67 int entries; local
70 entries = genwqe_ffdc_buff_size(cd, uid);
71 if (entries < 0)
74 if (entries == 0)
77 regs = kcalloc(entries, sizeof(*regs), GFP_KERNEL);
82 genwqe_ffdc_buff_read(cd, uid, regs, entries);
85 dbg_uidn_show(s, regs, entries);
115 dbg_uidn_show(s, cd->ffdc[uid].regs, cd->ffdc[uid].entries);
49 dbg_uidn_show(struct seq_file *s, struct genwqe_reg *regs, int entries) argument
[all...]
/drivers/infiniband/hw/mlx5/
H A Dcq.c602 int entries, struct mlx5_create_cq_mbox_in **cqb,
630 entries * ucmd.cqe_size,
645 ucmd.buf_addr, entries * ucmd.cqe_size, npages, page_shift, ncont);
688 int entries, int cqe_size,
704 err = alloc_cq_buf(dev, &cq->buf, entries, cqe_size);
737 struct ib_cq *mlx5_ib_create_cq(struct ib_device *ibdev, int entries, argument
751 if (entries < 0)
754 entries = roundup_pow_of_two(entries + 1);
755 if (entries > de
600 create_cq_user(struct mlx5_ib_dev *dev, struct ib_udata *udata, struct ib_ucontext *context, struct mlx5_ib_cq *cq, int entries, struct mlx5_create_cq_mbox_in **cqb, int *cqe_size, int *index, int *inlen) argument
687 create_cq_kernel(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, int cqe_size, struct mlx5_create_cq_mbox_in **cqb, int *index, int *inlen) argument
943 resize_user(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, struct ib_udata *udata, int *npas, int *page_shift, int *cqe_size) argument
981 resize_kernel(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, int cqe_size) argument
1065 mlx5_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) argument
[all...]
/drivers/of/
H A Ddynamic.c529 INIT_LIST_HEAD(&ocs->entries);
544 list_for_each_entry_safe_reverse(ce, cen, &ocs->entries, node)
567 list_for_each_entry(ce, &ocs->entries, node) {
571 list_for_each_entry_continue_reverse(ce, &ocs->entries, node)
580 list_for_each_entry(ce, &ocs->entries, node)
605 list_for_each_entry_reverse(ce, &ocs->entries, node) {
609 list_for_each_entry_continue(ce, &ocs->entries, node)
618 list_for_each_entry_reverse(ce, &ocs->entries, node)
661 list_add_tail(&ce->node, &ocs->entries);

Completed in 599 milliseconds

1234567891011