/drivers/misc/vmw_vmci/ |
H A D | vmci_handle_array.c | 66 array->entries[array->size] = handle; 80 if (vmci_handle_is_equal(array->entries[i], entry_handle)) { 81 handle = array->entries[i]; 83 array->entries[i] = array->entries[array->size]; 84 array->entries[array->size] = VMCI_INVALID_HANDLE; 101 handle = array->entries[array->size]; 102 array->entries[array->size] = VMCI_INVALID_HANDLE; 117 return array->entries[index]; 126 if (vmci_handle_is_equal(array->entries[ [all...] |
H A D | vmci_handle_array.h | 28 struct vmci_handle entries[]; member in struct:vmci_handle_arr
|
H A D | vmci_resource.c | 30 struct hlist_head entries[VMCI_RESOURCE_HASH_BUCKETS]; member in struct:vmci_hash_table 53 &vmci_resource_table.entries[idx], node) { 134 hlist_add_head_rcu(&resource->node, &vmci_resource_table.entries[idx]); 152 hlist_for_each_entry(r, &vmci_resource_table.entries[idx], node) {
|
/drivers/gpu/drm/nouveau/core/subdev/bios/ |
H A D | bit.c | 34 u8 entries = nv_ro08(bios, bios->bit_offset + 10); local 36 while (entries--) {
|
/drivers/crypto/qat/qat_dh895xcc/ |
H A D | adf_isr.c | 71 pci_dev_info->msix_entries.entries[i].entry = i; 74 pci_dev_info->msix_entries.entries, 108 struct msix_entry *msixe = pci_dev_info->msix_entries.entries; 150 struct msix_entry *msixe = pci_dev_info->msix_entries.entries; 166 struct msix_entry *entries; local 170 entries = kzalloc_node(msix_num_entries * sizeof(*entries), 172 if (!entries) 177 kfree(entries); 185 accel_dev->accel_pci_dev.msix_entries.entries [all...] |
/drivers/scsi/arm/ |
H A D | msgqueue.h | 31 struct msgqueue_entry entries[NR_MESSAGES]; member in struct:__anon5178
|
H A D | msgqueue.c | 59 msgq->free = &msgq->entries[0]; 62 msgq->entries[i].next = &msgq->entries[i + 1]; 64 msgq->entries[NR_MESSAGES - 1].next = NULL;
|
/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_wq.h | 67 /* Break the vnic_wq_buf allocations into blocks of 32/64 entries */ 70 #define VNIC_WQ_BUF_BLK_ENTRIES(entries) \ 71 ((unsigned int)((entries < VNIC_WQ_BUF_DFLT_BLK_ENTRIES) ? \ 73 #define VNIC_WQ_BUF_BLK_SZ(entries) \ 74 (VNIC_WQ_BUF_BLK_ENTRIES(entries) * sizeof(struct vnic_wq_buf)) 75 #define VNIC_WQ_BUF_BLKS_NEEDED(entries) \ 76 DIV_ROUND_UP(entries, VNIC_WQ_BUF_BLK_ENTRIES(entries))
|
H A D | vnic_rq.h | 55 /* Break the vnic_rq_buf allocations into blocks of 32/64 entries */ 58 #define VNIC_RQ_BUF_BLK_ENTRIES(entries) \ 59 ((unsigned int)((entries < VNIC_RQ_BUF_DFLT_BLK_ENTRIES) ? \ 61 #define VNIC_RQ_BUF_BLK_SZ(entries) \ 62 (VNIC_RQ_BUF_BLK_ENTRIES(entries) * sizeof(struct vnic_rq_buf)) 63 #define VNIC_RQ_BUF_BLKS_NEEDED(entries) \ 64 DIV_ROUND_UP(entries, VNIC_RQ_BUF_BLK_ENTRIES(entries))
|
/drivers/media/dvb-frontends/ |
H A D | dvb-pll.c | 74 } entries[12]; member in struct:dvb_pll_desc 87 .entries = { 110 .entries = { 133 .entries = { 151 .entries = { 166 .entries = { 188 .entries = { 202 .entries = { 228 .entries = { 261 .entries [all...] |
/drivers/gpu/drm/radeon/ |
H A D | r600_dpm.c | 824 radeon_table->entries = kzalloc(size, GFP_KERNEL); 825 if (!radeon_table->entries) 828 entry = &atom_table->entries[0]; 830 radeon_table->entries[i].clk = le16_to_cpu(entry->usClockLow) | 832 radeon_table->entries[i].v = le16_to_cpu(entry->usVoltage); 926 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); 937 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); 938 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); 949 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); 950 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); [all...] |
H A D | pptable.h | 467 UCHAR ucNumEntries; // Number of entries. 468 ATOM_PPLIB_Clock_Voltage_Dependency_Record entries[1]; // Dynamically allocate entries. member in struct:_ATOM_PPLIB_Clock_Voltage_Dependency_Table 483 UCHAR ucNumEntries; // Number of entries. 484 ATOM_PPLIB_Clock_Voltage_Limit_Record entries[1]; // Dynamically allocate entries. member in struct:_ATOM_PPLIB_Clock_Voltage_Limit_Table 507 UCHAR ucNumEntries; // Number of entries. 508 ATOM_PPLIB_CAC_Leakage_Record entries[1]; // Dynamically allocate entries. member in struct:_ATOM_PPLIB_CAC_Leakage_Table 522 UCHAR ucNumEntries; // Number of entries 523 ATOM_PPLIB_PhaseSheddingLimits_Record entries[1]; // Dynamically allocate entries. member in struct:_ATOM_PPLIB_PhaseSheddingLimits_Table 535 VCEClockInfo entries[1]; member in struct:_VCEClockInfoArray 547 ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record entries[1]; member in struct:_ATOM_PPLIB_VCE_Clock_Voltage_Limit_Table 559 ATOM_PPLIB_VCE_State_Record entries[1]; member in struct:_ATOM_PPLIB_VCE_State_Table 581 UVDClockInfo entries[1]; member in struct:_UVDClockInfoArray 593 ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record entries[1]; member in struct:_ATOM_PPLIB_UVD_Clock_Voltage_Limit_Table 612 ATOM_PPLIB_SAMClk_Voltage_Limit_Record entries[1]; member in struct:_ATOM_PPLIB_SAMClk_Voltage_Limit_Table 630 ATOM_PPLIB_ACPClk_Voltage_Limit_Record entries[1]; member in struct:_ATOM_PPLIB_ACPClk_Voltage_Limit_Table [all...] |
/drivers/gpu/drm/nouveau/core/subdev/mxm/ |
H A D | mxms.c | 108 u8 entries = 0; local 125 entries = (ROM32(desc[0]) & 0x01f00000) >> 20; 134 entries = (desc[1] & 0xf0) >> 4; 142 entries = desc[1] & 0x07; 163 for (i = 0; i < entries; i++, dump += recordlen) { 176 desc += headerlen + (entries * recordlen);
|
/drivers/md/ |
H A D | dm-bio-prison.c | 284 struct dm_deferred_entry entries[DEFERRED_SET_SIZE]; member in struct:dm_deferred_set 300 ds->entries[i].ds = ds; 301 ds->entries[i].count = 0; 302 INIT_LIST_HEAD(&ds->entries[i].work_items); 321 entry = ds->entries + ds->current_entry; 337 !ds->entries[ds->sweeper].count) { 338 list_splice_init(&ds->entries[ds->sweeper].work_items, head); 342 if ((ds->sweeper == ds->current_entry) && !ds->entries[ds->sweeper].count) 343 list_splice_init(&ds->entries[ds->sweeper].work_items, head); 369 !ds->entries[d [all...] |
/drivers/net/wireless/p54/ |
H A D | eeprom.c | 84 size_t entries; member in struct:p54_channel_list 157 if ((!list->entries) || (!list->band_channel_num[band])) 174 (i < list->entries); i++) { 245 for (i = list->entries; i >= 0; i--) { 252 if ((i < 0) && (list->entries < list->max_entries)) { 261 i = list->entries++; 330 if ((priv->iq_autocal_len != priv->curve_data->entries) || 331 (priv->iq_autocal_len != priv->output_limit->entries)) 336 max_channel_num = max_t(unsigned int, priv->output_limit->entries, 339 priv->curve_data->entries); 523 size_t db_len, entries; local 707 size_t payload_len, entries, entry_size, offset; local [all...] |
/drivers/uwb/ |
H A D | est.c | 51 u8 entries; member in struct:uwb_est 236 * right position (entries are sorted by type, event_high, vendor and 257 const struct uwb_est_entry *entry, size_t entries) 283 uwb_est[itr].entries = entries; 306 const struct uwb_est_entry *entry, size_t entries) 315 .entries = entries 362 if (event_low >= est->entries) { /* in range? */ 365 est->entries, event_lo 256 uwb_est_register(u8 type, u8 event_high, u16 vendor, u16 product, const struct uwb_est_entry *entry, size_t entries) argument 305 uwb_est_unregister(u8 type, u8 event_high, u16 vendor, u16 product, const struct uwb_est_entry *entry, size_t entries) argument [all...] |
/drivers/xen/xen-pciback/ |
H A D | pciback_ops.c | 199 struct msix_entry *entries; local 207 entries = kmalloc(op->value * sizeof(*entries), GFP_KERNEL); 208 if (entries == NULL) 212 entries[i].entry = op->msix_entries[i].entry; 213 entries[i].vector = op->msix_entries[i].vector; 216 result = pci_enable_msix_exact(dev, entries, op->value); 219 op->msix_entries[i].entry = entries[i].entry; 220 if (entries[i].vector) { 222 xen_pirq_from_irq(entries[ [all...] |
/drivers/acpi/apei/ |
H A D | erst.c | 381 ERST_TAB_ENTRY(erst_tab), erst_tab->entries); 439 u64 *entries; member in struct:erst_record_id_cache 488 u64 *entries; local 509 entries = erst_record_id_cache.entries; 511 if (entries[i] == id) 529 alloc_size = new_size * sizeof(entries[0]); 536 memcpy(new_entries, entries, 537 erst_record_id_cache.len * sizeof(entries[0])); 539 kfree(entries); 559 u64 *entries; local 604 u64 *entries; local 879 u64 *entries; local [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 169 struct ib_cq *mlx4_ib_create_cq(struct ib_device *ibdev, int entries, int vector, argument 178 if (entries < 1 || entries > dev->dev->caps.max_cqes) 185 entries = roundup_pow_of_two(entries + 1); 186 cq->ibcq.cqe = entries - 1; 201 ucmd.buf_addr, entries); 221 err = mlx4_ib_alloc_cq_buf(dev, &cq->buf, entries); 231 err = mlx4_cq_alloc(dev->dev, entries, &cq->buf.mtt, uar, 270 int entries) 269 mlx4_alloc_resize_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq *cq, int entries) argument 293 mlx4_alloc_resize_umem(struct mlx4_ib_dev *dev, struct mlx4_ib_cq *cq, int entries, struct ib_udata *udata) argument 358 mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) argument [all...] |
/drivers/iio/trigger/ |
H A D | Kconfig | 4 # When adding new entries keep the list in alphabetical order 22 Provides support for using SYSFS entries as IIO triggers.
|
/drivers/net/wireless/ath/ath6kl/ |
H A D | trace.h | 121 unsigned int entries, struct hif_scatter_item *list), 123 TP_ARGS(addr, flags, total_len, entries, list), 129 __field(unsigned int, entries) 131 __dynamic_array(unsigned int, len_array, entries) 142 __entry->entries = entries; 152 for (i = 0; i < entries; i++) { 164 "%s addr 0x%x flags 0x%x entries %d total_len %zd\n", 168 __entry->entries,
|
/drivers/net/wireless/rt2x00/ |
H A D | rt2x00queue.c | 559 * 2) Rule 1 can be broken when the available entries 824 if (fn(&queue->entries[i], data)) 829 if (fn(&queue->entries[i], data)) 834 if (fn(&queue->entries[i], data)) 857 entry = &queue->entries[queue->index[index]]; 1094 rt2x00dev->ops->lib->clear_entry(&queue->entries[i]); 1100 struct queue_entry *entries; local 1107 * Allocate all queue entries. 1109 entry_size = sizeof(*entries) + queue->priv_size; 1110 entries [all...] |
/drivers/misc/genwqe/ |
H A D | card_debugfs.c | 50 int entries) 55 for (i = 0; i < entries; i++) { 67 int entries; local 70 entries = genwqe_ffdc_buff_size(cd, uid); 71 if (entries < 0) 74 if (entries == 0) 77 regs = kcalloc(entries, sizeof(*regs), GFP_KERNEL); 82 genwqe_ffdc_buff_read(cd, uid, regs, entries); 85 dbg_uidn_show(s, regs, entries); 115 dbg_uidn_show(s, cd->ffdc[uid].regs, cd->ffdc[uid].entries); 49 dbg_uidn_show(struct seq_file *s, struct genwqe_reg *regs, int entries) argument [all...] |
/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 602 int entries, struct mlx5_create_cq_mbox_in **cqb, 630 entries * ucmd.cqe_size, 645 ucmd.buf_addr, entries * ucmd.cqe_size, npages, page_shift, ncont); 688 int entries, int cqe_size, 704 err = alloc_cq_buf(dev, &cq->buf, entries, cqe_size); 737 struct ib_cq *mlx5_ib_create_cq(struct ib_device *ibdev, int entries, argument 751 if (entries < 0) 754 entries = roundup_pow_of_two(entries + 1); 755 if (entries > de 600 create_cq_user(struct mlx5_ib_dev *dev, struct ib_udata *udata, struct ib_ucontext *context, struct mlx5_ib_cq *cq, int entries, struct mlx5_create_cq_mbox_in **cqb, int *cqe_size, int *index, int *inlen) argument 687 create_cq_kernel(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, int cqe_size, struct mlx5_create_cq_mbox_in **cqb, int *index, int *inlen) argument 943 resize_user(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, struct ib_udata *udata, int *npas, int *page_shift, int *cqe_size) argument 981 resize_kernel(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, int cqe_size) argument 1065 mlx5_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) argument [all...] |
/drivers/of/ |
H A D | dynamic.c | 529 INIT_LIST_HEAD(&ocs->entries); 544 list_for_each_entry_safe_reverse(ce, cen, &ocs->entries, node) 567 list_for_each_entry(ce, &ocs->entries, node) { 571 list_for_each_entry_continue_reverse(ce, &ocs->entries, node) 580 list_for_each_entry(ce, &ocs->entries, node) 605 list_for_each_entry_reverse(ce, &ocs->entries, node) { 609 list_for_each_entry_continue(ce, &ocs->entries, node) 618 list_for_each_entry_reverse(ce, &ocs->entries, node) 661 list_add_tail(&ce->node, &ocs->entries);
|