/drivers/cpuidle/governors/ |
H A D | ladder.c | 41 struct ladder_device_state states[CPUIDLE_STATE_MAX]; member in struct:ladder_device 56 ldev->states[old_idx].stats.promotion_count = 0; 57 ldev->states[old_idx].stats.demotion_count = 0; 80 last_state = &ldev->states[last_idx]; 82 if (drv->states[last_idx].flags & CPUIDLE_FLAG_TIME_VALID) { 84 drv->states[last_idx].exit_latency; 92 drv->states[last_idx + 1].exit_latency <= latency_req) { 103 drv->states[last_idx].exit_latency > latency_req) { 107 if (drv->states[i].exit_latency <= latency_req) 144 state = &drv->states[ [all...] |
H A D | menu.c | 83 * C states, especially those with large exit latencies, can have a real 89 * The busier the system, the less impact of C states is acceptable 289 drv->states[CPUIDLE_DRIVER_STATE_START].disable == 0) 297 struct cpuidle_state *s = &drv->states[i]; 344 struct cpuidle_state *target = &drv->states[last_idx]; 397 * menu_enable_device - scans a CPU's states and does setup
|
/drivers/regulator/ |
H A D | gpio-regulator.c | 48 struct gpio_regulator_state *states; member in struct:gpio_regulator_data 98 if (data->states[ptr].gpios == data->state) 99 return data->states[ptr].value; 112 if (data->states[ptr].value >= min && 113 data->states[ptr].value <= max) 114 target = data->states[ptr].gpios; 143 return data->states[selector].value; 199 drvdata->states = kmemdup(config->states, 203 if (drvdata->states [all...] |
/drivers/acpi/ |
H A D | processor_idle.c | 168 * Check, if one of the previous states already marked the lapic 200 int state = cx - pr->power.states; 263 * C/P/S0/S1 states when this bit is set. 289 pr->power.states[ACPI_STATE_C2].type = ACPI_STATE_C2; 290 pr->power.states[ACPI_STATE_C3].type = ACPI_STATE_C3; 303 pr->power.states[ACPI_STATE_C2].address = pr->pblk + 4; 304 pr->power.states[ACPI_STATE_C3].address = pr->pblk + 5; 307 pr->power.states[ACPI_STATE_C2].latency = acpi_gbl_FADT.C2latency; 308 pr->power.states[ACPI_STATE_C3].latency = acpi_gbl_FADT.C3latency; 318 pr->power.states[ACPI_STATE_C [all...] |
H A D | processor_perflib.c | 108 pr->performance->states[ppc]. 132 * (e.g. 0 = states 0..n; 1 = states 1..n; etc. 158 * 1: failure. OSPM has not changed the number of P-states in use 218 *limit = pr->performance->states[pr->performance_platform_limit]. 351 ACPI_DEBUG_PRINT((ACPI_DB_INFO, "Found %d performance states\n", 355 pr->performance->states = 358 if (!pr->performance->states) { 365 struct acpi_processor_px *px = &(pr->performance->states[i]); 377 kfree(pr->performance->states); [all...] |
H A D | power.c | 387 list = &acpi_dev->power.states[ACPI_STATE_D0].resources; 445 list = &acpi_dev->power.states[ACPI_STATE_D0].resources; 635 list = &device->power.states[i].resources; 658 return acpi_power_on_list(&device->power.states[state].resources); 684 &device->power.states[state].resources); 688 &device->power.states[device->power.state].resources);
|
H A D | sleep.c | 660 suspend_state_t states[] = { local 666 if (acpi_state < 6 && states[acpi_state]) 667 return pm_suspend(states[acpi_state]); 679 * @d_min_p: used to store the upper limit of allowed states range 718 * state) we can use for the corresponding S-states. Otherwise, the
|
H A D | scan.c | 870 * Enumerate supported power management states 873 struct acpi_device_power_state *ps = &device->power.states[i]; 905 /* Set defaults for D0 and D3 states (always valid) */ 906 device->power.states[ACPI_STATE_D0].flags.valid = 1; 907 device->power.states[ACPI_STATE_D0].power = 100; 908 device->power.states[ACPI_STATE_D3].flags.valid = 1; 909 device->power.states[ACPI_STATE_D3].power = 0; 912 if (device->power.states[ACPI_STATE_D3_HOT].flags.explicit_set) 913 device->power.states[ACPI_STATE_D3_COLD].flags.explicit_set = 1;
|
H A D | processor_throttling.c | 340 * t-states from high to low for better performance. 532 ACPI_DEBUG_PRINT((ACPI_DB_INFO, "Found %d throttling states\n", 707 state, pr->throttling.states[state].performance)); 935 ACPI_DEBUG_PRINT((ACPI_DB_INFO, "No throttling states\n")); 955 pr->throttling.states[i].performance = 1000 - step * i; 956 pr->throttling.states[i].power = 1000 - step * i; 1028 (pr->throttling.states[state].performance ? pr-> 1029 throttling.states[state].performance / 10 : 0))); 1170 * affected cpu to update the T-states. 1241 ACPI_DEBUG_PRINT((ACPI_DB_INFO, "Found %d throttling states\ [all...] |
/drivers/cpufreq/ |
H A D | acpi-cpufreq.c | 93 if (value == perf->states[i].status) 108 if (msr == perf->states[data->freq_table[i].index].status) 327 cmd.val = (u32) perf->states[next_perf_state].control; 333 cmd.val = (u32) perf->states[next_perf_state].control; 346 freqs.old = perf->states[perf->state].core_frequency * 1000; 392 unsigned long freqn = perf->states[0].core_frequency * 1000; 396 freqn = perf->states[i+1].core_frequency * 1000; 407 return perf->states[0].core_frequency * 1000; 606 if ((perf->states[i].transition_latency * 1000) > 609 perf->states[ [all...] |
H A D | e_powersaver.c | 220 int states; local 360 /* Calc number of p-states supported */ 362 states = max_multiplier - min_multiplier + 1; 364 states = 2; 368 + (states + 1) * sizeof(struct cpufreq_frequency_table),
|
H A D | powernow-k8.c | 691 printk(KERN_ERR PFX "no p states to transition\n"); 831 control = data->acpi_data.states[index].control; 930 index = data->acpi_data.states[i].control & HW_PSTATE_MASK; 957 data->acpi_data.states[i].core_frequency * 1000; 976 status = data->acpi_data.states[i].status; 980 control = data->acpi_data.states[i].control; 1008 if (freq != (data->acpi_data.states[i].core_frequency * 1000)) { 1012 (data->acpi_data.states[i].core_frequency 1034 int cur_latency = data->acpi_data.states[i].transition_latency 1035 + data->acpi_data.states[ [all...] |
/drivers/cpuidle/ |
H A D | driver.c | 35 * c-states. 39 drv->states[i].power_usage = -1 - i;
|
H A D | cpuidle.c | 59 struct cpuidle_state *target_state = &drv->states[index]; 91 struct cpuidle_state *s = &drv->states[i]; 100 return drv->states[dead_state].enter_dead(dev, dead_state); 109 * @next_state: index into drv->states of the state to enter 290 struct cpuidle_state *state = &drv->states[0];
|
/drivers/pinctrl/ |
H A D | core.h | 53 * @states: a list of states for this device 59 struct list_head states; member in struct:pinctrl 65 * @node: list not for struct pinctrl's @states field
|
H A D | core.c | 466 list_for_each_entry(state, &p->states, node) 488 list_add_tail(&state->node, &p->states); 581 INIT_LIST_HEAD(&p->states); 643 list_for_each_entry_safe(state, n1, &p->states, node) { 1059 list_for_each_entry(state, &p->states, node) {
|
/drivers/net/ethernet/mellanox/mlx4/ |
H A D | qp.c | 485 enum mlx4_qp_state states[] = { local 492 for (i = 0; i < ARRAY_SIZE(states) - 1; i++) { 494 context->flags |= cpu_to_be32(states[i + 1] << 28); 495 err = mlx4_qp_modify(dev, mtt, states[i], states[i + 1], 500 states[i + 1], err); 504 *qp_state = states[i + 1];
|
/drivers/xen/ |
H A D | xen-acpi-processor.c | 76 cx = &_pr->power.states[i]; 117 set_xen_guest_handle(op.u.set_pminfo.power.states, dst_cx_states); 123 pr_debug("ACPI CPU%u - C-states uploaded.\n", _pr->acpi_id); 125 cx = &_pr->power.states[i]; 160 memcpy(&(dst_states[i]), &(_pr->performance->states[i]), 234 set_xen_guest_handle(dst_perf->states, dst_states); 255 pr_debug("ACPI CPU%u - P-states uploaded.\n", _pr->acpi_id); 259 (u32) perf->states[i].core_frequency, 260 (u32) perf->states[i].power, 261 (u32) perf->states[ [all...] |
/drivers/input/ |
H A D | ff-memless.c | 62 struct ml_effect_state states[FF_MEMLESS_EFFECTS]; member in struct:ml_device 136 state = &ml->states[i]; 340 state = &ml->states[i]; 426 __clear_bit(FF_EFFECT_PLAYING, &ml->states[i].flags); 437 struct ml_effect_state *state = &ml->states[effect_id]; 468 struct ml_effect_state *state = &ml->states[effect->id]; 542 ml->states[i].effect = &ff->effects[i];
|
/drivers/input/misc/ |
H A D | gpio_tilt_polled.c | 58 if (pdata->states[i].gpios == state) 59 tilt_state = &pdata->states[i];
|
/drivers/iommu/ |
H A D | amd_iommu_v2.c | 51 struct pri_queue pri[PRI_QUEUE_SIZE]; /* PRI tag states */ 61 struct pasid_state **states; member in struct:device_state 192 root = dev_state->states; 386 free_pasid_states_level2(dev_state->states); 388 free_pasid_states_level1(dev_state->states); 392 free_page((unsigned long)dev_state->states); 782 dev_state->states = (void *)get_zeroed_page(GFP_KERNEL); 783 if (dev_state->states == NULL) 818 free_page((unsigned long)dev_state->states); 850 /* Get rid of any remaining pasid states */ [all...] |
/drivers/block/ |
H A D | pktcdvd.c | 1542 static void pkt_count_states(struct pktcdvd_device *pd, int *states) argument 1548 states[i] = 0; 1552 states[pkt->state]++; 1596 int states[PACKET_NUM_STATES]; local 1597 pkt_count_states(pd, states); 1599 states[0], states[1], states[2], states[3], 1600 states[ 2630 int states[PACKET_NUM_STATES]; local [all...] |
/drivers/idle/ |
H A D | intel_idle.c | 31 * All CPUs have same idle states as boot CPU 103 * Set this flag for states where the HW flushes the TLB for us 247 struct cpuidle_state *state = &drv->states[index]; 477 " states deeper than C2"); 479 drv->states[drv->state_count] = /* structure copy */
|
/drivers/scsi/ |
H A D | raid_class.c | 186 #define raid_attr_ro_states(attr, states, code) \ 190 name = raid_##states##_name(rd->attr); \
|
H A D | aha152x.c | 447 * internal states of the host 659 * driver states 668 } states[] = { variable in typeref:struct:__anon4044 2542 DPRINTK(debug_phases, LEAD "start %s %s(%s)\n", CMDINFO(CURRENT_SC), states[STATE].name, states[PREVSTATE].name, states[LASTSTATE].name); 2548 if(PREVSTATE!=STATE && states[PREVSTATE].end) 2549 states[PREVSTATE].end(shpnt); 2556 if(states[PREVSTATE].spio && !states[STAT [all...] |