/drivers/cpufreq/ |
H A D | cpufreq_performance.c | 19 static int cpufreq_governor_performance(struct cpufreq_policy *policy, argument 26 policy->max, event); 27 __cpufreq_driver_target(policy, policy->max, 56 MODULE_DESCRIPTION("CPUfreq policy governor 'performance'");
|
H A D | cpufreq_powersave.c | 19 static int cpufreq_governor_powersave(struct cpufreq_policy *policy, argument 26 policy->min, event); 27 __cpufreq_driver_target(policy, policy->min, 56 MODULE_DESCRIPTION("CPUfreq policy governor 'powersave'");
|
H A D | unicore2-cpufreq.c | 28 static int ucv2_verify_speed(struct cpufreq_policy *policy) argument 30 if (policy->cpu) 33 cpufreq_verify_within_cpu_limits(policy); 37 static int ucv2_target(struct cpufreq_policy *policy, argument 44 freqs.old = policy->cur; 47 cpufreq_freq_transition_begin(policy, &freqs); 48 ret = clk_set_rate(policy->clk, target_freq * 1000); 49 cpufreq_freq_transition_end(policy, &freqs, ret); 54 static int __init ucv2_cpu_init(struct cpufreq_policy *policy) argument 56 if (policy [all...] |
H A D | cpufreq_userspace.c | 26 * @policy: pointer to policy struct where freq is being set 31 static int cpufreq_set(struct cpufreq_policy *policy, unsigned int freq) argument 35 pr_debug("cpufreq_set for cpu %u, freq %u kHz\n", policy->cpu, freq); 38 if (!per_cpu(cpu_is_managed, policy->cpu)) 41 ret = __cpufreq_driver_target(policy, freq, CPUFREQ_RELATION_L); 47 static ssize_t show_speed(struct cpufreq_policy *policy, char *buf) argument 49 return sprintf(buf, "%u\n", policy->cur); 52 static int cpufreq_governor_userspace(struct cpufreq_policy *policy, argument 55 unsigned int cpu = policy [all...] |
H A D | dbx500-cpufreq.c | 22 static int dbx500_cpufreq_target(struct cpufreq_policy *policy, argument 29 static int dbx500_cpufreq_init(struct cpufreq_policy *policy) argument 31 policy->clk = armss_clk; 32 return cpufreq_generic_init(policy, freq_table, 20 * 1000);
|
H A D | at32ap-cpufreq.c | 29 static int at32_set_target(struct cpufreq_policy *policy, unsigned int index) argument 33 old_freq = policy->cur; 44 clk_set_rate(policy->clk, new_freq * 1000); 52 static int at32_cpufreq_driver_init(struct cpufreq_policy *policy) argument 58 if (policy->cpu != 0) 70 policy->cpuinfo.transition_latency = 0; 98 policy->clk = cpuclk; 101 retval = cpufreq_table_validate_and_show(policy, freq_table);
|
H A D | cris-artpec3-cpufreq.c | 30 static int cris_freq_target(struct cpufreq_policy *policy, unsigned int state) argument 50 static int cris_freq_cpu_init(struct cpufreq_policy *policy) argument 52 return cpufreq_generic_init(policy, cris_freq_table, 1000000);
|
H A D | cris-etraxfs-cpufreq.c | 30 static int cris_freq_target(struct cpufreq_policy *policy, unsigned int state) argument 50 static int cris_freq_cpu_init(struct cpufreq_policy *policy) argument 52 return cpufreq_generic_init(policy, cris_freq_table, 1000000);
|
H A D | ppc_cbe_cpufreq.c | 68 static int cbe_cpufreq_cpu_init(struct cpufreq_policy *policy) argument 76 cpu = of_get_cpu_node(policy->cpu, NULL); 81 pr_debug("init cpufreq on CPU %d\n", policy->cpu); 86 if (!cbe_get_cpu_pmd_regs(policy->cpu) || 87 !cbe_get_cpu_mic_tm_regs(policy->cpu)) { 113 policy->cpuinfo.transition_latency = 25000; 115 cur_pmode = cbe_cpufreq_get_pmode(policy->cpu); 118 policy->cur = cbe_freqs[cur_pmode].frequency; 121 cpumask_copy(policy->cpus, cpu_sibling_mask(policy 129 cbe_cpufreq_target(struct cpufreq_policy *policy, unsigned int cbe_pmode_new) argument [all...] |
H A D | sa1100-cpufreq.c | 180 static int sa1100_target(struct cpufreq_policy *policy, unsigned int ppcr) argument 198 static int __init sa1100_cpu_init(struct cpufreq_policy *policy) argument 200 return cpufreq_generic_init(policy, sa11x0_freq_table, CPUFREQ_ETERNAL);
|
H A D | sc520_freq.c | 56 static int sc520_freq_target(struct cpufreq_policy *policy, unsigned int state) argument 75 static int sc520_freq_cpu_init(struct cpufreq_policy *policy) argument 84 /* cpuinfo and default policy values */ 85 policy->cpuinfo.transition_latency = 1000000; /* 1ms */ 87 return cpufreq_table_validate_and_show(policy, sc520_freq_table);
|
H A D | sh-cpufreq.c | 41 static int sh_cpufreq_target(struct cpufreq_policy *policy, argument 45 unsigned int cpu = policy->cpu; 62 if (freq < (policy->min * 1000) || freq > (policy->max * 1000)) 71 cpufreq_freq_transition_begin(policy, &freqs); 74 cpufreq_freq_transition_end(policy, &freqs, 0); 81 static int sh_cpufreq_verify(struct cpufreq_policy *policy) argument 83 struct clk *cpuclk = &per_cpu(sh_cpuclk, policy->cpu); 88 return cpufreq_frequency_table_verify(policy, freq_table); 90 cpufreq_verify_within_cpu_limits(policy); 99 sh_cpufreq_cpu_init(struct cpufreq_policy *policy) argument 141 sh_cpufreq_cpu_exit(struct cpufreq_policy *policy) argument [all...] |
H A D | tegra-cpufreq.c | 50 static unsigned int tegra_get_intermediate(struct cpufreq_policy *policy, argument 57 * - we are already at it, i.e. policy->cur == ifreq 60 if ((freq_table[index].frequency == ifreq) || (policy->cur == ifreq)) 66 static int tegra_target_intermediate(struct cpufreq_policy *policy, argument 92 static int tegra_target(struct cpufreq_policy *policy, unsigned int index) argument 137 static int tegra_cpu_init(struct cpufreq_policy *policy) argument 141 if (policy->cpu >= NUM_CPUS) 148 ret = cpufreq_generic_init(policy, freq_table, 300 * 1000); 155 policy->clk = cpu_clk; 156 policy 160 tegra_cpu_exit(struct cpufreq_policy *policy) argument [all...] |
H A D | amd_freq_sensitivity.c | 40 static unsigned int amd_powersave_bias_target(struct cpufreq_policy *policy, argument 47 struct cpu_data_t *data = &per_cpu(cpu_data, policy->cpu); 48 struct dbs_data *od_data = policy->governor_data; 51 od_data->cdata->get_cpu_dbs_info_s(policy->cpu); 56 rdmsr_on_cpu(policy->cpu, MSR_AMD64_FREQ_SENSITIVITY_ACTUAL, 58 rdmsr_on_cpu(policy->cpu, MSR_AMD64_FREQ_SENSITIVITY_REFERENCE, 65 freq_next = policy->cur; 74 freq_next = policy->cur; 85 if (data->freq_prev == policy->cur) 86 freq_next = policy [all...] |
H A D | davinci-cpufreq.c | 41 static int davinci_verify_speed(struct cpufreq_policy *policy) argument 48 return cpufreq_frequency_table_verify(policy, freq_table); 50 if (policy->cpu) 53 cpufreq_verify_within_cpu_limits(policy); 54 policy->min = clk_round_rate(armclk, policy->min * 1000) / 1000; 55 policy->max = clk_round_rate(armclk, policy->max * 1000) / 1000; 56 cpufreq_verify_within_limits(policy, policy 61 davinci_target(struct cpufreq_policy *policy, unsigned int idx) argument 95 davinci_cpu_init(struct cpufreq_policy *policy) argument [all...] |
H A D | elanfreq.c | 108 static int elanfreq_target(struct cpufreq_policy *policy, argument 147 static int elanfreq_cpu_init(struct cpufreq_policy *policy) argument 166 /* cpuinfo and default policy values */ 167 policy->cpuinfo.transition_latency = CPUFREQ_ETERNAL; 169 return cpufreq_table_validate_and_show(policy, elanfreq_table);
|
H A D | exynos-cpufreq.c | 47 struct cpufreq_policy *policy = cpufreq_cpu_get(0); local 55 old_freq = policy->cur; 58 * The policy max have been changed so that we cannot get proper 60 * policy and get the index from the raw frequency table. 124 cpufreq_cpu_put(policy); 129 static int exynos_target(struct cpufreq_policy *policy, unsigned int index) argument 134 static int exynos_cpufreq_cpu_init(struct cpufreq_policy *policy) argument 136 policy->clk = exynos_info->cpu_clk; 137 policy->suspend_freq = locking_frequency; 138 return cpufreq_generic_init(policy, exynos_inf [all...] |
H A D | integrator-cpufreq.c | 57 * Validate the speed policy. 59 static int integrator_verify_policy(struct cpufreq_policy *policy) argument 63 cpufreq_verify_within_cpu_limits(policy); 65 vco = icst_hz_to_vco(&cclk_params, policy->max * 1000); 66 policy->max = icst_hz(&cclk_params, vco) / 1000; 68 vco = icst_hz_to_vco(&cclk_params, policy->min * 1000); 69 policy->min = icst_hz(&cclk_params, vco) / 1000; 71 cpufreq_verify_within_cpu_limits(policy); 76 static int integrator_set_target(struct cpufreq_policy *policy, argument 81 int cpu = policy 181 integrator_cpufreq_init(struct cpufreq_policy *policy) argument [all...] |
H A D | kirkwood-cpufreq.c | 54 static int kirkwood_cpufreq_target(struct cpufreq_policy *policy, argument 90 static int kirkwood_cpufreq_cpu_init(struct cpufreq_policy *policy) argument 92 return cpufreq_generic_init(policy, kirkwood_freq_table, 5000);
|
H A D | loongson2_cpufreq.c | 48 static int loongson2_cpufreq_target(struct cpufreq_policy *policy, argument 51 unsigned int cpu = policy->cpu; 65 clk_set_rate(policy->clk, freq * 1000); 70 static int loongson2_cpufreq_cpu_init(struct cpufreq_policy *policy) argument 101 policy->clk = cpuclk; 102 return cpufreq_generic_init(policy, &loongson2_clockmod_table[0], 0); 105 static int loongson2_cpufreq_exit(struct cpufreq_policy *policy) argument 107 clk_put(policy->clk);
|
H A D | omap-cpufreq.c | 42 static int omap_target(struct cpufreq_policy *policy, unsigned int index) argument 49 old_freq = policy->cur; 53 ret = clk_round_rate(policy->clk, freq); 91 ret = clk_set_rate(policy->clk, new_freq * 1000); 99 clk_set_rate(policy->clk, old_freq * 1000); 113 static int omap_cpu_init(struct cpufreq_policy *policy) argument 117 policy->clk = clk_get(NULL, "cpufreq_ck"); 118 if (IS_ERR(policy->clk)) 119 return PTR_ERR(policy->clk); 126 __func__, policy 144 omap_cpu_exit(struct cpufreq_policy *policy) argument [all...] |
H A D | ppc_cbe_cpufreq_pmi.c | 96 struct cpufreq_policy *policy = data; local 101 * and CPUFREQ_NOTIFY policy events?) 106 cbe_freqs = cpufreq_frequency_get_table(policy->cpu); 107 node = cbe_cpu_to_node(policy->cpu); 115 cpufreq_verify_within_limits(policy, 0,
|
/drivers/block/drbd/ |
H A D | drbd_nla.c | 31 const struct nla_policy *policy) 37 err = nla_parse_nested(tb, maxtype, nla, policy); 30 drbd_nla_parse_nested(struct nlattr *tb[], int maxtype, struct nlattr *nla, const struct nla_policy *policy) argument
|
/drivers/net/wireless/cw1200/ |
H A D | txrx.h | 36 struct tx_policy policy; member in struct:tx_policy_cache_entry 45 spinlock_t lock; /* Protect policy cache */ 49 /* TX policy cache */ 50 /* Intention of TX policy cache is an overcomplicated WSM API. 52 * It uses "tx retry policy id" instead, so driver code has to sync 53 * linux tx retry sequences with a retry policy table in the device.
|
/drivers/staging/lustre/lustre/ptlrpc/ |
H A D | nrs_fifo.c | 30 * Network Request Scheduler (NRS) FIFO policy 32 * Handles RPCs in a FIFO manner, as received from the network. This policy is 34 * default and fallback policy for all types of RPCs on all PTLRPC service 36 * the policy is the one enabled at PTLRPC service partition startup time, and 37 * fallback means the policy is used to handle RPCs that are not handled 38 * successfully or are not handled at all by any primary policy that may be 58 * The FIFO policy is a logical wrapper around previous, non-NRS functionality. 67 * Is called before the policy transitions into 69 * policy-specific private data structure. 71 * \param[in] policy Th 79 nrs_fifo_start(struct ptlrpc_nrs_policy *policy) argument 101 nrs_fifo_stop(struct ptlrpc_nrs_policy *policy) argument 129 nrs_fifo_res_get(struct ptlrpc_nrs_policy *policy, struct ptlrpc_nrs_request *nrq, const struct ptlrpc_nrs_resource *parent, struct ptlrpc_nrs_resource **resp, bool moving_req) argument 160 nrs_fifo_req_get(struct ptlrpc_nrs_policy *policy, bool peek, bool force) argument 194 nrs_fifo_req_add(struct ptlrpc_nrs_policy *policy, struct ptlrpc_nrs_request *nrq) argument 216 nrs_fifo_req_del(struct ptlrpc_nrs_policy *policy, struct ptlrpc_nrs_request *nrq) argument 233 nrs_fifo_req_stop(struct ptlrpc_nrs_policy *policy, struct ptlrpc_nrs_request *nrq) argument [all...] |