processor_core.c revision 78ed8bd2944b6400f742306e5fe9d1b9b6bf18ba
1/* 2 * Copyright (C) 2005 Intel Corporation 3 * Copyright (C) 2009 Hewlett-Packard Development Company, L.P. 4 * 5 * Alex Chiang <achiang@hp.com> 6 * - Unified x86/ia64 implementations 7 * Venkatesh Pallipadi <venkatesh.pallipadi@intel.com> 8 * - Added _PDC for platforms with Intel CPUs 9 */ 10#include <linux/dmi.h> 11 12#include <acpi/acpi_drivers.h> 13#include <acpi/processor.h> 14 15#include "internal.h" 16 17#define PREFIX "ACPI: " 18#define _COMPONENT ACPI_PROCESSOR_COMPONENT 19ACPI_MODULE_NAME("processor_core"); 20 21static int set_no_mwait(const struct dmi_system_id *id) 22{ 23 printk(KERN_NOTICE PREFIX "%s detected - " 24 "disabling mwait for CPU C-states\n", id->ident); 25 idle_nomwait = 1; 26 return 0; 27} 28 29static struct dmi_system_id __cpuinitdata processor_idle_dmi_table[] = { 30 { 31 set_no_mwait, "IFL91 board", { 32 DMI_MATCH(DMI_BIOS_VENDOR, "COMPAL"), 33 DMI_MATCH(DMI_SYS_VENDOR, "ZEPTO"), 34 DMI_MATCH(DMI_PRODUCT_VERSION, "3215W"), 35 DMI_MATCH(DMI_BOARD_NAME, "IFL91") }, NULL}, 36 { 37 set_no_mwait, "Extensa 5220", { 38 DMI_MATCH(DMI_BIOS_VENDOR, "Phoenix Technologies LTD"), 39 DMI_MATCH(DMI_SYS_VENDOR, "Acer"), 40 DMI_MATCH(DMI_PRODUCT_VERSION, "0100"), 41 DMI_MATCH(DMI_BOARD_NAME, "Columbia") }, NULL}, 42 {}, 43}; 44 45#ifdef CONFIG_SMP 46static struct acpi_table_madt *madt; 47 48static int map_lapic_id(struct acpi_subtable_header *entry, 49 u32 acpi_id, int *apic_id) 50{ 51 struct acpi_madt_local_apic *lapic = 52 (struct acpi_madt_local_apic *)entry; 53 if ((lapic->lapic_flags & ACPI_MADT_ENABLED) && 54 lapic->processor_id == acpi_id) { 55 *apic_id = lapic->id; 56 return 1; 57 } 58 return 0; 59} 60 61static int map_x2apic_id(struct acpi_subtable_header *entry, 62 int device_declaration, u32 acpi_id, int *apic_id) 63{ 64 struct acpi_madt_local_x2apic *apic = 65 (struct acpi_madt_local_x2apic *)entry; 66 u32 tmp = apic->local_apic_id; 67 68 /* Only check enabled APICs*/ 69 if (!(apic->lapic_flags & ACPI_MADT_ENABLED)) 70 return 0; 71 72 /* Device statement declaration type */ 73 if (device_declaration) { 74 if (apic->uid == acpi_id) 75 goto found; 76 } 77 78 return 0; 79found: 80 *apic_id = tmp; 81 return 1; 82} 83 84static int map_lsapic_id(struct acpi_subtable_header *entry, 85 int device_declaration, u32 acpi_id, int *apic_id) 86{ 87 struct acpi_madt_local_sapic *lsapic = 88 (struct acpi_madt_local_sapic *)entry; 89 u32 tmp = (lsapic->id << 8) | lsapic->eid; 90 91 /* Only check enabled APICs*/ 92 if (!(lsapic->lapic_flags & ACPI_MADT_ENABLED)) 93 return 0; 94 95 /* Device statement declaration type */ 96 if (device_declaration) { 97 if (entry->length < 16) 98 printk(KERN_ERR PREFIX 99 "Invalid LSAPIC with Device type processor (SAPIC ID %#x)\n", 100 tmp); 101 else if (lsapic->uid == acpi_id) 102 goto found; 103 /* Processor statement declaration type */ 104 } else if (lsapic->processor_id == acpi_id) 105 goto found; 106 107 return 0; 108found: 109 *apic_id = tmp; 110 return 1; 111} 112 113static int map_madt_entry(int type, u32 acpi_id) 114{ 115 unsigned long madt_end, entry; 116 int apic_id = -1; 117 118 if (!madt) 119 return apic_id; 120 121 entry = (unsigned long)madt; 122 madt_end = entry + madt->header.length; 123 124 /* Parse all entries looking for a match. */ 125 126 entry += sizeof(struct acpi_table_madt); 127 while (entry + sizeof(struct acpi_subtable_header) < madt_end) { 128 struct acpi_subtable_header *header = 129 (struct acpi_subtable_header *)entry; 130 if (header->type == ACPI_MADT_TYPE_LOCAL_APIC) { 131 if (map_lapic_id(header, acpi_id, &apic_id)) 132 break; 133 } else if (header->type == ACPI_MADT_TYPE_LOCAL_X2APIC) { 134 if (map_x2apic_id(header, type, acpi_id, &apic_id)) 135 break; 136 } else if (header->type == ACPI_MADT_TYPE_LOCAL_SAPIC) { 137 if (map_lsapic_id(header, type, acpi_id, &apic_id)) 138 break; 139 } 140 entry += header->length; 141 } 142 return apic_id; 143} 144 145static int map_mat_entry(acpi_handle handle, int type, u32 acpi_id) 146{ 147 struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL }; 148 union acpi_object *obj; 149 struct acpi_subtable_header *header; 150 int apic_id = -1; 151 152 if (ACPI_FAILURE(acpi_evaluate_object(handle, "_MAT", NULL, &buffer))) 153 goto exit; 154 155 if (!buffer.length || !buffer.pointer) 156 goto exit; 157 158 obj = buffer.pointer; 159 if (obj->type != ACPI_TYPE_BUFFER || 160 obj->buffer.length < sizeof(struct acpi_subtable_header)) { 161 goto exit; 162 } 163 164 header = (struct acpi_subtable_header *)obj->buffer.pointer; 165 if (header->type == ACPI_MADT_TYPE_LOCAL_APIC) { 166 map_lapic_id(header, acpi_id, &apic_id); 167 } else if (header->type == ACPI_MADT_TYPE_LOCAL_SAPIC) { 168 map_lsapic_id(header, type, acpi_id, &apic_id); 169 } 170 171exit: 172 if (buffer.pointer) 173 kfree(buffer.pointer); 174 return apic_id; 175} 176 177int acpi_get_cpuid(acpi_handle handle, int type, u32 acpi_id) 178{ 179 int i; 180 int apic_id = -1; 181 182 apic_id = map_mat_entry(handle, type, acpi_id); 183 if (apic_id == -1) 184 apic_id = map_madt_entry(type, acpi_id); 185 if (apic_id == -1) 186 return apic_id; 187 188 for_each_possible_cpu(i) { 189 if (cpu_physical_id(i) == apic_id) 190 return i; 191 } 192 return -1; 193} 194EXPORT_SYMBOL_GPL(acpi_get_cpuid); 195#endif 196 197 198static void acpi_set_pdc_bits(u32 *buf) 199{ 200 buf[0] = ACPI_PDC_REVISION_ID; 201 buf[1] = 1; 202 203 /* Enable coordination with firmware's _TSD info */ 204 buf[2] = ACPI_PDC_SMP_T_SWCOORD; 205 206 /* Twiddle arch-specific bits needed for _PDC */ 207 arch_acpi_set_pdc_bits(buf); 208} 209 210static struct acpi_object_list *acpi_processor_alloc_pdc(void) 211{ 212 struct acpi_object_list *obj_list; 213 union acpi_object *obj; 214 u32 *buf; 215 216 /* allocate and initialize pdc. It will be used later. */ 217 obj_list = kmalloc(sizeof(struct acpi_object_list), GFP_KERNEL); 218 if (!obj_list) { 219 printk(KERN_ERR "Memory allocation error\n"); 220 return NULL; 221 } 222 223 obj = kmalloc(sizeof(union acpi_object), GFP_KERNEL); 224 if (!obj) { 225 printk(KERN_ERR "Memory allocation error\n"); 226 kfree(obj_list); 227 return NULL; 228 } 229 230 buf = kmalloc(12, GFP_KERNEL); 231 if (!buf) { 232 printk(KERN_ERR "Memory allocation error\n"); 233 kfree(obj); 234 kfree(obj_list); 235 return NULL; 236 } 237 238 acpi_set_pdc_bits(buf); 239 240 obj->type = ACPI_TYPE_BUFFER; 241 obj->buffer.length = 12; 242 obj->buffer.pointer = (u8 *) buf; 243 obj_list->count = 1; 244 obj_list->pointer = obj; 245 246 return obj_list; 247} 248 249/* 250 * _PDC is required for a BIOS-OS handshake for most of the newer 251 * ACPI processor features. 252 */ 253static int 254acpi_processor_eval_pdc(acpi_handle handle, struct acpi_object_list *pdc_in) 255{ 256 acpi_status status = AE_OK; 257 258 if (idle_nomwait) { 259 /* 260 * If mwait is disabled for CPU C-states, the C2C3_FFH access 261 * mode will be disabled in the parameter of _PDC object. 262 * Of course C1_FFH access mode will also be disabled. 263 */ 264 union acpi_object *obj; 265 u32 *buffer = NULL; 266 267 obj = pdc_in->pointer; 268 buffer = (u32 *)(obj->buffer.pointer); 269 buffer[2] &= ~(ACPI_PDC_C_C2C3_FFH | ACPI_PDC_C_C1_FFH); 270 271 } 272 status = acpi_evaluate_object(handle, "_PDC", pdc_in, NULL); 273 274 if (ACPI_FAILURE(status)) 275 ACPI_DEBUG_PRINT((ACPI_DB_INFO, 276 "Could not evaluate _PDC, using legacy perf. control.\n")); 277 278 return status; 279} 280 281static int early_pdc_done; 282 283void acpi_processor_set_pdc(acpi_handle handle) 284{ 285 struct acpi_object_list *obj_list; 286 287 if (arch_has_acpi_pdc() == false) 288 return; 289 290 if (early_pdc_done) 291 return; 292 293 obj_list = acpi_processor_alloc_pdc(); 294 if (!obj_list) 295 return; 296 297 acpi_processor_eval_pdc(handle, obj_list); 298 299 kfree(obj_list->pointer->buffer.pointer); 300 kfree(obj_list->pointer); 301 kfree(obj_list); 302} 303EXPORT_SYMBOL_GPL(acpi_processor_set_pdc); 304 305static int early_pdc_optin; 306static int set_early_pdc_optin(const struct dmi_system_id *id) 307{ 308 early_pdc_optin = 1; 309 return 0; 310} 311 312static int param_early_pdc_optin(char *s) 313{ 314 early_pdc_optin = 1; 315 return 1; 316} 317__setup("acpi_early_pdc_eval", param_early_pdc_optin); 318 319static struct dmi_system_id __cpuinitdata early_pdc_optin_table[] = { 320 { 321 set_early_pdc_optin, "HP Envy", { 322 DMI_MATCH(DMI_BIOS_VENDOR, "Hewlett-Packard"), 323 DMI_MATCH(DMI_PRODUCT_NAME, "HP Envy") }, NULL}, 324 { 325 set_early_pdc_optin, "HP Pavilion dv6", { 326 DMI_MATCH(DMI_BIOS_VENDOR, "Hewlett-Packard"), 327 DMI_MATCH(DMI_PRODUCT_NAME, "HP Pavilion dv6") }, NULL}, 328 { 329 set_early_pdc_optin, "HP Pavilion dv7", { 330 DMI_MATCH(DMI_BIOS_VENDOR, "Hewlett-Packard"), 331 DMI_MATCH(DMI_PRODUCT_NAME, "HP Pavilion dv7") }, NULL}, 332 {}, 333}; 334 335static acpi_status 336early_init_pdc(acpi_handle handle, u32 lvl, void *context, void **rv) 337{ 338 acpi_processor_set_pdc(handle); 339 return AE_OK; 340} 341 342void __init acpi_early_processor_set_pdc(void) 343{ 344 345#ifdef CONFIG_SMP 346 if (ACPI_FAILURE(acpi_get_table(ACPI_SIG_MADT, 0, 347 (struct acpi_table_header **)&madt))) 348 madt = NULL; 349#endif 350 351 /* 352 * Check whether the system is DMI table. If yes, OSPM 353 * should not use mwait for CPU-states. 354 */ 355 dmi_check_system(processor_idle_dmi_table); 356 357 /* 358 * Allow systems to opt-in to early _PDC evaluation. 359 */ 360 dmi_check_system(early_pdc_optin_table); 361 if (!early_pdc_optin) 362 return; 363 364 acpi_walk_namespace(ACPI_TYPE_PROCESSOR, ACPI_ROOT_OBJECT, 365 ACPI_UINT32_MAX, 366 early_init_pdc, NULL, NULL, NULL); 367 368 early_pdc_done = 1; 369} 370