/arch/m68k/mac/ |
H A D | oss.c | 69 int events = oss->irq_pending & local 73 if ((console_loglevel == 10) && !(events & OSS_IP_SCSI)) { 74 printk("oss_irq: irq %u events = 0x%04X\n", irq, 79 if (events & OSS_IP_IOPSCC) { 84 if (events & OSS_IP_SCSI) { 89 if (events & OSS_IP_IOPISM) { 103 int events, irq_bit, i; local 105 events = oss->irq_pending & OSS_IP_NUBUS; 106 if (!events) 111 printk("oss_nubus_irq: events [all...] |
H A D | baboon.c | 51 unsigned char events; local 59 events = baboon->mb_ifr & 0x07; 60 if (!events) 66 if (events & irq_bit) { 72 } while(events >= irq_bit); 76 baboon->mb_ifr &= ~events;
|
H A D | via.c | 452 unsigned char irq_bit, events; local 454 events = via1[vIFR] & via1[vIER] & 0x7F; 455 if (!events) 461 if (events & irq_bit) { 467 } while (events >= irq_bit); 473 unsigned char irq_bit, events; local 475 events = via2[gIFR] & via2[gIER] & 0x7F; 476 if (!events) 482 if (events & irq_bit) { 488 } while (events > 499 unsigned char slot_bit, events; local [all...] |
H A D | psc.c | 122 unsigned char irq_bit, events; local 129 events = psc_read_byte(pIFR) & psc_read_byte(pIER) & 0xF; 130 if (!events) 136 if (events & irq_bit) { 142 } while (events >= irq_bit);
|
/arch/sparc/kernel/ |
H A D | sparc_ksyms_32.c | 18 short events; member in struct:poll
|
H A D | sparc_ksyms_64.c | 21 short events; member in struct:poll
|
/arch/um/include/shared/ |
H A D | irq_user.h | 17 int events; member in struct:irq_fd
|
/arch/arm/kernel/ |
H A D | perf_event_xscale.c | 166 struct perf_event *event = cpuc->events[idx]; 201 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 224 raw_spin_lock_irqsave(&events->pmu_lock, flags); 229 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); 237 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 258 raw_spin_lock_irqsave(&events->pmu_lock, flags); 263 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); 290 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 292 raw_spin_lock_irqsave(&events->pmu_lock, flags); 296 raw_spin_unlock_irqrestore(&events 302 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 542 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 588 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 654 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 666 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local [all...] |
H A D | perf_event_v6.c | 13 * performance counters can export events to the event bus, and the event bus 14 * itself can be monitored. This requires that we *don't* export the events to 63 * The hardware events that we support. We do support cache operations but 131 * The hardware events that we support. We do support cache operations but 265 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 288 raw_spin_lock_irqsave(&events->pmu_lock, flags); 293 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); 320 struct perf_event *event = cpuc->events[idx]; 345 * Handle the pending perf events. 359 struct pmu_hw_events *events local 371 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 412 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 447 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local [all...] |
H A D | sys_oabi-compat.c | 248 __u32 events; member in struct:oabi_epoll_event 264 kernel.events = user.events; 274 struct oabi_epoll_event __user *events, 292 __put_user_error(kbuf[i].events, &events->events, err); 293 __put_user_error(kbuf[i].data, &events->data, err); 294 events++; 273 sys_oabi_epoll_wait(int epfd, struct oabi_epoll_event __user *events, int maxevents, int timeout) argument
|
H A D | perf_event_cpu.c | 123 printk_once("perf/ARM: No irqs for PMU defined, sampling events not supported\n"); 174 struct pmu_hw_events *events = &per_cpu(cpu_hw_events, cpu); local 175 events->events = per_cpu(hw_events, cpu); 176 events->used_mask = per_cpu(used_mask, cpu); 177 raw_spin_lock_init(&events->pmu_lock);
|
H A D | perf_event_v7.c | 28 * Note: An implementation may not be able to count all of these events 61 /* These events are defined by the PMUv2 supplement (ARM DDI 0457A). */ 144 * Cortex-A8 HW events mapping 146 * The hardware events that we support. We do support cache operations but 197 * Cortex-A9 HW events mapping 241 * Cortex-A5 HW events mapping 287 * Cortex-A15 HW events mapping 336 * Cortex-A7 HW events mapping 385 * Cortex-A12 HW events mapping 435 * Krait HW events mappin 708 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 754 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 846 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 857 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 1290 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 1316 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local [all...] |
/arch/arm64/kernel/ |
H A D | trace-events-emulation.h | 34 #define TRACE_INCLUDE_FILE trace-events-emulation
|
H A D | perf_event.c | 40 * ARMv8 supports a maximum of 32 events. 274 hw_events->events[idx] = NULL; 304 hw_events->events[idx] = event; 638 /* Required events. */ 650 /* Common architectural events. */ 662 /* Common microarchitectural events. */ 677 /* PMUv3 HW events mapping. */ 1017 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 1023 raw_spin_lock_irqsave(&events->pmu_lock, flags); 1045 raw_spin_unlock_irqrestore(&events 1051 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 1136 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 1147 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); local 1316 struct pmu_hw_events *events = &per_cpu(cpu_hw_events, cpu); local [all...] |
/arch/x86/xen/ |
H A D | multicalls.h | 4 #include <trace/events/xen.h>
|
H A D | suspend.c | 6 #include <xen/events.h>
|
H A D | trace.c | 62 #include <trace/events/xen.h>
|
/arch/um/os-Linux/ |
H A D | irq.c | 52 int os_create_pollfd(int fd, int events, void *tmp_pfd, int size_tmpfds) argument 72 .events = events,
|
/arch/arm/mach-omap2/ |
H A D | prm44xx_54xx.h | 55 extern void omap44xx_prm_read_pending_irqs(unsigned long *events);
|
/arch/powerpc/include/asm/ |
H A D | fsl_gtm.h | 45 extern void gtm_ack_timer16(struct gtm_timer *tmr, u16 events);
|
/arch/metag/kernel/perf/ |
H A D | perf_event.c | 288 cpuc->events[idx] = event; 357 cpuc->events[idx] = NULL; 600 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); local 605 raw_spin_lock_irqsave(&events->pmu_lock, flags); 668 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); 673 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); local 692 * least two events that count events that are core global and ignore 696 raw_spin_lock_irqsave(&events->pmu_lock, flags); 702 raw_spin_unlock_irqrestore(&events 721 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); local [all...] |
H A D | perf_event.h | 26 * one time, with the returned count being an aggregate of events. A small 27 * number of events are thread global, i.e. they count the aggregate of all 28 * threads' events, regardless of the thread selected. 42 * struct cpu_hw_events - a processor core's performance events 43 * @events: an array of perf_events active for a given index. 51 struct perf_event *events[MAX_HWEVENTS]; member in struct:cpu_hw_events
|
/arch/um/kernel/ |
H A D | irq.c | 68 int events, err, n; local 80 events = UM_POLLIN | UM_POLLPRI; 81 else events = UM_POLLOUT; 87 .events = events, 109 n = os_create_pollfd(fd, events, tmp_pfd, n);
|
/arch/arm64/include/asm/ |
H A D | pmu.h | 24 /* The events for a given PMU register set. */ 27 * The events that are active on the PMU for the given index. 29 struct perf_event **events; member in struct:pmu_hw_events
|
/arch/powerpc/platforms/powernv/ |
H A D | opal.c | 251 static void opal_do_notifier(uint64_t events) argument 260 changed_mask = last_notified_mask ^ events; 261 last_notified_mask = events; 269 events, (void *)changed_mask); 290 /* Process pending events */ 362 unsigned long events, void *change) 364 if (events & OPAL_EVENT_MSG_PENDING) 609 __be64 events; local 611 opal_handle_interrupt(virq_to_hw(irq), &events); local 613 opal_do_notifier(be64_to_cpu(events)); 361 opal_message_notify(struct notifier_block *nb, unsigned long events, void *change) argument [all...] |