Searched refs:pinned (Results 1 - 19 of 19) sorted by relevance

/external/ltp/testcases/kernel/power_management/
H A Dpm_cpu_consolidation.py63 pinned="no"
65 trigger_ebizzy (options.smt_value, "partial", duration, background, pinned)
74 trigger_kernbench (options.smt_value, "partial", background, pinned, "no")
124 pinned ="no"
131 trigger_workld( options.smt_value, options.work_ld, options.stress, duration, background, pinned, perf_test)
H A Dpm_ilb_test.py42 pinned="yes"
44 trigger_workld(options.smt_level,options.work_ld, "single_job", duration, background, pinned, "no")
/external/strace/
H A Dperf_event_struct.h18 pinned :1, member in struct:perf_event_attr
H A Dperf.c266 ", pinned=%u"
282 attr->pinned,
/external/linux-kselftest/tools/testing/selftests/powerpc/pmu/ebb/
H A Dcpu_event_pinned_vs_ebb_test.c18 * Tests a pinned cpu event vs an EBB - in that order. The pinned cpu event
26 event->attr.pinned = 1;
H A Dtask_event_pinned_vs_ebb_test.c18 * Tests a pinned per-task event vs an EBB - in that order. The pinned per-task
26 event->attr.pinned = 1;
H A Devent_attributes_test.c117 leader.attr.pinned = 0;
118 /* Expected to fail, leader isn't pinned */
H A Debb.c346 e->attr.pinned = 1;
/external/ltp/testcases/kernel/power_management/lib/
H A Dpm_sched_mc.py315 def trigger_ebizzy (sched_smt, stress, duration, background, pinned):
341 if pinned == "yes":
366 def trigger_kernbench (sched_smt, stress, background, pinned, perf_test):
407 if pinned == "yes":
437 def trigger_workld(sched_smt, workload, stress, duration, background, pinned, perf_test):
443 trigger_ebizzy (sched_smt, stress, duration, background, pinned)
445 trigger_kernbench (sched_smt, stress, background, pinned, perf_test)
/external/tensorflow/tensorflow/contrib/verbs/
H A DREADME.md27 TensorFlow dynamically allocates memory for tensors that are to be sent or received. This causes difficulty for RDMA operations where pinned memory is required. Few remedies are possible:
28 1. The memory is pinned, transferred, then unpinned for each and every tensor to be transferred. This incurs significant operation overhead since pinning and unpinning memory for each dynamically generated tensor is slow.
29 2. Buffer is pre-allocated and pinned for each tensor. This incurs large memory overhead and extra copying from the tensor to its pinned buffer, but may still be faster than the former.
/external/libevent/
H A Dbuffer.c282 /* Free all trailing chains in 'buf' that are neither pinned nor empty, prior
805 * read-pinned chains. The first pinned chain is saved in first, and the
806 * last in last. If src has no read-pinned chains, first and last are set
812 struct evbuffer_chain *chain, **pinned; local
821 pinned = src->last_with_datap;
822 if (!CHAIN_PINNED_R(*pinned))
823 pinned = &(*pinned)->next;
824 EVUTIL_ASSERT(CHAIN_PINNED_R(*pinned));
853 RESTORE_PINNED(struct evbuffer *src, struct evbuffer_chain *pinned, struct evbuffer_chain *last) argument
963 struct evbuffer_chain *pinned, *last; local
1052 struct evbuffer_chain *pinned, *last; local
[all...]
/external/perf_data_converter/src/quipper/kernel/
H A Dperf_event.h269 pinned : 1, /* must always be on PMU */ member in struct:perf_event_attr
/external/strace/tests/
H A Dperf_event_open.c80 pinned :1, member in struct:pea_flags
228 ", pinned=%u"
243 attr->pinned,
/external/strace/tests-m32/
H A Dperf_event_open.c80 pinned :1, member in struct:pea_flags
228 ", pinned=%u"
243 attr->pinned,
/external/strace/tests-mx32/
H A Dperf_event_open.c80 pinned :1, member in struct:pea_flags
228 ", pinned=%u"
243 attr->pinned,
/external/kernel-headers/original/uapi/linux/
H A Dperf_event.h335 pinned : 1, /* must always be on PMU */ member in struct:perf_event_attr
/external/perf_data_converter/src/quipper/
H A Dperf_serializer.cc71 S(pinned);
123 S(pinned);
/external/v8/tools/gcmole/
H A Dgcmole.lua337 -- It is pinned and always present.
/external/valgrind/include/vki/
H A Dvki-linux.h2924 pinned : 1, /* must always be on PMU */ member in struct:vki_perf_event_attr

Completed in 416 milliseconds