/external/compiler-rt/SDKs/linux/usr/include/sys/ |
H A D | mman.h | 24 #define PROT_EXEC 0x4 macro
|
/external/valgrind/main/none/tests/ |
H A D | munmap_exe.c | 6 /* Point of this is that the fd of an PROT_EXEC segment is -1, so Valgrind 14 m = mmap(NULL, 100, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0);
|
/external/libffi/testsuite/libffi.special/ |
H A D | ffitestcxx.h | 69 page = mmap (NULL, size, PROT_READ | PROT_WRITE | PROT_EXEC, 73 page = mmap (NULL, size, PROT_READ | PROT_WRITE | PROT_EXEC,
|
/external/compiler-rt/lib/ |
H A D | enable_execute_stack.c | 56 (void) mprotect((void *)startPage, length, PROT_READ | PROT_WRITE | PROT_EXEC);
|
/external/kernel-headers/original/asm-generic/ |
H A D | mman-common.h | 11 #define PROT_EXEC 0x4 /* page can be executed */ macro
|
/external/libffi/testsuite/libffi.call/ |
H A D | ffitest.h | 69 page = mmap (NULL, size, PROT_READ | PROT_WRITE | PROT_EXEC, 73 page = mmap (NULL, size, PROT_READ | PROT_WRITE | PROT_EXEC,
|
/external/valgrind/main/memcheck/tests/ |
H A D | mempool.c | 37 p = (pool *)mmap(0, sizeof(pool), PROT_READ|PROT_WRITE|PROT_EXEC, 40 PROT_READ|PROT_WRITE|PROT_EXEC, 59 PROT_READ|PROT_WRITE|PROT_EXEC,
|
H A D | sigaltstack.c | 21 char *stk = (char *)mmap(0, size, PROT_READ|PROT_WRITE|PROT_EXEC,
|
H A D | mempool2.c | 40 p = (pool *)mmap(0, sizeof(pool), PROT_READ|PROT_WRITE|PROT_EXEC, 43 PROT_READ|PROT_WRITE|PROT_EXEC, 62 PROT_READ|PROT_WRITE|PROT_EXEC,
|
H A D | custom_alloc.c | 16 void* p = mmap( 0, SUPERBLOCK_SIZE, PROT_READ|PROT_WRITE|PROT_EXEC,
|
/external/kernel-headers/original/asm-mips/ |
H A D | mman.h | 13 * implementation does not necessarily support PROT_EXEC or PROT_WRITE 20 #define PROT_EXEC 0x04 /* page can be executed */ macro
|
/external/qemu/android/utils/ |
H A D | mapfile.h | 31 #define PROT_EXEC 0x4 macro
|
H A D | mapfile.c | 193 if ((prot & (PROT_WRITE | PROT_EXEC)) == 0) { 196 } else if ((prot & (PROT_WRITE | PROT_EXEC)) == 197 (PROT_WRITE | PROT_EXEC)) { 203 } else if ((prot & PROT_EXEC) == PROT_EXEC) {
|
/external/valgrind/main/drd/tests/ |
H A D | custom_alloc.c | 16 void* p = mmap( 0, SUPERBLOCK_SIZE, PROT_READ|PROT_WRITE|PROT_EXEC,
|
/external/valgrind/main/massif/tests/ |
H A D | custom_alloc.c | 16 void* p = mmap( 0, SUPERBLOCK_SIZE, PROT_READ|PROT_WRITE|PROT_EXEC,
|
/external/webkit/Source/JavaScriptCore/wtf/ |
H A D | PageAllocationAligned.cpp | 47 protection |= PROT_EXEC; 50 vm_map(current_task(), &address, size, alignmentMask, flags, MEMORY_OBJECT_NULL, 0, FALSE, protection, PROT_READ | PROT_WRITE | PROT_EXEC, VM_INHERIT_DEFAULT); local
|
H A D | OSAllocatorPosix.cpp | 53 protection |= PROT_EXEC;
|
/external/compiler-rt/test/Unit/ |
H A D | clear_cache_test.c | 57 if (mprotect(start, end-start, PROT_READ|PROT_WRITE|PROT_EXEC) != 0)
|
/external/llvm/lib/Support/Unix/ |
H A D | Memory.inc | 61 void *pa = ::mmap(start, pageSize*NumPages, PROT_READ|PROT_EXEC, 64 void *pa = ::mmap(start, pageSize*NumPages, PROT_READ|PROT_WRITE|PROT_EXEC,
|
/external/valgrind/main/memcheck/tests/linux/ |
H A D | stack_changes.c | 37 stack = (void *)mmap(0, STACK_SIZE, PROT_READ|PROT_WRITE|PROT_EXEC,
|
/external/valgrind/main/none/tests/amd64/ |
H A D | smc1.c | 103 code = mmap(NULL, 20, PROT_READ|PROT_WRITE|PROT_EXEC,
|
/external/valgrind/main/perf/ |
H A D | bigcode.c | 39 PROT_EXEC|PROT_WRITE,
|
/external/libffi/src/ |
H A D | closures.c | 51 don't attempt PROT_EXEC|PROT_WRITE mapping at all, as that 374 ptr = mmap (NULL, length, (prot & ~PROT_WRITE) | PROT_EXEC, 425 ptr = mmap (start, length, prot | PROT_EXEC, flags, fd, offset); 432 with ((prot & ~PROT_WRITE) | PROT_EXEC) and mremap with
|
/external/oprofile/module/ia64/ |
H A D | op_syscalls.c | 215 if ((prot & PROT_EXEC) && ret >= 0) 228 if ((prot & PROT_EXEC) && ret >= 0)
|
/external/oprofile/module/x86/ |
H A D | op_syscalls.c | 175 if ((prot & PROT_EXEC) && ret >= 0) 199 if (a.prot&PROT_EXEC)
|