/art/compiler/optimizing/ |
H A D | code_generator.h | 172 void Compile(CodeAllocator* allocator); 195 virtual void Finalize(CodeAllocator* allocator);
|
H A D | nodes.h | 1674 void SetAndCopyParentChain(ArenaAllocator* allocator, HEnvironment* parent) { argument 1676 parent_->SetAndCopyParentChain(allocator, parent); 1678 parent_ = new (allocator) HEnvironment(allocator, *parent, holder_); 1681 parent_->SetAndCopyParentChain(allocator, parent->GetParent()); 1913 ArenaAllocator* allocator = GetBlock()->GetGraph()->GetArena(); local 1914 environment_ = new (allocator) HEnvironment(allocator, *environment, this); 1917 environment_->SetAndCopyParentChain(allocator, environment->GetParent()); 1924 ArenaAllocator* allocator local [all...] |
H A D | intrinsics_mips64.cc | 1472 ArenaAllocator* allocator, 1488 slow_path = new (allocator) IntrinsicSlowPathMIPS64(invoke); 1497 slow_path = new (allocator) IntrinsicSlowPathMIPS64(invoke); 1469 GenerateStringIndexOf(HInvoke* invoke, Mips64Assembler* assembler, CodeGeneratorMIPS64* codegen, ArenaAllocator* allocator, bool start_at_zero) argument
|
H A D | intrinsics_x86.cc | 1374 ArenaAllocator* allocator, 1376 LocationSummary* locations = new (allocator) LocationSummary(invoke, 1382 // allocator to do that, anyways. We can still do the constant check by checking the parameter 1401 ArenaAllocator* allocator, 1428 slow_path = new (allocator) IntrinsicSlowPathX86(invoke); 1436 slow_path = new (allocator) IntrinsicSlowPathX86(invoke); 2197 // Check that the register allocator did not assign the location 1373 CreateStringIndexOfLocations(HInvoke* invoke, ArenaAllocator* allocator, bool start_at_zero) argument 1398 GenerateStringIndexOf(HInvoke* invoke, X86Assembler* assembler, CodeGeneratorX86* codegen, ArenaAllocator* allocator, bool start_at_zero) argument
|
H A D | intrinsics_x86_64.cc | 1476 ArenaAllocator* allocator, 1478 LocationSummary* locations = new (allocator) LocationSummary(invoke, 1484 // allocator to do that, anyways. We can still do the constant check by checking the parameter 1503 ArenaAllocator* allocator, 1530 slow_path = new (allocator) IntrinsicSlowPathX86_64(invoke); 1538 slow_path = new (allocator) IntrinsicSlowPathX86_64(invoke); 2221 // Check that the register allocator did not assign the location 1475 CreateStringIndexOfLocations(HInvoke* invoke, ArenaAllocator* allocator, bool start_at_zero) argument 1500 GenerateStringIndexOf(HInvoke* invoke, X86_64Assembler* assembler, CodeGeneratorX86_64* codegen, ArenaAllocator* allocator, bool start_at_zero) argument
|
H A D | intrinsics_arm.cc | 1115 ArenaAllocator* allocator, 1131 slow_path = new (allocator) IntrinsicSlowPathARM(invoke); 1141 slow_path = new (allocator) IntrinsicSlowPathARM(invoke); 1605 // the code generator. Furthermore, the register allocator creates fixed live intervals 1607 // the input and output locations are unallocated, the register allocator runs out of 1631 // the code generator. Furthermore, the register allocator creates fixed live intervals 1633 // the input and output locations are unallocated, the register allocator runs out of 1112 GenerateVisitStringIndexOf(HInvoke* invoke, ArmAssembler* assembler, CodeGeneratorARM* codegen, ArenaAllocator* allocator, bool start_at_zero) argument
|
H A D | intrinsics_mips.cc | 2063 ArenaAllocator* allocator) { 2078 slow_path = new (allocator) IntrinsicSlowPathMIPS(invoke); 2095 slow_path = new (allocator) IntrinsicSlowPathMIPS(invoke); 2059 GenerateStringIndexOf(HInvoke* invoke, bool start_at_zero, MipsAssembler* assembler, CodeGeneratorMIPS* codegen, ArenaAllocator* allocator) argument
|
H A D | code_generator_mips64.cc | 430 void CodeGeneratorMIPS64::Finalize(CodeAllocator* allocator) { argument 457 CodeGenerator::Finalize(allocator); 4043 // Nothing to do, this should be removed during prepare for register allocator. 4048 // Nothing to do, this should be removed during prepare for register allocator.
|
H A D | code_generator_arm64.cc | 704 // the register allocator instead; but we prefer not to, as 936 void CodeGeneratorARM64::Finalize(CodeAllocator* allocator) { argument 941 CodeGenerator::Finalize(allocator); 2255 // TODO: Extend the register allocator to support a new 4729 // Nothing to do, this should be removed during prepare for register allocator. 4734 // Nothing to do, this should be removed during prepare for register allocator.
|
H A D | code_generator_mips.cc | 484 void CodeGeneratorMIPS::Finalize(CodeAllocator* allocator) { argument 511 CodeGenerator::Finalize(allocator); 5039 // Nothing to do, this should be removed during prepare for register allocator. 5044 // Nothing to do, this should be removed during prepare for register allocator.
|
H A D | code_generator_arm.cc | 56 // D31 cannot be split into two S registers, and the register allocator only works on 531 // the register allocator instead; but we prefer not to, as 799 void CodeGeneratorARM::Finalize(CodeAllocator* allocator) { argument 824 CodeGenerator::Finalize(allocator); 3288 // don't clash with high registers which the register allocator currently guarantees. 3394 // Register allocator doesn't create partial overlap. 4595 // TODO: Extend the register allocator to support a new 6623 // Nothing to do, this should be removed during prepare for register allocator. 6628 // Nothing to do, this should be removed during prepare for register allocator.
|
H A D | code_generator_x86.cc | 548 // the register allocator instead; but we prefer not to, as 2189 // Make the output overlap to please the register allocator. This greatly simplifies 2201 // Make the output overlap to please the register allocator. This greatly simplifies 2623 // TODO: enhance register allocator to ask for stack temporaries. 2676 // TODO: enhance register allocator to ask for stack temporaries. 3185 // TODO: enhance register allocator to ask for stack temporaries. 4699 // The register allocator does not support multiple 4718 // NB: We could make the register allocator understand fp_reg <-> core_reg moves but given the 5218 // In case of a byte operation, the register allocator does not support multiple 5340 // TODO: Extend the register allocator t 7286 Finalize(CodeAllocator* allocator) argument [all...] |
H A D | code_generator_x86_64.cc | 569 // the register allocator instead; but we prefer not to, as 3235 // TODO: enhance register allocator to ask for stack temporaries. 4830 // TODO: Extend the register allocator to support a new 6468 // Nothing to do, this should be removed during prepare for register allocator. 6473 // Nothing to do, this should be removed during prepare for register allocator. 6709 void CodeGeneratorX86_64::Finalize(CodeAllocator* allocator) { argument 6727 CodeGenerator::Finalize(allocator);
|
/art/runtime/gc/allocator/ |
H A D | rosalloc.h | 29 #include "base/allocator.h" 41 namespace allocator { namespace in namespace:art::gc 43 // A runs-of-slots memory allocator. 695 // The base address of the memory region that's managed by this allocator. 703 // the end of the memory region that's currently managed by this allocator. 707 // the end of the memory region that's ever managed by this allocator. 767 // Whether this allocator is running under Valgrind. 770 // The base address of the memory region that's managed by this allocator. 772 // The end address of the memory region that's managed by this allocator. 843 // If kThreadUnsafe is true then the allocator ma [all...] |
H A D | rosalloc.cc | 36 namespace allocator { namespace in namespace:art::gc 2193 } // namespace allocator
|
/art/runtime/base/ |
H A D | hash_set.h | 60 class Pred = std::equal_to<T>, class Alloc = std::allocator<T>>
|
/art/runtime/ |
H A D | thread.cc | 47 #include "gc/allocator/rosalloc.h" 1623 gc::allocator::RosAlloc::GetDedicatedFullRun());
|