Searched defs:start_index (Results 1 - 5 of 5) sorted by relevance
/art/compiler/dex/quick/ |
H A D | gen_invoke.cc | 762 const size_t start_index = skip_this ? 1 : 0; local 773 size_t stack_map_start = std::max(in_to_reg_storage_mapping.GetEndMappedIn(), start_index); 795 for (size_t i = start_index; i < stack_map_start + regs_left_to_pass_via_stack; i++) { 838 for (size_t i = start_index; i < stack_map_start; i++) {
|
/art/compiler/optimizing/ |
H A D | builder.cc | 741 size_t start_index = 0; local 749 start_index = 1; 753 uint32_t argument_index = start_index; 755 start_index = 1; 757 for (size_t i = start_index;
|
H A D | intrinsics_x86.cc | 1034 Register start_index = locations->InAt(2).AsRegister<Register>(); local 1036 // Do a start_index check. 1037 __ cmpl(start_index, string_length); 1042 __ cmpl(start_index, Immediate(0)); 1043 __ cmovl(kGreater, counter, start_index); 1045 // Move to the start of the string: string_obj + value_offset + 2 * start_index. 1048 // Now update ecx (the repne scasw work counter). We have string.length - start_index left to
|
H A D | intrinsics_x86_64.cc | 959 CpuRegister start_index = locations->InAt(2).AsRegister<CpuRegister>(); local 961 // Do a start_index check. 962 __ cmpl(start_index, string_length); 967 __ cmpl(start_index, Immediate(0)); 968 __ cmov(kGreater, counter, start_index, false); // 32-bit copy is enough. 970 // Move to the start of the string: string_obj + value_offset + 2 * start_index. 973 // Now update ecx, the work counter: it's gonna be string.length - start_index.
|
/art/runtime/ |
H A D | utils.cc | 892 unsigned int start_index = 0; local 896 while (start_index < s.size()) { 897 if (!isspace(s[start_index])) { 900 start_index++; 904 while (end_index >= start_index) { 912 if (end_index < start_index) { 916 return s.substr(start_index, end_index - start_index + 1);
|
Completed in 123 milliseconds