Lines Matching refs:r9

74     ldr         r9, [src], pstep            ; p3
89 uqsub8 r6, r9, r10 ; p3 - p2
90 uqsub8 r7, r10, r9 ; p2 - p3
101 ldr r9, [src], pstep ; q0
110 uqsub8 r11, r12, r9 ; p0 - q0
111 uqsub8 r12, r9, r12 ; q0 - p0
118 uqsub8 r7, r9, r10 ; q0 - q1
120 uqsub8 r6, r10, r9 ; q1 - q0
122 uqsub8 r9, r11, r10 ; q2 - q1
129 orr r10, r9, r10 ; abs (q2-q1)
137 uqsub8 r9, r11, r12 ; q2 - q3
141 orr r10, r10, r9 ; abs (q3-q2)
168 ldr r9, [src], pstep ; q0
173 eor r9, r9, r12 ; q0 offset to convert to a signed value
176 str r9, [sp] ; store qs0 temporarily
182 qsub8 r8, r9, r8 ; vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
187 ldr r9, c0x03030303 ; r9 = 3 --modified for vp8
196 qadd8 r8 , r7 , r9 ; Filter2 (r8) = vp8_signed_char_clamp(vp8_filter+3)
199 mov r9, #0
200 shadd8 r8 , r8 , r9 ; Filter2 >>= 3
201 shadd8 r7 , r7 , r9 ; vp8_filter >>= 3
202 shadd8 r8 , r8 , r9
203 shadd8 r7 , r7 , r9
204 shadd8 lr , r8 , r9 ; lr: Filter2
205 shadd8 r7 , r7 , r9 ; r7: filter
208 ;sel lr, r11, r9
210 ;sel r8, r11, r9
217 ldr r9, [sp, #4] ; load ps0
222 qadd8 r9, r9, lr ; u = vp8_signed_char_clamp(ps0 + Filter2)
241 eor r9, r9, r12 ; *op0 = u^0x80
242 str r9, [src], pstep ; store op0 result
262 ldrne r9, [src], pstep ; p3
283 ldr r9, [src], pstep ; p3
299 uqsub8 r6, r9, r10 ; p3 - p2
300 uqsub8 r7, r10, r9 ; p2 - p3
312 ldr r9, [src], pstep ; q0
321 uqsub8 r11, r12, r9 ; p0 - q0
322 uqsub8 r12, r9, r12 ; q0 - p0
329 uqsub8 r7, r9, r10 ; q0 - q1
331 uqsub8 r6, r10, r9 ; q1 - q0
333 uqsub8 r9, r11, r10 ; q2 - q1
341 orr r10, r9, r10 ; abs (q2-q1)
349 uqsub8 r9, r11, r12 ; q2 - q3
353 orr r10, r10, r9 ; abs (q3-q2)
381 ldr r9, [src], pstep ; q0
386 eor r9, r9, r12 ; qs0
389 qsub8 r12, r9, r8 ; vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
394 str r9, [sp] ; store qs0 temporarily
400 ldr r9, c0x04040404
409 qadd8 r8 , r12 , r9 ; Filter1 (r8) = vp8_signed_char_clamp(Filter2+4)
420 ldr r9, [sp] ; load qs0
423 qsub8 r9 , r9, r8 ; qs0 = vp8_signed_char_clamp(qs0 - Filter1)
428 ;qadd8 r12 , r12 , r9 ; Filter2 = vp8_signed_char_clamp(Filter2+4)
431 ;usub8 lr, r8, r9 ; s = (s==4)*-1
434 ;usub8 r8, r9, r8
436 ;ldr r9, [sp] ; load qs0
441 ;qsub8 r9 , r9, r12 ; qs0 = vp8_signed_char_clamp(qs0 - Filter2)
476 qsub8 r8, r9, r10 ; s = vp8_signed_char_clamp(qs0 - u)
492 smlabb r9, r10, lr, r7
496 ssat r9, #8, r9, asr #7
502 pkhbt r10, r9, r10, lsl #16
504 ldr r9, [sp, #8] ; load qs1
515 qsub8 r8, r9, r10 ; s = vp8_signed_char_clamp(qs1 - u)
527 ldr r9, [src] ; load q2
553 eor r9, r9, lr
559 qsub8 r10, r9, r10 ; s = vp8_signed_char_clamp(qs2 - u)
571 ldrne r9, [src], pstep ; p3
608 TRANSPOSE_MATRIX r6, r7, r8, lr, r9, r10, r11, r12
610 uqsub8 r7, r9, r10 ; p3 - p2
611 uqsub8 r8, r10, r9 ; p2 - p3
612 uqsub8 r9, r10, r11 ; p2 - p1
615 orr r10, r9, r10 ; abs (p2-p1)
642 TRANSPOSE_MATRIX r6, r7, r8, lr, r9, r10, r11, r12
659 uqsub8 r6, r11, r9 ; p0 - q0
660 uqsub8 r7, r9, r11 ; q0 - p0
668 uqsub8 r11, r10, r9 ; q1 - q0
670 uqsub8 r12, r9, r10 ; q0 - q1
673 orr r9, r11, r12 ; abs (q1-q0)
674 uqsub8 r8, r9, r2 ; compare to limit
675 uqsub8 r10, r9, r3 ; compare to thresh
683 ldr r9, [sp] ; load the compared result
694 orr r9, r9, r10
699 usub8 r9, r12, r9
704 ldrh r9, [src, #-2]
712 pkhbt r11, r9, r10, lsl #16
714 ldrh r9, [src, #-2]
722 pkhbt lr, r9, r10, lsl #16
724 ;transpose r12, r11, r6, lr to r7, r8, r9, r10
725 TRANSPOSE_MATRIX r12, r11, r6, lr, r7, r8, r9, r10
734 eor r9, r9, r12 ; q0 offset to convert to a signed value
737 str r9, [sp] ; store qs0 temporarily
743 qsub8 r8, r9, r8 ; vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
748 ldr r9, c0x03030303 ; r9 = 3 --modified for vp8
759 qadd8 r8 , r7 , r9 ; Filter2 (r8) = vp8_signed_char_clamp(vp8_filter+3)
762 mov r9, #0
763 shadd8 r8 , r8 , r9 ; Filter2 >>= 3
764 shadd8 r7 , r7 , r9 ; vp8_filter >>= 3
765 shadd8 r8 , r8 , r9
766 shadd8 r7 , r7 , r9
767 shadd8 lr , r8 , r9 ; lr: filter2
768 shadd8 r7 , r7 , r9 ; r7: filter
771 ;sel lr, r11, r9
773 ;sel r8, r11, r9
780 ldr r9, [sp, #4] ; load ps0
785 qadd8 r9, r9, lr ; u = vp8_signed_char_clamp(ps0 + Filter2)
789 eor r9, r9, r12
816 TRANSPOSE_MATRIX r11, r9, r8, r10, r6, r7, r12, lr
877 TRANSPOSE_MATRIX r6, r7, r8, lr, r9, r10, r11, r12
879 uqsub8 r7, r9, r10 ; p3 - p2
880 uqsub8 r8, r10, r9 ; p2 - p3
881 uqsub8 r9, r10, r11 ; p2 - p1
884 orr r10, r9, r10 ; abs (p2-p1)
911 TRANSPOSE_MATRIX r6, r7, r8, lr, r9, r10, r11, r12
928 uqsub8 r6, r11, r9 ; p0 - q0
929 uqsub8 r7, r9, r11 ; q0 - p0
937 uqsub8 r11, r10, r9 ; q1 - q0
939 uqsub8 r12, r9, r10 ; q0 - q1
942 orr r9, r11, r12 ; abs (q1-q0)
943 uqsub8 r8, r9, r2 ; compare to limit
944 uqsub8 r10, r9, r3 ; compare to thresh
952 ldr r9, [sp] ; load the compared result
964 orr r9, r9, r10
969 usub8 r9, r12, r9
977 ldrh r9, [src, #-2]
985 pkhbt r11, r9, r10, lsl #16
987 ldrh r9, [src, #-2]
994 pkhbt lr, r9, r10, lsl #16
997 TRANSPOSE_MATRIX r12, r11, r6, lr, r7, r8, r9, r10
1006 eor r9, r9, r12 ; qs0
1009 qsub8 r12, r9, r8 ; vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
1014 str r9, [sp] ; store qs0 temporarily
1020 ldr r9, c0x04040404
1030 qadd8 r8 , r12 , r9 ; Filter1 (r8) = vp8_signed_char_clamp(Filter2+4)
1041 ldr r9, [sp] ; load qs0
1044 qsub8 r9 , r9, r8 ; qs0 = vp8_signed_char_clamp(qs0 - Filter1)
1049 ;qadd8 r12 , r12 , r9 ; Filter2 = vp8_signed_char_clamp(Filter2+4)
1052 ;usub8 lr, r8, r9 ; s = (s==4)*-1
1055 ;usub8 r8, r9, r8
1057 ;ldr r9, [sp] ; load qs0
1062 ;qsub8 r9 , r9, r12 ; qs0 = vp8_signed_char_clamp(qs0 - Filter2)
1097 qsub8 r8, r9, r10 ; s = vp8_signed_char_clamp(qs0 - u)
1125 smlabb r9, r10, lr, r7
1129 ssat r9, #8, r9, asr #7
1135 pkhbt r10, r9, r10, lsl #16
1137 ldr r9, [sp, #8] ; load qs1
1148 qsub8 r8, r9, r10 ; s = vp8_signed_char_clamp(qs1 - u)
1156 ldrb r9, [src], pstep ; load q2 for 1/7th difference across boundary
1169 orr r9, r9, r7, lsl #8
1179 orr r9, r9, r7, lsl #16
1186 orr r9, r9, r7, lsl #24
1189 eor r9, r9, lr
1219 qsub8 r10, r9, r10 ; s = vp8_signed_char_clamp(qs2 - u)