/external/chromium_org/v8/test/mjsunit/compiler/ |
H A D | regress-gap.js | 39 function select(n, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10) { 50 v9 = v10; 51 v10 = tmp; 56 function select_while(n, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10) { 68 v9 = v10; 69 v10 = tmp;
|
/external/chromium_org/third_party/libjpeg_turbo/simd/ |
H A D | jsimd_arm64_neon.S | 330 mov v10.16b, v14.16b 339 smlal v10.4s, ROW5L.4h, XFIX_2_053119869_MINUS_2_562915447 341 smlsl v10.4s, ROW3L.4h, XFIX_2_562915447 351 add v2.4s, v6.4s, v10.4s 353 sub v6.4s, v6.4s, v10.4s 354 saddl v10.4s, ROW0L.4h, ROW4L.4h 360 shl v10.4s, v10.4s, #13 363 add v4.4s, v10.4s, v12.4s 366 sub v2.4s, v10 [all...] |
/external/clang/test/CodeGen/ |
H A D | vector-alignment.c | 31 double __attribute__((vector_size(40))) v10; variable 32 // CHECK: @v10 {{.*}}, align 32
|
/external/llvm/test/MC/AArch64/ |
H A D | neon-scalar-by-elem-mul.s | 27 fmulx s13, s21, v10.s[3] 34 // CHECK: fmulx s13, s21, v10.s[3] // encoding: [0xad,0x9a,0xaa,0x7f]
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/ppc/ |
H A D | variance_subpixel_altivec.asm | 205 load_c v10, b_0123_b, 0, r12, r0 208 hfilter_8 v0, v10, v11, 1 209 hfilter_8 v1, v10, v11, 1 210 hfilter_8 v2, v10, v11, 1 211 hfilter_8 v3, v10, v11, 1 219 hfilter_8 v4, v10, v11, 0 261 load_c v10, b_hilo_b, 0, r12, r0 263 vperm v0, v0, v1, v10 264 vperm v1, v2, v3, v10 296 load_c v10, b_0123_ [all...] |
H A D | loopfilter_filters_altivec.asm | 118 Tpair v20,v21, v2,v10 132 Tpair v10,v11, v21,v29 244 ;# The input/output is in registers v0...v7. We use v10...v17 as mirrors; 248 ;# result lands in the "mirror" registers v10...v17 251 Tpair v10, v11, v0, v4 258 Tpair v0, v1, v10, v14 288 vmrghb v10, v21, v29 298 vmrghb v20, v2, v10 299 vmrglb v21, v2, v10 320 vmrghb v10, v2 [all...] |
H A D | filter_bilinear_altivec.asm | 79 vperm v24, v21, v21, v10 ;# v20 = 0123 1234 2345 3456 178 load_c v10, b_0123_b, 0, r9, r12 263 load_c v10, b_0123_b, 0, r9, r12 315 load_c v10, b_hilo_b, 0, r9, r10 317 vperm v0, v0, v1, v10 318 vperm v2, v2, v3, v10 349 load_c v10, b_0123_b, 0, r9, r12 417 load_c v10, b_hilo_b, 0, r9, r10 419 vperm v0, v0, v1, v10 420 vperm v2, v2, v3, v10 [all...] |
H A D | filter_altivec.asm | 49 lvx v10, 0, r9 ;# v10..v14 = first 5 rows 125 luma_v v10, v11, v12, v13, v14, v15 126 luma_v v11, v12, v13, v14, v15, v10 131 luma_v v12, v13, v14, v15, v10, v11 132 luma_v v13, v14, v15, v10, v11, v12 137 luma_v v14, v15, v10, v11, v12, v13 138 luma_v v15, v10, v11, v12, v13, v14 509 load_c v10, b_hilo, 0, r9, r10 511 vperm v0, v0, v1, v10 [all...] |
/external/libvpx/libvpx/vp8/common/ppc/ |
H A D | variance_subpixel_altivec.asm | 205 load_c v10, b_0123_b, 0, r12, r0 208 hfilter_8 v0, v10, v11, 1 209 hfilter_8 v1, v10, v11, 1 210 hfilter_8 v2, v10, v11, 1 211 hfilter_8 v3, v10, v11, 1 219 hfilter_8 v4, v10, v11, 0 261 load_c v10, b_hilo_b, 0, r12, r0 263 vperm v0, v0, v1, v10 264 vperm v1, v2, v3, v10 296 load_c v10, b_0123_ [all...] |
H A D | loopfilter_filters_altivec.asm | 118 Tpair v20,v21, v2,v10 132 Tpair v10,v11, v21,v29 244 ;# The input/output is in registers v0...v7. We use v10...v17 as mirrors; 248 ;# result lands in the "mirror" registers v10...v17 251 Tpair v10, v11, v0, v4 258 Tpair v0, v1, v10, v14 288 vmrghb v10, v21, v29 298 vmrghb v20, v2, v10 299 vmrglb v21, v2, v10 320 vmrghb v10, v2 [all...] |
H A D | filter_bilinear_altivec.asm | 79 vperm v24, v21, v21, v10 ;# v20 = 0123 1234 2345 3456 178 load_c v10, b_0123_b, 0, r9, r12 263 load_c v10, b_0123_b, 0, r9, r12 315 load_c v10, b_hilo_b, 0, r9, r10 317 vperm v0, v0, v1, v10 318 vperm v2, v2, v3, v10 349 load_c v10, b_0123_b, 0, r9, r12 417 load_c v10, b_hilo_b, 0, r9, r10 419 vperm v0, v0, v1, v10 420 vperm v2, v2, v3, v10 [all...] |
H A D | filter_altivec.asm | 49 lvx v10, 0, r9 ;# v10..v14 = first 5 rows 125 luma_v v10, v11, v12, v13, v14, v15 126 luma_v v11, v12, v13, v14, v15, v10 131 luma_v v12, v13, v14, v15, v10, v11 132 luma_v v13, v14, v15, v10, v11, v12 137 luma_v v14, v15, v10, v11, v12, v13 138 luma_v v15, v10, v11, v12, v13, v14 509 load_c v10, b_hilo, 0, r9, r10 511 vperm v0, v0, v1, v10 [all...] |
/external/libhevc/common/arm64/ |
H A D | ihevc_sao_band_offset_chroma.s | 163 LD1 {v10.8b},[x14],#8 //band_table_v.val[1] 223 ADD v14.8b, v10.8b , v30.8b //band_table_v.val[1] = vadd_u8(band_table_v.val[1], band_pos_v) 235 ADD v10.8b, v14.8b , v28.8b //band_table_v.val[1] = vadd_u8(band_table_v.val[1], vdup_n_u8(pi1_sao_offset_v[2])) 266 cmhs v19.8b, v29.8b , v10.8b //vcle_u8(band_table.val[1], vdup_n_u8(16)) 268 ORR v10.8b, v10.8b , v19.8b //band_table.val[1] = vorr_u8(band_table.val[1], au1_cmp) 280 AND v10.8b, v10.8b , v19.8b //band_table.val[1] = vand_u8(band_table.val[1], au1_cmp) 296 mov v9.d[1],v10.d[0] 297 mov v10 [all...] |
H A D | ihevc_inter_pred_chroma_horz.s | 197 ld1 { v10.2s},[x4],x11 //vector load pu1_src 240 umull v22.8h, v10.8b, v25.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 295 ld1 { v10.2s},[x4],x11 //vector load pu1_src 354 umull v22.8h, v10.8b, v25.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 394 ld1 { v10.2s},[x4],x11 //vector load pu1_src 425 umull v22.8h, v10.8b, v25.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 499 umull v10.8h, v5.8b, v25.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 500 umlsl v10.8h, v4.8b, v24.8b //mul_res = vmlsl_u8(src[0_2], coeffabs_2)// 504 umlal v10.8h, v6.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 505 umlsl v10 [all...] |
H A D | ihevc_itrans_recon_32x32.s | 211 ld1 {v10.4h},[x0],x6 230 smull v20.4s, v10.4h, v0.4h[0] 234 smull v22.4s, v10.4h, v0.4h[0] 237 smull v16.4s, v10.4h, v0.4h[0] 240 smull v18.4s, v10.4h, v0.4h[0] 280 ld1 {v10.4h},[x0],x6 300 smlal v20.4s, v10.4h, v2.4h[0] 304 smlal v22.4s, v10.4h, v6.4h[0] 307 smlsl v16.4s, v10.4h, v6.4h[0] 310 smlsl v18.4s, v10 [all...] |
H A D | ihevc_intra_pred_chroma_mode_27_to_33.s | 169 umull v10.8h, v23.8b, v30.8b //(i row)vmull_u8(ref_main_idx, dup_const_32_fract) 172 umlal v10.8h, v9.8b, v31.8b //(i row)vmull_u8(ref_main_idx_1, dup_const_fract) 187 rshrn v10.8b, v10.8h,#5 //(i row)shift_res = vrshrn_n_u16(add_res, 5) 201 st1 {v10.8b},[x2],#8 //(i row) 230 umull v10.8h, v23.8b, v30.8b //(v)vmull_u8(ref_main_idx, dup_const_32_fract) 233 umlal v10.8h, v9.8b, v31.8b //(v)vmull_u8(ref_main_idx_1, dup_const_fract) 253 rshrn v10.8b, v10.8h,#5 //(v)shift_res = vrshrn_n_u16(add_res, 5) 267 st1 {v10 [all...] |
H A D | ihevc_intra_pred_filters_luma_mode_19_to_25.s | 284 umull v10.8h, v23.8b, v30.8b //(i row)vmull_u8(ref_main_idx, dup_const_32_fract) 287 umlal v10.8h, v9.8b, v31.8b //(i row)vmull_u8(ref_main_idx_1, dup_const_fract) 301 rshrn v10.8b, v10.8h,#5 //(i row)shift_res = vrshrn_n_u16(add_res, 5) 315 st1 {v10.8b},[x2],#8 //(i row) 342 umull v10.8h, v23.8b, v30.8b //(v)vmull_u8(ref_main_idx, dup_const_32_fract) 345 umlal v10.8h, v9.8b, v31.8b //(v)vmull_u8(ref_main_idx_1, dup_const_fract) 364 rshrn v10.8b, v10.8h,#5 //(v)shift_res = vrshrn_n_u16(add_res, 5) 378 st1 {v10 [all...] |
H A D | ihevc_intra_pred_luma_mode_27_to_33.s | 174 umull v10.8h, v23.8b, v30.8b //(i row)vmull_u8(ref_main_idx, dup_const_32_fract) 177 umlal v10.8h, v9.8b, v31.8b //(i row)vmull_u8(ref_main_idx_1, dup_const_fract) 192 rshrn v10.8b, v10.8h,#5 //(i row)shift_res = vrshrn_n_u16(add_res, 5) 206 st1 {v10.8b},[x2],#8 //(i row) 235 umull v10.8h, v23.8b, v30.8b //(v)vmull_u8(ref_main_idx, dup_const_32_fract) 238 umlal v10.8h, v9.8b, v31.8b //(v)vmull_u8(ref_main_idx_1, dup_const_fract) 258 rshrn v10.8b, v10.8h,#5 //(v)shift_res = vrshrn_n_u16(add_res, 5) 272 st1 {v10 [all...] |
H A D | ihevc_inter_pred_chroma_horz_w16out.s | 212 ld1 { v10.2s},[x4],x11 //vector load pu1_src 252 umull v22.8h, v10.8b, v25.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 302 ld1 { v10.2s},[x4],x11 //vector load pu1_src 347 umull v22.8h, v10.8b, v25.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 391 ld1 { v10.2s},[x4],x11 //vector load pu1_src 415 umull v22.8h, v10.8b, v25.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 482 umull v10.8h, v5.8b, v25.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 483 umlsl v10.8h, v4.8b, v24.8b //mul_res = vmlsl_u8(src[0_2], coeffabs_2)// 486 umlal v10.8h, v6.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 487 umlsl v10 [all...] |
H A D | ihevc_inter_pred_filters_luma_horz.s | 246 umull v10.8h, v15.8b, v27.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 248 umlsl v10.8h, v14.8b, v26.8b //mul_res = vmlsl_u8(src[0_2], coeffabs_2)// 250 umlal v10.8h, v16.8b, v28.8b //mul_res = vmlal_u8(src[0_4], coeffabs_4)// 252 umlsl v10.8h, v17.8b, v29.8b //mul_res = vmlsl_u8(src[0_5], coeffabs_5)// 254 umlal v10.8h, v18.8b, v30.8b //mul_res = vmlal_u8(src[0_6], coeffabs_6)// 255 umlsl v10.8h, v19.8b, v31.8b //mul_res = vmlsl_u8(src[0_7], coeffabs_7)// 257 umlsl v10.8h, v12.8b, v24.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 258 umlal v10.8h, v13.8b, v25.8b //mul_res = vmlal_u8(src[0_1], coeffabs_1)// 262 sqrshrun v8.8b, v10.8h,#6 //right shift and saturating narrow result 2 376 umull v10 [all...] |
H A D | ihevc_inter_pred_luma_horz_w16out.s | 432 umull v10.8h, v15.8b, v27.8b //mul_res = vmull_u8(src[0_3], coeffabs_3)// 434 umlsl v10.8h, v14.8b, v26.8b //mul_res = vmlsl_u8(src[0_2], coeffabs_2)// 436 umlal v10.8h, v16.8b, v28.8b //mul_res = vmlal_u8(src[0_4], coeffabs_4)// 438 umlsl v10.8h, v17.8b, v29.8b //mul_res = vmlsl_u8(src[0_5], coeffabs_5)// 440 umlal v10.8h, v18.8b, v30.8b //mul_res = vmlal_u8(src[0_6], coeffabs_6)// 441 umlsl v10.8h, v19.8b, v31.8b //mul_res = vmlsl_u8(src[0_7], coeffabs_7)// 443 umlsl v10.8h, v12.8b, v24.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 444 umlal v10.8h, v13.8b, v25.8b //mul_res = vmlal_u8(src[0_1], coeffabs_1)// 450 st1 {v10.8h},[x10],#16 //store the result pu1_dst 558 umull v10 [all...] |
H A D | ihevc_itrans_recon_8x8.s | 197 ld1 {v10.4h},[x0],#8 208 smull v22.4s, v10.4h, v0.4h[0] //// y4 * cos4(part of c0 and c1) 242 add v10.4s, v20.4s , v22.4s //// c0 = y0 * cos4 + y4 * cos4(part of a0 and a1) 250 add v14.4s, v10.4s , v6.4s //// a0 = c0 + d0(part of x0,x7) 251 sub v10.4s, v10.4s , v6.4s //// a3 = c0 - d0(part of x3,x4) 264 add v26.4s, v10.4s , v30.4s //// a3 + b3(part of x3) 265 sub v30.4s, v10.4s , v30.4s //// a3 - b3(part of x4) 274 sqrshrn v10.4h, v30.4s,#shift_stage1_idct //// x4 = (a3 - b3 + rnd) >> 7(shift_stage1_idct) 321 sub v10 [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/ppc/ |
H A D | fdct_altivec.asm | 64 vmsumshm v10, v0, v8, v6 65 vmsumshm v10, v1, v9, v10 66 vsraw v10, v10, v7 ;# v10 = A0 A1 B0 B1 72 vpkuwum v10, v10, v11 ;# v10 = A0 A1 B0 B1 A2 A3 B2 B3 73 vperm \Dst, v10, v1 [all...] |
/external/libvpx/libvpx/vp8/encoder/ppc/ |
H A D | fdct_altivec.asm | 64 vmsumshm v10, v0, v8, v6 65 vmsumshm v10, v1, v9, v10 66 vsraw v10, v10, v7 ;# v10 = A0 A1 B0 B1 72 vpkuwum v10, v10, v11 ;# v10 = A0 A1 B0 B1 A2 A3 B2 B3 73 vperm \Dst, v10, v1 [all...] |
/external/pixman/test/ |
H A D | rotate-test.c | 26 #define TRANSFORM(v00, v01, v10, v11) \ 28 { v10, v11, HEIGHT * pixman_fixed_1 / 2 }, \
|